Compare commits
63 Commits
reduce-fro
...
readme
Author | SHA1 | Date | |
---|---|---|---|
eead35f822
|
|||
d7bef20ffc
|
|||
c2c8c81759
|
|||
98dbf5dfe2
|
|||
4236f33cb9
|
|||
60ce90b633
|
|||
c52fc69560
|
|||
e9a13c662e
|
|||
93625f12b1
|
|||
b7ed3f83b8
|
|||
f1a47c9621
|
|||
7843309f0a
|
|||
cebc9077c3
|
|||
f862e124a6
|
|||
1bb75b28eb
|
|||
0a7337a313
|
|||
fcd1ae78c3
|
|||
ec1b4100a3
|
|||
37ec05bf3b
|
|||
3434802b31
|
|||
bdbc72ab39
|
|||
4cc8f60076
|
|||
cf5c4c6224
|
|||
f69aee1c9c
|
|||
8c493f8fab
|
|||
4fce832db5
|
|||
bfc669cd11
|
|||
4056b3a739
|
|||
f7f0fdb980
|
|||
6c149e844b
|
|||
f5529bba24
|
|||
5da918c04b
|
|||
2eaf0f4ed5
|
|||
c42b687870
|
|||
d8adb4f52d
|
|||
5e7f440b54
|
|||
824536ddbe
|
|||
7fbdd0606c
|
|||
581f6b7b8f
|
|||
404b228cbf
|
|||
e2ff90cd5f
|
|||
cd700ebd42
|
|||
c44a52b7ae
|
|||
b3b36bc3de
|
|||
5e0d7b174c
|
|||
b437bbbf70
|
|||
85fdb28cc6
|
|||
b894d68a7a
|
|||
ef26509b87
|
|||
de82eb6c55
|
|||
0ba4c1e6fa
|
|||
0f1c16ab53
|
|||
5692bc96d5
|
|||
d15fee75ca
|
|||
93047c8706
|
|||
c84b78d3e2
|
|||
c6e3881966
|
|||
b713042359
|
|||
5505209760
|
|||
6aa28f7bfb
|
|||
07d43bfcb4
|
|||
6eb78d7307
|
|||
bcedf32bec
|
2
.gitattributes
vendored
Normal file
2
.gitattributes
vendored
Normal file
@@ -0,0 +1,2 @@
|
||||
medias/presentation.png filter=lfs diff=lfs merge=lfs -text
|
||||
medias/presentation.mp4 filter=lfs diff=lfs merge=lfs -text
|
124
README.md
Normal file
124
README.md
Normal file
@@ -0,0 +1,124 @@
|
||||
# 🚧 Carrramba! Encore raté! 🚧
|
||||
|
||||
Resident of the Ile-de-France,
|
||||
|
||||
- Tired of missing your bus/train/metro ?
|
||||
- Tired of having to walk to know where your bus/train/metro will stop ?
|
||||
- Are you looking to display its next passages in the same way as Ile De France Mobilité ?
|
||||
|
||||
**Visit [carrramba.adrien.run](https://carrramba.adrien.run/)**
|
||||
|
||||
[](medias/presentation.mp4)
|
||||
|
||||
# Technical stack
|
||||
|
||||
The following figure list the building blocks (docker images) and their interactions to provide the service:
|
||||
|
||||
```plantuml
|
||||
@startuml
|
||||
!include <C4/C4_Container>
|
||||
|
||||
!define ICONURL https://raw.githubusercontent.com/tupadr3/plantuml-icon-font-sprites/v2.4.0
|
||||
!includeurl ICONURL/common.puml
|
||||
|
||||
!includeurl ICONURL/font-awesome-5/users.puml
|
||||
!includeurl ICONURL/devicons2/redis.puml
|
||||
!includeurl ICONURL/devicons2/sqlalchemy.puml
|
||||
!includeurl ICONURL/devicons2/postgresql.puml
|
||||
|
||||
title Carrramba-encore-rate - Back-end
|
||||
|
||||
LAYOUT_WITH_LEGEND()
|
||||
|
||||
Person(user, "User", "The service user", $sprite="users")
|
||||
|
||||
System_Boundary(cb1, "carrramba-encore-rate") {
|
||||
Container(app, "SPA", "SolidJS", "The graphical interface used by users to consume services provided")
|
||||
Container(api, "carrramba-encore-rate-api", "FastAPI / Pydantic 2 / SQLAlchemy 2", "Provides the functionalities serving API endpoints", $sprite="sqlalchemy")
|
||||
ContainerDb(db, "Postgres", "PostgreSQL database", "Stores stops/stop areas/lines/shapes and associations information.", $sprite="postgresql")
|
||||
ContainerDb(cache, "In-Memory cache", "Redis", "Store previously computed results (stop, line, destination, next passages, shapes).", $sprite="redis")
|
||||
|
||||
Container(db_updater, "db-updater", "Sync the service static data with the IDFM ones")
|
||||
}
|
||||
|
||||
System_Boundary(idfm, "IDFM") {
|
||||
Container_Ext(static_, "Static data")
|
||||
Container_Ext(dynamic_, "Dynamic data")
|
||||
}
|
||||
|
||||
Rel(user, app, "Uses")
|
||||
|
||||
Rel(app, api, "Uses", "JSON/HTTPS")
|
||||
|
||||
Rel_R(api, db, "Reads from", "sync, PSQL")
|
||||
Rel_L(api, cache, "Reads from and writes to", "sync, REdis Serialization Protocol")
|
||||
Rel(api, dynamic_, "Get next passages", "JSON/HTTPS")
|
||||
|
||||
Rel_L(db_updater, db, "Writes to", "sync, PSQL")
|
||||
Rel(db_updater, static_, "Get stops, lines andshapes")
|
||||
@enduml
|
||||
```
|
||||
|
||||
## Back-end
|
||||
|
||||
Conventional but efficient:
|
||||
|
||||
- [FastAPI](https://fastapi.tiangolo.com/): _FastAPI is a modern, fast (high-performance), web framework for building
|
||||
APIs with Python 3.8+ based on standard Python type hints._
|
||||
- [Pydantic 2](https://docs.pydantic.dev/latest/): _Pydantic is the most widely used data validation library for
|
||||
Python._
|
||||
- [Sqlalchemy 2](https://www.sqlalchemy.org/): _SQLAlchemy is the Python SQL toolkit and Object Relational Mapper that
|
||||
gives application developers the full power and flexibility of SQL._
|
||||
|
||||
The [Msgspec](https://github.com/jcrist/msgspec) library is used to serialize JSON objects collected from IDFM API
|
||||
endpoints, this library being faster than Pydantic.
|
||||
|
||||
## Front-end
|
||||
|
||||
The component is developed using the [SolidJS](https://www.solidjs.com/) library. It brings the following interesting specificities:
|
||||
|
||||
- Share with React the same programmatic structures and support for component.
|
||||
- Fine-grained reactivity architecture: no virtual DOM used to update the components, the browser DOM is directly
|
||||
updated by SolidJS.
|
||||
|
||||
The front-end tries to be as close as possible to the design defined by the IDFM for the displays deployed by the
|
||||
transport operators in Ile-de-France. These specifications are public and available here:
|
||||
[PRIM-IDFM](https://prim.iledefrance-mobilites.fr/en/chartes-et-prescriptions).
|
||||
|
||||
|
||||
# TODO
|
||||
|
||||
## Features
|
||||
|
||||
- [ ] Integration with [Matrix.org](https://matrix.org/ecosystem/integrations/) ecosystem: make the app able to send message to a room when a bus/train/metro will
|
||||
arriving in X minutes.
|
||||
- [ ] Add the capability for the users to pin his/her favorite stops.
|
||||
- [ ] Add the address to the stop location.
|
||||
|
||||
## Docs
|
||||
|
||||
- [ ] Describe how to build the front-end and back-end docker images.
|
||||
- [ ] Describe how to deploy them using `docker compose`.
|
||||
- [ ] Add back-end API description ([openAPI](https://www.openapis.org/)) + generate documentation
|
||||
([mkdoc](https://www.mkdocs.org/), [Redoc](https://github.com/Redocly/redoc)). The best would be to build and
|
||||
deploy it (as docker images) using CI/CD... need to test [Agola](https://agola.io/), [Jaypore
|
||||
CI](https://www.jayporeci.in/) (pipelines configured in Python),
|
||||
[Woodpecker](https://woodpecker-ci.org/docs/intro) or [Drone](https://www.drone.io/) before.
|
||||
|
||||
## Front-end
|
||||
|
||||
- [ ] Add unit tests.
|
||||
- [ ] Make the StopNameInput component liquid.
|
||||
- [ ] Liquid to responsive Design.
|
||||
|
||||
## Back-end
|
||||
|
||||
- [ ] Add unit tests.
|
||||
- [ ] Use [alembic](https://alembic.sqlalchemy.org/en/latest/) to manage the future updates of the database schemas.
|
||||
- [ ] Add the capability to reload the application configuration on configuration file update (e.g.: credential update by the vault).
|
||||
- [ ] Rework how the database is updated with the IDFM data: For now the database is cleaned before refilling. It could
|
||||
be useful to avoid to empty the database and only apply deltas (update/add/remove rows).
|
||||
- [ ] Could be nice to compare FastAPI with [Litestar](https://litestar.dev/).
|
||||
|
||||
<!-- LocalWords: specificities
|
||||
-->
|
12
backend/.dockerignore
Normal file
12
backend/.dockerignore
Normal file
@@ -0,0 +1,12 @@
|
||||
.dir-locals.el
|
||||
.dockerignore
|
||||
.gitignore
|
||||
**/.mypy_cache
|
||||
**/.ruff_cache
|
||||
.venv
|
||||
**/__pycache__
|
||||
config
|
||||
docker
|
||||
poetry.lock
|
||||
tests
|
||||
Dockerfile
|
36
backend/Dockerfile.api_server
Normal file
36
backend/Dockerfile.api_server
Normal file
@@ -0,0 +1,36 @@
|
||||
FROM python:3.11-slim as builder
|
||||
|
||||
RUN pip install poetry
|
||||
|
||||
ENV POETRY_NO_INTERACTION=1 \
|
||||
POETRY_VIRTUALENVS_IN_PROJECT=1 \
|
||||
POETRY_VIRTUALENVS_CREATE=1 \
|
||||
POETRY_CACHE_DIR=/tmp/poetry_cache
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
COPY pyproject.toml /app
|
||||
|
||||
RUN poetry install --only=main --no-root && \
|
||||
rm -rf ${POETRY_CACHE_DIR}
|
||||
|
||||
|
||||
FROM python:3.11-slim as runtime
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
RUN apt update && \
|
||||
apt install -y --no-install-recommends libpq5 && \
|
||||
apt clean && \
|
||||
rm -rf /var/lib/apt/lists/*
|
||||
|
||||
env VIRTUAL_ENV=/app/.venv \
|
||||
PATH="/app/.venv/bin:$PATH"
|
||||
|
||||
COPY --from=builder ${VIRTUAL_ENV} ${VIRTUAL_ENV}
|
||||
|
||||
COPY api /app/api
|
||||
COPY config.sample.yaml .
|
||||
COPY api_server.py .
|
||||
|
||||
CMD ["./api_server.py"]
|
41
backend/Dockerfile.db_updater
Normal file
41
backend/Dockerfile.db_updater
Normal file
@@ -0,0 +1,41 @@
|
||||
FROM python:3.11-slim as builder
|
||||
|
||||
RUN apt update && \
|
||||
apt install -y --no-install-recommends proj-bin && \
|
||||
apt clean && \
|
||||
rm -rf /var/lib/apt/lists/*
|
||||
|
||||
RUN pip install poetry
|
||||
|
||||
ENV POETRY_NO_INTERACTION=1 \
|
||||
POETRY_VIRTUALENVS_IN_PROJECT=1 \
|
||||
POETRY_VIRTUALENVS_CREATE=1 \
|
||||
POETRY_CACHE_DIR=/tmp/poetry_cache
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
COPY ./pyproject.toml /app
|
||||
|
||||
RUN poetry install --only=db_updater --no-root && \
|
||||
rm -rf ${POETRY_CACHE_DIR}
|
||||
|
||||
|
||||
FROM python:3.11-slim as runtime
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
RUN apt update && \
|
||||
apt install -y --no-install-recommends libpq5 && \
|
||||
apt clean && \
|
||||
rm -rf /var/lib/apt/lists/*
|
||||
|
||||
env VIRTUAL_ENV=/app/.venv \
|
||||
PATH="/app/.venv/bin:$PATH"
|
||||
|
||||
COPY --from=builder ${VIRTUAL_ENV} ${VIRTUAL_ENV}
|
||||
|
||||
COPY api /app/api
|
||||
COPY config.sample.yaml .
|
||||
COPY db_updater.py .
|
||||
|
||||
CMD ["./db_updater.py"]
|
@@ -0,0 +1,9 @@
|
||||
|
||||
|
||||
# Architecture
|
||||
|
||||
|
||||
|
||||
The following schema shows the components used to satisfy the `api` service:
|
||||
|
||||

|
||||
|
58
backend/api/db/base_class.py
Normal file
58
backend/api/db/base_class.py
Normal file
@@ -0,0 +1,58 @@
|
||||
from __future__ import annotations
|
||||
|
||||
from logging import getLogger
|
||||
from typing import Self, Sequence, TYPE_CHECKING
|
||||
|
||||
from sqlalchemy import select
|
||||
from sqlalchemy.exc import IntegrityError
|
||||
from sqlalchemy.orm import DeclarativeBase
|
||||
|
||||
if TYPE_CHECKING:
|
||||
from .db import Database
|
||||
|
||||
logger = getLogger(__name__)
|
||||
|
||||
|
||||
class Base(DeclarativeBase):
|
||||
db: Database | None = None
|
||||
|
||||
@classmethod
|
||||
async def add(cls, objs: Sequence[Self]) -> bool:
|
||||
if cls.db is not None and (session := await cls.db.get_session()) is not None:
|
||||
|
||||
try:
|
||||
async with session.begin():
|
||||
session.add_all(objs)
|
||||
|
||||
except IntegrityError as err:
|
||||
logger.warning(err)
|
||||
return await cls.merge(objs)
|
||||
|
||||
except AttributeError as err:
|
||||
logger.error(err)
|
||||
return False
|
||||
|
||||
return True
|
||||
|
||||
@classmethod
|
||||
async def merge(cls, objs: Sequence[Self]) -> bool:
|
||||
if cls.db is not None and (session := await cls.db.get_session()) is not None:
|
||||
|
||||
async with session.begin():
|
||||
for obj in objs:
|
||||
await session.merge(obj)
|
||||
|
||||
return True
|
||||
|
||||
return False
|
||||
|
||||
@classmethod
|
||||
async def get_by_id(cls, id_: int | str) -> Self | None:
|
||||
if cls.db is not None and (session := await cls.db.get_session()) is not None:
|
||||
|
||||
async with session.begin():
|
||||
stmt = select(cls).where(cls.id == id_)
|
||||
res = await session.execute(stmt)
|
||||
return res.scalar_one_or_none()
|
||||
|
||||
return None
|
76
backend/api/db/db.py
Normal file
76
backend/api/db/db.py
Normal file
@@ -0,0 +1,76 @@
|
||||
from asyncio import sleep
|
||||
from logging import getLogger
|
||||
from typing import Annotated, AsyncIterator
|
||||
|
||||
from fastapi import Depends
|
||||
from opentelemetry.instrumentation.sqlalchemy import SQLAlchemyInstrumentor
|
||||
from sqlalchemy import text
|
||||
from sqlalchemy.exc import OperationalError, SQLAlchemyError
|
||||
from sqlalchemy.ext.asyncio import (
|
||||
async_sessionmaker,
|
||||
AsyncEngine,
|
||||
AsyncSession,
|
||||
create_async_engine,
|
||||
)
|
||||
|
||||
from .base_class import Base
|
||||
from ..settings import DatabaseSettings
|
||||
|
||||
logger = getLogger(__name__)
|
||||
|
||||
|
||||
class Database:
|
||||
def __init__(self) -> None:
|
||||
self._async_engine: AsyncEngine | None = None
|
||||
self._async_session_local: async_sessionmaker[AsyncSession] | None = None
|
||||
|
||||
async def get_session(self) -> AsyncSession | None:
|
||||
try:
|
||||
return self._async_session_local() # type: ignore
|
||||
|
||||
except (SQLAlchemyError, AttributeError) as e:
|
||||
logger.exception(e)
|
||||
|
||||
return None
|
||||
|
||||
# TODO: Preserve UserLastStopSearchResults table from drop.
|
||||
async def connect(
|
||||
self, settings: DatabaseSettings, clear_static_data: bool = False
|
||||
) -> bool:
|
||||
password = settings.password
|
||||
path = (
|
||||
f"{settings.driver}://{settings.user}:"
|
||||
f"{password.get_secret_value() if password is not None else ''}"
|
||||
f"@{settings.host}:{settings.port}/{settings.name}"
|
||||
)
|
||||
self._async_engine = create_async_engine(
|
||||
path, pool_pre_ping=True, pool_size=10, max_overflow=20
|
||||
)
|
||||
|
||||
if self._async_engine is not None:
|
||||
SQLAlchemyInstrumentor().instrument(engine=self._async_engine.sync_engine)
|
||||
|
||||
self._async_session_local = async_sessionmaker(
|
||||
bind=self._async_engine,
|
||||
# autoflush=False,
|
||||
expire_on_commit=False,
|
||||
class_=AsyncSession,
|
||||
)
|
||||
|
||||
ret = False
|
||||
while not ret:
|
||||
try:
|
||||
async with self._async_engine.begin() as session:
|
||||
if clear_static_data:
|
||||
await session.run_sync(Base.metadata.drop_all)
|
||||
await session.run_sync(Base.metadata.create_all)
|
||||
ret = True
|
||||
except OperationalError as err:
|
||||
logger.error(err)
|
||||
await sleep(1)
|
||||
|
||||
return True
|
||||
|
||||
async def disconnect(self) -> None:
|
||||
if self._async_engine is not None:
|
||||
await self._async_engine.dispose()
|
38
backend/api/dependencies.py
Normal file
38
backend/api/dependencies.py
Normal file
@@ -0,0 +1,38 @@
|
||||
from os import environ
|
||||
|
||||
from fastapi_cache.backends.redis import RedisBackend
|
||||
from redis import asyncio as aioredis
|
||||
from yaml import safe_load
|
||||
|
||||
from .db import db
|
||||
from .idfm_interface.idfm_interface import IdfmInterface
|
||||
from .settings import CacheSettings, Settings
|
||||
|
||||
|
||||
CONFIG_PATH = environ.get("CONFIG_PATH", "./config.sample.yaml")
|
||||
|
||||
|
||||
def load_settings(path: str) -> Settings:
|
||||
with open(path, "r") as config_file:
|
||||
config = safe_load(config_file)
|
||||
|
||||
return Settings(**config)
|
||||
|
||||
|
||||
settings = load_settings(CONFIG_PATH)
|
||||
|
||||
idfm_interface = IdfmInterface(settings.idfm_api_key.get_secret_value(), db)
|
||||
|
||||
|
||||
def init_redis_backend(settings: CacheSettings) -> RedisBackend:
|
||||
login = f"{settings.user}:{settings.password}@" if settings.user is not None else ""
|
||||
|
||||
url = f"redis://{login}{settings.host}:{settings.port}"
|
||||
|
||||
redis_connections_pool = aioredis.from_url(
|
||||
url, encoding="utf8", decode_responses=True
|
||||
)
|
||||
return RedisBackend(redis_connections_pool)
|
||||
|
||||
|
||||
redis_backend = init_redis_backend(settings.cache)
|
@@ -1,5 +1,3 @@
|
||||
from .idfm_interface import IdfmInterface
|
||||
|
||||
from .idfm_types import (
|
||||
Coordinate,
|
||||
Destinations,
|
||||
@@ -38,7 +36,6 @@ __all__ = [
|
||||
"Coordinate",
|
||||
"Destinations",
|
||||
"FramedVehicleJourney",
|
||||
"IdfmInterface",
|
||||
"IdfmLineState",
|
||||
"IdfmOperator",
|
||||
"IdfmResponse",
|
115
backend/api/idfm_interface/idfm_interface.py
Normal file
115
backend/api/idfm_interface/idfm_interface.py
Normal file
@@ -0,0 +1,115 @@
|
||||
from collections import defaultdict
|
||||
from re import compile as re_compile
|
||||
from typing import ByteString
|
||||
|
||||
from aiofiles import open as async_open
|
||||
from aiohttp import ClientSession
|
||||
from msgspec import ValidationError
|
||||
from msgspec.json import Decoder
|
||||
|
||||
from .idfm_types import Destinations as IdfmDestinations, IdfmResponse, IdfmState
|
||||
from ..db import Database
|
||||
from ..models import Line, Stop, StopArea
|
||||
|
||||
|
||||
class IdfmInterface:
|
||||
|
||||
IDFM_ROOT_URL = "https://prim.iledefrance-mobilites.fr/marketplace"
|
||||
IDFM_STOP_MON_URL = f"{IDFM_ROOT_URL}/stop-monitoring"
|
||||
|
||||
OPERATOR_RE = re_compile(r"[^:]+:Operator::([^:]+):")
|
||||
LINE_RE = re_compile(r"[^:]+:Line::C([^:]+):")
|
||||
|
||||
def __init__(self, api_key: str, database: Database) -> None:
|
||||
self._api_key = api_key
|
||||
self._database = database
|
||||
|
||||
self._http_headers = {"Accept": "application/json", "apikey": self._api_key}
|
||||
|
||||
self._response_json_decoder = Decoder(type=IdfmResponse)
|
||||
|
||||
async def startup(self) -> None:
|
||||
...
|
||||
|
||||
@staticmethod
|
||||
def _format_line_id(line_id: str) -> int:
|
||||
return int(line_id[1:] if line_id[0] == "C" else line_id)
|
||||
|
||||
async def render_line_picto(self, line: Line) -> tuple[None | str, None | str]:
|
||||
line_picto_path = line_picto_format = None
|
||||
target = f"/tmp/{line.id}_repr"
|
||||
|
||||
picto = line.picto
|
||||
if picto is not None:
|
||||
if (picto_data := await self._get_line_picto(line)) is not None:
|
||||
async with async_open(target, "wb") as fd:
|
||||
await fd.write(bytes(picto_data))
|
||||
line_picto_path = target
|
||||
line_picto_format = picto.mime_type
|
||||
|
||||
return (line_picto_path, line_picto_format)
|
||||
|
||||
async def _get_line_picto(self, line: Line) -> ByteString | None:
|
||||
data = None
|
||||
|
||||
picto = line.picto
|
||||
if picto is not None and picto.url is not None:
|
||||
headers = (
|
||||
self._http_headers if picto.url.startswith(self.IDFM_ROOT_URL) else None
|
||||
)
|
||||
|
||||
async with ClientSession(headers=headers) as session:
|
||||
async with session.get(picto.url) as response:
|
||||
data = await response.read()
|
||||
|
||||
return data
|
||||
|
||||
async def get_next_passages(self, stop_point_id: int) -> IdfmResponse | None:
|
||||
ret = None
|
||||
params = {"MonitoringRef": f"STIF:StopPoint:Q:{stop_point_id}:"}
|
||||
|
||||
async with ClientSession(headers=self._http_headers) as session:
|
||||
async with session.get(self.IDFM_STOP_MON_URL, params=params) as response:
|
||||
|
||||
if response.status == 200:
|
||||
data = await response.read()
|
||||
try:
|
||||
ret = self._response_json_decoder.decode(data)
|
||||
except ValidationError as err:
|
||||
print(err)
|
||||
|
||||
return ret
|
||||
|
||||
async def get_destinations(self, stop_id: int) -> IdfmDestinations | None:
|
||||
destinations: IdfmDestinations = defaultdict(set)
|
||||
|
||||
if (stop := await Stop.get_by_id(stop_id)) is not None:
|
||||
expected_stop_ids = {stop.id}
|
||||
elif (stop_area := await StopArea.get_by_id(stop_id)) is not None:
|
||||
expected_stop_ids = {stop.id for stop in stop_area.stops}
|
||||
else:
|
||||
return None
|
||||
|
||||
if (res := await self.get_next_passages(stop_id)) is not None:
|
||||
|
||||
for delivery in res.Siri.ServiceDelivery.StopMonitoringDelivery:
|
||||
if delivery.Status == IdfmState.true:
|
||||
for stop_visit in delivery.MonitoredStopVisit:
|
||||
|
||||
monitoring_ref = stop_visit.MonitoringRef.value
|
||||
|
||||
try:
|
||||
monitored_stop_id = int(monitoring_ref.split(":")[-2])
|
||||
except (IndexError, ValueError):
|
||||
print(f"Unable to get stop id from {monitoring_ref}")
|
||||
continue
|
||||
|
||||
journey = stop_visit.MonitoredVehicleJourney
|
||||
if (
|
||||
dst_names := journey.DestinationName
|
||||
) and monitored_stop_id in expected_stop_ids:
|
||||
raw_line_id = journey.LineRef.value.split(":")[-2]
|
||||
line_id = IdfmInterface._format_line_id(raw_line_id)
|
||||
destinations[line_id].add(dst_names[0].value)
|
||||
|
||||
return destinations
|
@@ -116,24 +116,33 @@ class StopArea(Struct):
|
||||
record_timestamp: datetime
|
||||
|
||||
|
||||
class ConnectionArea(Struct):
|
||||
class ConnectionAreaFields(Struct, kw_only=True):
|
||||
zdcid: str
|
||||
zdcversion: str
|
||||
zdccreated: datetime
|
||||
zdcchanged: datetime
|
||||
zdcname: str
|
||||
zdcxepsg2154: int
|
||||
zdcyepsg2154: int
|
||||
zdcxepsg2154: int | None = None
|
||||
zdcyepsg2154: int | None = None
|
||||
zdctown: str
|
||||
zdcpostalregion: str
|
||||
zdctype: StopAreaType
|
||||
|
||||
|
||||
class ConnectionArea(Struct):
|
||||
datasetid: str
|
||||
recordid: str
|
||||
fields: ConnectionAreaFields
|
||||
record_timestamp: datetime
|
||||
|
||||
|
||||
class StopAreaStopAssociationFields(Struct, kw_only=True):
|
||||
arrid: str # TODO: use int ?
|
||||
artid: str | None = None
|
||||
arrversion: str
|
||||
zdcid: str
|
||||
pdeid: str | None = None
|
||||
pdeversion: int | None = None
|
||||
version: int
|
||||
zdaid: str
|
||||
zdaversion: str
|
||||
@@ -149,6 +158,7 @@ class StopAreaStopAssociation(Struct):
|
||||
|
||||
class IdfmLineState(Enum):
|
||||
active = "active"
|
||||
available_soon = "prochainement active"
|
||||
|
||||
|
||||
class LinePicto(Struct, rename={"id_": "id"}):
|
||||
@@ -197,7 +207,7 @@ class Line(Struct):
|
||||
|
||||
Lines = dict[str, Line]
|
||||
|
||||
Destinations = dict[str, set[str]]
|
||||
Destinations = dict[int, set[str]]
|
||||
|
||||
|
||||
# TODO: Set structs frozen
|
15
backend/api/idfm_interface/ratp_types.py
Normal file
15
backend/api/idfm_interface/ratp_types.py
Normal file
@@ -0,0 +1,15 @@
|
||||
from msgspec import Struct
|
||||
|
||||
|
||||
class PictoFieldsFile(Struct, rename={"id_": "id"}):
|
||||
id_: str
|
||||
height: int
|
||||
width: int
|
||||
filename: str
|
||||
thumbnail: bool
|
||||
format: str
|
||||
|
||||
|
||||
class Picto(Struct):
|
||||
indices_commerciaux: str
|
||||
noms_des_fichiers: PictoFieldsFile | None = None
|
@@ -5,13 +5,11 @@ from typing import Iterable, Self, Sequence
|
||||
from sqlalchemy import (
|
||||
BigInteger,
|
||||
Boolean,
|
||||
Column,
|
||||
Enum,
|
||||
ForeignKey,
|
||||
Integer,
|
||||
select,
|
||||
String,
|
||||
Table,
|
||||
)
|
||||
from sqlalchemy.orm import Mapped, mapped_column, relationship, selectinload
|
||||
from sqlalchemy.sql.expression import tuple_
|
||||
@@ -25,12 +23,14 @@ from ..idfm_interface.idfm_types import (
|
||||
)
|
||||
from .stop import _Stop
|
||||
|
||||
line_stop_association_table = Table(
|
||||
"line_stop_association_table",
|
||||
Base.metadata,
|
||||
Column("line_id", ForeignKey("lines.id")),
|
||||
Column("stop_id", ForeignKey("_stops.id")),
|
||||
)
|
||||
|
||||
class LineStopAssociations(Base):
|
||||
|
||||
id = mapped_column(BigInteger, primary_key=True)
|
||||
line_id = mapped_column(BigInteger, ForeignKey("lines.id"))
|
||||
stop_id = mapped_column(BigInteger, ForeignKey("_stops.id"))
|
||||
|
||||
__tablename__ = "line_stop_associations"
|
||||
|
||||
|
||||
class LinePicto(Base):
|
||||
@@ -53,7 +53,7 @@ class Line(Base):
|
||||
|
||||
db = db
|
||||
|
||||
id = mapped_column(String, primary_key=True)
|
||||
id = mapped_column(BigInteger, primary_key=True)
|
||||
|
||||
short_name = mapped_column(String)
|
||||
name = mapped_column(String, nullable=False)
|
||||
@@ -68,7 +68,7 @@ class Line(Base):
|
||||
colour_web_hexa = mapped_column(String, nullable=False)
|
||||
text_colour_hexa = mapped_column(String, nullable=False)
|
||||
|
||||
operator_id = mapped_column(String)
|
||||
operator_id = mapped_column(Integer)
|
||||
operator_name = mapped_column(String)
|
||||
|
||||
accessibility = mapped_column(Enum(IdfmState), nullable=False)
|
||||
@@ -83,7 +83,7 @@ class Line(Base):
|
||||
|
||||
stops: Mapped[list[_Stop]] = relationship(
|
||||
"_Stop",
|
||||
secondary=line_stop_association_table,
|
||||
secondary="line_stop_associations",
|
||||
back_populates="lines",
|
||||
lazy="selectin",
|
||||
)
|
||||
@@ -94,23 +94,24 @@ class Line(Base):
|
||||
async def get_by_name(
|
||||
cls, name: str, operator_name: None | str = None
|
||||
) -> Sequence[Self] | None:
|
||||
session = cls.db.session
|
||||
if session is None:
|
||||
return None
|
||||
if (session := await cls.db.get_session()) is not None:
|
||||
|
||||
filters = {"name": name}
|
||||
if operator_name is not None:
|
||||
filters["operator_name"] = operator_name
|
||||
async with session.begin():
|
||||
filters = {"name": name}
|
||||
if operator_name is not None:
|
||||
filters["operator_name"] = operator_name
|
||||
|
||||
stmt = (
|
||||
select(cls)
|
||||
.filter_by(**filters)
|
||||
.options(selectinload(cls.stops), selectinload(cls.picto))
|
||||
)
|
||||
res = await session.execute(stmt)
|
||||
lines = res.scalars().all()
|
||||
stmt = (
|
||||
select(cls)
|
||||
.filter_by(**filters)
|
||||
.options(selectinload(cls.stops), selectinload(cls.picto))
|
||||
)
|
||||
res = await session.execute(stmt)
|
||||
lines = res.scalars().all()
|
||||
|
||||
return lines
|
||||
return lines
|
||||
|
||||
return None
|
||||
|
||||
@classmethod
|
||||
async def _add_picto_to_line(cls, line: str | Self, picto: LinePicto) -> None:
|
||||
@@ -133,57 +134,63 @@ class Line(Base):
|
||||
|
||||
@classmethod
|
||||
async def add_pictos(cls, line_to_pictos: Iterable[tuple[str, LinePicto]]) -> bool:
|
||||
session = cls.db.session
|
||||
if session is None:
|
||||
return False
|
||||
if (session := await cls.db.get_session()) is not None:
|
||||
|
||||
await asyncio_gather(
|
||||
*[cls._add_picto_to_line(line, picto) for line, picto in line_to_pictos]
|
||||
)
|
||||
async with session.begin():
|
||||
await asyncio_gather(
|
||||
*[
|
||||
cls._add_picto_to_line(line, picto)
|
||||
for line, picto in line_to_pictos
|
||||
]
|
||||
)
|
||||
|
||||
await session.commit()
|
||||
return True
|
||||
|
||||
return True
|
||||
return False
|
||||
|
||||
@classmethod
|
||||
async def add_stops(cls, line_to_stop_ids: Iterable[tuple[str, str, int]]) -> int:
|
||||
session = cls.db.session
|
||||
if session is None:
|
||||
return 0
|
||||
if (session := await cls.db.get_session()) is not None:
|
||||
|
||||
line_names_ops, stop_ids = set(), set()
|
||||
for line_name, operator_name, stop_id in line_to_stop_ids:
|
||||
line_names_ops.add((line_name, operator_name))
|
||||
stop_ids.add(stop_id)
|
||||
async with session.begin():
|
||||
|
||||
lines_res = await session.execute(
|
||||
select(Line).where(
|
||||
tuple_(Line.name, Line.operator_name).in_(line_names_ops)
|
||||
)
|
||||
)
|
||||
line_names_ops, stop_ids = set(), set()
|
||||
for line_name, operator_name, stop_id in line_to_stop_ids:
|
||||
line_names_ops.add((line_name, operator_name))
|
||||
stop_ids.add(stop_id)
|
||||
|
||||
lines = defaultdict(list)
|
||||
for line in lines_res.scalars():
|
||||
lines[(line.name, line.operator_name)].append(line)
|
||||
|
||||
stops_res = await session.execute(select(_Stop).where(_Stop.id.in_(stop_ids)))
|
||||
stops = {stop.id: stop for stop in stops_res.scalars()}
|
||||
|
||||
found = 0
|
||||
for line_name, operator_name, stop_id in line_to_stop_ids:
|
||||
if (stop := stops.get(stop_id)) is not None:
|
||||
if (stop_lines := lines.get((line_name, operator_name))) is not None:
|
||||
for stop_line in stop_lines:
|
||||
stop_line.stops.append(stop)
|
||||
found += 1
|
||||
else:
|
||||
print(f"No line found for {line_name}/{operator_name}")
|
||||
else:
|
||||
print(
|
||||
f"No stop found for {stop_id} id"
|
||||
f"(used by {line_name}/{operator_name})"
|
||||
lines_res = await session.execute(
|
||||
select(Line).where(
|
||||
tuple_(Line.name, Line.operator_name).in_(line_names_ops)
|
||||
)
|
||||
)
|
||||
|
||||
await session.commit()
|
||||
lines = defaultdict(list)
|
||||
for line in lines_res.scalars():
|
||||
lines[(line.name, line.operator_name)].append(line)
|
||||
|
||||
return found
|
||||
stops_res = await session.execute(
|
||||
select(_Stop).where(_Stop.id.in_(stop_ids))
|
||||
)
|
||||
stops = {stop.id: stop for stop in stops_res.scalars()}
|
||||
|
||||
found = 0
|
||||
for line_name, operator_name, stop_id in line_to_stop_ids:
|
||||
if (stop := stops.get(stop_id)) is not None:
|
||||
if (
|
||||
stop_lines := lines.get((line_name, operator_name))
|
||||
) is not None:
|
||||
for stop_line in stop_lines:
|
||||
stop_line.stops.append(stop)
|
||||
found += 1
|
||||
else:
|
||||
print(f"No line found for {line_name}/{operator_name}")
|
||||
else:
|
||||
print(
|
||||
f"No stop found for {stop_id} id"
|
||||
f"(used by {line_name}/{operator_name})"
|
||||
)
|
||||
|
||||
return found
|
||||
|
||||
return 0
|
275
backend/api/models/stop.py
Normal file
275
backend/api/models/stop.py
Normal file
@@ -0,0 +1,275 @@
|
||||
from __future__ import annotations
|
||||
|
||||
from logging import getLogger
|
||||
from typing import Iterable, Sequence, TYPE_CHECKING
|
||||
|
||||
from sqlalchemy import (
|
||||
BigInteger,
|
||||
Computed,
|
||||
desc,
|
||||
Enum,
|
||||
Float,
|
||||
ForeignKey,
|
||||
func,
|
||||
Integer,
|
||||
JSON,
|
||||
select,
|
||||
String,
|
||||
)
|
||||
from sqlalchemy.orm import (
|
||||
mapped_column,
|
||||
Mapped,
|
||||
relationship,
|
||||
selectinload,
|
||||
with_polymorphic,
|
||||
)
|
||||
from sqlalchemy.schema import Index
|
||||
from sqlalchemy_utils.types.ts_vector import TSVectorType
|
||||
|
||||
from ..db import Base, db
|
||||
from ..idfm_interface.idfm_types import TransportMode, IdfmState, StopAreaType
|
||||
|
||||
if TYPE_CHECKING:
|
||||
from .line import Line
|
||||
|
||||
|
||||
logger = getLogger(__name__)
|
||||
|
||||
class StopAreaStopAssociations(Base):
|
||||
|
||||
id = mapped_column(BigInteger, primary_key=True)
|
||||
stop_id = mapped_column(BigInteger, ForeignKey("_stops.id"))
|
||||
stop_area_id = mapped_column(BigInteger, ForeignKey("stop_areas.id"))
|
||||
|
||||
__tablename__ = "stop_area_stop_associations"
|
||||
|
||||
|
||||
class _Stop(Base):
|
||||
|
||||
db = db
|
||||
|
||||
id = mapped_column(BigInteger, primary_key=True)
|
||||
kind = mapped_column(String)
|
||||
|
||||
name = mapped_column(String, nullable=False, index=True)
|
||||
town_name = mapped_column(String, nullable=False)
|
||||
postal_region = mapped_column(Integer, nullable=False)
|
||||
epsg3857_x = mapped_column(Float, nullable=False)
|
||||
epsg3857_y = mapped_column(Float, nullable=False)
|
||||
|
||||
version = mapped_column(String, nullable=False)
|
||||
created_ts = mapped_column(BigInteger)
|
||||
changed_ts = mapped_column(BigInteger, nullable=False)
|
||||
|
||||
lines: Mapped[list[Line]] = relationship(
|
||||
"Line",
|
||||
secondary="line_stop_associations",
|
||||
back_populates="stops",
|
||||
lazy="selectin",
|
||||
)
|
||||
areas: Mapped[list["StopArea"]] = relationship(
|
||||
"StopArea",
|
||||
secondary="stop_area_stop_associations",
|
||||
back_populates="stops",
|
||||
)
|
||||
connection_area_id: Mapped[int] = mapped_column(
|
||||
ForeignKey("connection_areas.id"), nullable=True
|
||||
)
|
||||
connection_area: Mapped["ConnectionArea"] = relationship(
|
||||
back_populates="stops", lazy="selectin"
|
||||
)
|
||||
|
||||
names_tsv = mapped_column(
|
||||
TSVectorType("name", "town_name", regconfig="french"),
|
||||
Computed("to_tsvector('french', name || ' ' || town_name)", persisted=True),
|
||||
)
|
||||
|
||||
__tablename__ = "_stops"
|
||||
__mapper_args__ = {"polymorphic_identity": "_stops", "polymorphic_on": kind}
|
||||
__table_args__ = (
|
||||
Index(
|
||||
"names_tsv_idx",
|
||||
names_tsv,
|
||||
postgresql_ops={"name": "gin_trgm_ops"},
|
||||
postgresql_using="gin",
|
||||
),
|
||||
)
|
||||
|
||||
@classmethod
|
||||
async def get_by_name(cls, name: str) -> Sequence[_Stop] | None:
|
||||
if (session := await cls.db.get_session()) is not None:
|
||||
|
||||
async with session.begin():
|
||||
descendants = with_polymorphic(_Stop, "*")
|
||||
|
||||
match_stmt = descendants.names_tsv.match(
|
||||
name, postgresql_regconfig="french"
|
||||
)
|
||||
ranking_stmt = func.ts_rank_cd(
|
||||
descendants.names_tsv, func.plainto_tsquery("french", name)
|
||||
)
|
||||
stmt = (
|
||||
select(descendants).filter(match_stmt).order_by(desc(ranking_stmt))
|
||||
)
|
||||
|
||||
res = await session.execute(stmt)
|
||||
stops = res.scalars().all()
|
||||
|
||||
return stops
|
||||
|
||||
return None
|
||||
|
||||
|
||||
class Stop(_Stop):
|
||||
|
||||
id = mapped_column(BigInteger, ForeignKey("_stops.id"), primary_key=True)
|
||||
|
||||
transport_mode = mapped_column(Enum(TransportMode), nullable=False)
|
||||
accessibility = mapped_column(Enum(IdfmState), nullable=False)
|
||||
visual_signs_available = mapped_column(Enum(IdfmState), nullable=False)
|
||||
audible_signs_available = mapped_column(Enum(IdfmState), nullable=False)
|
||||
|
||||
record_id = mapped_column(String, nullable=False)
|
||||
record_ts = mapped_column(BigInteger, nullable=False)
|
||||
|
||||
__tablename__ = "stops"
|
||||
__mapper_args__ = {"polymorphic_identity": "stops", "polymorphic_load": "inline"}
|
||||
|
||||
|
||||
class StopArea(_Stop):
|
||||
|
||||
id = mapped_column(BigInteger, ForeignKey("_stops.id"), primary_key=True)
|
||||
|
||||
type = mapped_column(Enum(StopAreaType), nullable=False)
|
||||
|
||||
stops: Mapped[list["Stop"]] = relationship(
|
||||
"Stop",
|
||||
secondary="stop_area_stop_associations",
|
||||
back_populates="areas",
|
||||
lazy="selectin",
|
||||
)
|
||||
|
||||
__tablename__ = "stop_areas"
|
||||
__mapper_args__ = {
|
||||
"polymorphic_identity": "stop_areas",
|
||||
"polymorphic_load": "inline",
|
||||
}
|
||||
|
||||
@classmethod
|
||||
async def add_stops(
|
||||
cls, stop_area_to_stop_ids: Iterable[tuple[int, int]]
|
||||
) -> int | None:
|
||||
if (session := await cls.db.get_session()) is not None:
|
||||
|
||||
async with session.begin():
|
||||
|
||||
stop_area_ids, stop_ids = set(), set()
|
||||
for stop_area_id, stop_id in stop_area_to_stop_ids:
|
||||
stop_area_ids.add(stop_area_id)
|
||||
stop_ids.add(stop_id)
|
||||
|
||||
stop_areas_res = await session.scalars(
|
||||
select(StopArea)
|
||||
.where(StopArea.id.in_(stop_area_ids))
|
||||
.options(selectinload(StopArea.stops))
|
||||
)
|
||||
stop_areas: dict[int, StopArea] = {
|
||||
stop_area.id: stop_area for stop_area in stop_areas_res.all()
|
||||
}
|
||||
|
||||
stop_res = await session.execute(
|
||||
select(Stop).where(Stop.id.in_(stop_ids))
|
||||
)
|
||||
stops: dict[int, Stop] = {stop.id: stop for stop in stop_res.scalars()}
|
||||
|
||||
found = 0
|
||||
for stop_area_id, stop_id in stop_area_to_stop_ids:
|
||||
if (stop_area := stop_areas.get(stop_area_id)) is not None:
|
||||
if (stop := stops.get(stop_id)) is not None:
|
||||
stop_area.stops.append(stop)
|
||||
found += 1
|
||||
else:
|
||||
print(f"No stop found for {stop_id} id")
|
||||
else:
|
||||
print(f"No stop area found for {stop_area_id}")
|
||||
|
||||
return found
|
||||
|
||||
return None
|
||||
|
||||
|
||||
class StopShape(Base):
|
||||
|
||||
db = db
|
||||
|
||||
id = mapped_column(BigInteger, primary_key=True) # Same id than ConnectionArea
|
||||
type = mapped_column(Integer, nullable=False)
|
||||
epsg3857_bbox = mapped_column(JSON)
|
||||
epsg3857_points = mapped_column(JSON)
|
||||
|
||||
__tablename__ = "stop_shapes"
|
||||
|
||||
|
||||
class ConnectionArea(Base):
|
||||
|
||||
db = db
|
||||
|
||||
id = mapped_column(BigInteger, primary_key=True)
|
||||
|
||||
name = mapped_column(String, nullable=False)
|
||||
town_name = mapped_column(String, nullable=False)
|
||||
postal_region = mapped_column(String, nullable=False)
|
||||
epsg3857_x = mapped_column(Float, nullable=False)
|
||||
epsg3857_y = mapped_column(Float, nullable=False)
|
||||
transport_mode = mapped_column(Enum(StopAreaType), nullable=False)
|
||||
|
||||
version = mapped_column(String, nullable=False)
|
||||
created_ts = mapped_column(BigInteger)
|
||||
changed_ts = mapped_column(BigInteger, nullable=False)
|
||||
|
||||
stops: Mapped[list["_Stop"]] = relationship(back_populates="connection_area")
|
||||
|
||||
__tablename__ = "connection_areas"
|
||||
|
||||
# TODO: Merge with StopArea.add_stops
|
||||
@classmethod
|
||||
async def add_stops(
|
||||
cls, conn_area_to_stop_ids: Iterable[tuple[int, int]]
|
||||
) -> int | None:
|
||||
if (session := await cls.db.get_session()) is not None:
|
||||
|
||||
async with session.begin():
|
||||
|
||||
conn_area_ids, stop_ids = set(), set()
|
||||
for conn_area_id, stop_id in conn_area_to_stop_ids:
|
||||
conn_area_ids.add(conn_area_id)
|
||||
stop_ids.add(stop_id)
|
||||
|
||||
conn_area_res = await session.execute(
|
||||
select(ConnectionArea)
|
||||
.where(ConnectionArea.id.in_(conn_area_ids))
|
||||
.options(selectinload(ConnectionArea.stops))
|
||||
)
|
||||
conn_areas: dict[int, ConnectionArea] = {
|
||||
conn.id: conn for conn in conn_area_res.scalars()
|
||||
}
|
||||
|
||||
stop_res = await session.execute(
|
||||
select(Stop).where(Stop.id.in_(stop_ids))
|
||||
)
|
||||
stops: dict[int, Stop] = {stop.id: stop for stop in stop_res.scalars()}
|
||||
|
||||
found = 0
|
||||
for conn_area_id, stop_id in conn_area_to_stop_ids:
|
||||
if (conn_area := conn_areas.get(conn_area_id)) is not None:
|
||||
if (stop := stops.get(stop_id)) is not None:
|
||||
conn_area.stops.append(stop)
|
||||
found += 1
|
||||
else:
|
||||
print(f"No stop found for {stop_id} id")
|
||||
else:
|
||||
print(f"No connection area found for {conn_area_id}")
|
||||
|
||||
return found
|
||||
|
||||
return None
|
27
backend/api/models/user.py
Normal file
27
backend/api/models/user.py
Normal file
@@ -0,0 +1,27 @@
|
||||
from sqlalchemy import BigInteger, ForeignKey, String
|
||||
from sqlalchemy.orm import Mapped, mapped_column, relationship
|
||||
|
||||
from ..db import Base, db
|
||||
from .stop import _Stop
|
||||
|
||||
|
||||
class UserLastStopSearchStopAssociations(Base):
|
||||
id = mapped_column(BigInteger, primary_key=True)
|
||||
user_mxid = mapped_column(
|
||||
String, ForeignKey("user_last_stop_search_results.user_mxid")
|
||||
)
|
||||
stop_id = mapped_column(BigInteger, ForeignKey("_stops.id"))
|
||||
|
||||
__tablename__ = "user_last_stop_search_stop_associations"
|
||||
|
||||
|
||||
class UserLastStopSearchResults(Base):
|
||||
db = db
|
||||
|
||||
user_mxid = mapped_column(String, primary_key=True)
|
||||
request_content = mapped_column(String, nullable=False)
|
||||
stops: Mapped[_Stop] = relationship(
|
||||
_Stop, secondary="user_last_stop_search_stop_associations"
|
||||
)
|
||||
|
||||
__tablename__ = "user_last_stop_search_results"
|
0
backend/api/routers/__init__.py
Normal file
0
backend/api/routers/__init__.py
Normal file
34
backend/api/routers/line.py
Normal file
34
backend/api/routers/line.py
Normal file
@@ -0,0 +1,34 @@
|
||||
from fastapi import APIRouter, HTTPException
|
||||
from fastapi_cache.decorator import cache
|
||||
|
||||
from ..models import Line
|
||||
from ..schemas import Line as LineSchema, TransportMode
|
||||
|
||||
|
||||
router = APIRouter(prefix="/line", tags=["line"])
|
||||
|
||||
|
||||
@router.get("/{line_id}", response_model=LineSchema)
|
||||
@cache(namespace="line")
|
||||
async def get_line(line_id: int) -> LineSchema:
|
||||
line: Line | None = await Line.get_by_id(line_id)
|
||||
|
||||
if line is None:
|
||||
raise HTTPException(status_code=404, detail=f'Line "{line_id}" not found')
|
||||
|
||||
return LineSchema(
|
||||
id=line.id,
|
||||
shortName=line.short_name,
|
||||
name=line.name,
|
||||
status=line.status,
|
||||
transportMode=TransportMode.from_idfm_transport_mode(
|
||||
line.transport_mode, line.transport_submode
|
||||
),
|
||||
backColorHexa=line.colour_web_hexa,
|
||||
foreColorHexa=line.text_colour_hexa,
|
||||
operatorId=line.operator_id,
|
||||
accessibility=line.accessibility,
|
||||
visualSignsAvailable=line.visual_signs_available,
|
||||
audibleSignsAvailable=line.audible_signs_available,
|
||||
stopIds=[stop.id for stop in line.stops],
|
||||
)
|
176
backend/api/routers/stop.py
Normal file
176
backend/api/routers/stop.py
Normal file
@@ -0,0 +1,176 @@
|
||||
from collections import defaultdict
|
||||
from datetime import datetime
|
||||
from typing import Sequence
|
||||
|
||||
from fastapi import APIRouter, HTTPException
|
||||
from fastapi_cache.decorator import cache
|
||||
|
||||
from ..idfm_interface import Destinations as IdfmDestinations, TrainStatus
|
||||
from ..models import Stop, StopArea, StopShape
|
||||
from ..schemas import (
|
||||
NextPassage as NextPassageSchema,
|
||||
NextPassages as NextPassagesSchema,
|
||||
Stop as StopSchema,
|
||||
StopArea as StopAreaSchema,
|
||||
StopShape as StopShapeSchema,
|
||||
)
|
||||
from ..dependencies import idfm_interface
|
||||
|
||||
|
||||
router = APIRouter(prefix="/stop", tags=["stop"])
|
||||
|
||||
|
||||
def _format_stop(stop: Stop) -> StopSchema:
|
||||
return StopSchema(
|
||||
id=stop.id,
|
||||
name=stop.name,
|
||||
town=stop.town_name,
|
||||
epsg3857_x=stop.epsg3857_x,
|
||||
epsg3857_y=stop.epsg3857_y,
|
||||
lines=[line.id for line in stop.lines],
|
||||
)
|
||||
|
||||
|
||||
def optional_datetime_to_ts(dt: datetime | None) -> int | None:
|
||||
return int(dt.timestamp()) if dt else None
|
||||
|
||||
|
||||
# TODO: Add limit support
|
||||
@router.get("/")
|
||||
@cache(namespace="stop")
|
||||
async def get_stop(
|
||||
name: str = "", limit: int = 10
|
||||
) -> Sequence[StopAreaSchema | StopSchema] | None:
|
||||
|
||||
matching_stops = await Stop.get_by_name(name)
|
||||
if matching_stops is None:
|
||||
return None
|
||||
|
||||
formatted: list[StopAreaSchema | StopSchema] = []
|
||||
stop_areas: dict[int, StopArea] = {}
|
||||
stops: dict[int, Stop] = {}
|
||||
for stop in matching_stops:
|
||||
if isinstance(stop, StopArea):
|
||||
stop_areas[stop.id] = stop
|
||||
elif isinstance(stop, Stop):
|
||||
stops[stop.id] = stop
|
||||
|
||||
for stop_area in stop_areas.values():
|
||||
|
||||
formatted_stops = []
|
||||
for stop in stop_area.stops:
|
||||
formatted_stops.append(_format_stop(stop))
|
||||
try:
|
||||
del stops[stop.id]
|
||||
except KeyError as err:
|
||||
print(err)
|
||||
|
||||
formatted.append(
|
||||
StopAreaSchema(
|
||||
id=stop_area.id,
|
||||
name=stop_area.name,
|
||||
town=stop_area.town_name,
|
||||
type=stop_area.type,
|
||||
lines=[line.id for line in stop_area.lines],
|
||||
stops=formatted_stops,
|
||||
)
|
||||
)
|
||||
|
||||
formatted.extend(_format_stop(stop) for stop in stops.values())
|
||||
|
||||
return formatted
|
||||
|
||||
|
||||
@router.get("/{stop_id}/nextPassages")
|
||||
@cache(namespace="stop-nextPassages", expire=30)
|
||||
async def get_next_passages(stop_id: int) -> NextPassagesSchema | None:
|
||||
res = await idfm_interface.get_next_passages(stop_id)
|
||||
if res is None:
|
||||
return None
|
||||
|
||||
service_delivery = res.Siri.ServiceDelivery
|
||||
stop_monitoring_deliveries = service_delivery.StopMonitoringDelivery
|
||||
|
||||
by_line_by_dst_passages: dict[
|
||||
int, dict[str, list[NextPassageSchema]]
|
||||
] = defaultdict(lambda: defaultdict(list))
|
||||
|
||||
for delivery in stop_monitoring_deliveries:
|
||||
for stop_visit in delivery.MonitoredStopVisit:
|
||||
|
||||
journey = stop_visit.MonitoredVehicleJourney
|
||||
|
||||
# re.match will return None if the given journey.LineRef.value is not valid.
|
||||
try:
|
||||
line_id_match = idfm_interface.LINE_RE.match(journey.LineRef.value)
|
||||
line_id = int(line_id_match.group(1)) # type: ignore
|
||||
except (AttributeError, TypeError, ValueError) as err:
|
||||
raise HTTPException(
|
||||
status_code=404, detail=f'Line "{journey.LineRef.value}" not found'
|
||||
) from err
|
||||
|
||||
call = journey.MonitoredCall
|
||||
|
||||
dst_names = call.DestinationDisplay
|
||||
dsts = [dst.value for dst in dst_names] if dst_names else []
|
||||
arrivalPlatformName = (
|
||||
call.ArrivalPlatformName.value if call.ArrivalPlatformName else None
|
||||
)
|
||||
|
||||
next_passage = NextPassageSchema(
|
||||
line=line_id,
|
||||
operator=journey.OperatorRef.value,
|
||||
destinations=dsts,
|
||||
atStop=call.VehicleAtStop,
|
||||
aimedArrivalTs=optional_datetime_to_ts(call.AimedArrivalTime),
|
||||
expectedArrivalTs=optional_datetime_to_ts(call.ExpectedArrivalTime),
|
||||
arrivalPlatformName=arrivalPlatformName,
|
||||
aimedDepartTs=optional_datetime_to_ts(call.AimedDepartureTime),
|
||||
expectedDepartTs=optional_datetime_to_ts(call.ExpectedDepartureTime),
|
||||
arrivalStatus=call.ArrivalStatus
|
||||
if call.ArrivalStatus is not None
|
||||
else TrainStatus.unknown,
|
||||
departStatus=call.DepartureStatus
|
||||
if call.DepartureStatus is not None
|
||||
else TrainStatus.unknown,
|
||||
)
|
||||
|
||||
by_line_passages = by_line_by_dst_passages[line_id]
|
||||
# TODO: by_line_passages[dst].extend(dsts) instead ?
|
||||
for dst in dsts:
|
||||
by_line_passages[dst].append(next_passage)
|
||||
|
||||
return NextPassagesSchema(
|
||||
ts=int(service_delivery.ResponseTimestamp.timestamp()),
|
||||
passages=by_line_by_dst_passages,
|
||||
)
|
||||
|
||||
|
||||
@router.get("/{stop_id}/destinations")
|
||||
@cache(namespace="stop-destinations", expire=30)
|
||||
async def get_stop_destinations(
|
||||
stop_id: int,
|
||||
) -> IdfmDestinations | None:
|
||||
destinations = await idfm_interface.get_destinations(stop_id)
|
||||
|
||||
return destinations
|
||||
|
||||
|
||||
@router.get("/{stop_id}/shape")
|
||||
@cache(namespace="stop-shape")
|
||||
async def get_stop_shape(stop_id: int) -> StopShapeSchema | None:
|
||||
if (await Stop.get_by_id(stop_id)) is not None or (
|
||||
await StopArea.get_by_id(stop_id)
|
||||
) is not None:
|
||||
shape_id = stop_id
|
||||
|
||||
if (shape := await StopShape.get_by_id(shape_id)) is not None:
|
||||
return StopShapeSchema(
|
||||
id=shape.id,
|
||||
type=shape.type,
|
||||
epsg3857_bbox=shape.epsg3857_bbox,
|
||||
epsg3857_points=shape.epsg3857_points,
|
||||
)
|
||||
|
||||
msg = f"No shape found for stop {stop_id}"
|
||||
raise HTTPException(status_code=404, detail=msg)
|
@@ -46,15 +46,15 @@ class TransportMode(StrEnum):
|
||||
|
||||
|
||||
class Line(BaseModel):
|
||||
id: str
|
||||
id: int
|
||||
shortName: str
|
||||
name: str
|
||||
status: IdfmLineState
|
||||
transportMode: TransportMode
|
||||
backColorHexa: str
|
||||
foreColorHexa: str
|
||||
operatorId: str
|
||||
operatorId: int
|
||||
accessibility: IdfmState
|
||||
visualSignsAvailable: IdfmState
|
||||
audibleSignsAvailable: IdfmState
|
||||
stopIds: list[str]
|
||||
stopIds: list[int]
|
@@ -4,7 +4,7 @@ from ..idfm_interface.idfm_types import TrainStatus
|
||||
|
||||
|
||||
class NextPassage(BaseModel):
|
||||
line: str
|
||||
line: int
|
||||
operator: str
|
||||
destinations: list[str]
|
||||
atStop: bool
|
||||
@@ -19,4 +19,4 @@ class NextPassage(BaseModel):
|
||||
|
||||
class NextPassages(BaseModel):
|
||||
ts: int
|
||||
passages: dict[str, dict[str, list[NextPassage]]]
|
||||
passages: dict[int, dict[str, list[NextPassage]]]
|
@@ -9,7 +9,7 @@ class Stop(BaseModel):
|
||||
town: str
|
||||
epsg3857_x: float
|
||||
epsg3857_y: float
|
||||
lines: list[str]
|
||||
lines: list[int]
|
||||
|
||||
|
||||
class StopArea(BaseModel):
|
||||
@@ -17,7 +17,7 @@ class StopArea(BaseModel):
|
||||
name: str
|
||||
town: str
|
||||
type: StopAreaType
|
||||
lines: list[str] # SNCF lines are linked to stop areas and not stops.
|
||||
lines: list[int] # SNCF lines are linked to stop areas and not stops.
|
||||
stops: list[Stop]
|
||||
|
||||
|
74
backend/api/settings.py
Normal file
74
backend/api/settings.py
Normal file
@@ -0,0 +1,74 @@
|
||||
from __future__ import annotations
|
||||
|
||||
from typing import Annotated
|
||||
|
||||
from pydantic import BaseModel, SecretStr
|
||||
from pydantic.functional_validators import model_validator
|
||||
from pydantic_settings import (
|
||||
BaseSettings,
|
||||
PydanticBaseSettingsSource,
|
||||
SettingsConfigDict,
|
||||
)
|
||||
|
||||
|
||||
class HttpSettings(BaseModel):
|
||||
host: str = "127.0.0.1"
|
||||
port: int = 8080
|
||||
cert: str | None = None
|
||||
|
||||
|
||||
class DatabaseSettings(BaseModel):
|
||||
name: str
|
||||
host: str
|
||||
port: int
|
||||
driver: str = "postgresql+psycopg"
|
||||
user: str
|
||||
password: Annotated[SecretStr, check_user_password]
|
||||
|
||||
|
||||
class CacheSettings(BaseModel):
|
||||
enable: bool = False
|
||||
host: str = "127.0.0.1"
|
||||
port: int = 6379
|
||||
user: str | None = None
|
||||
password: Annotated[SecretStr | None, check_user_password] = None
|
||||
|
||||
|
||||
@model_validator(mode="after")
|
||||
def check_user_password(self) -> DatabaseSettings | CacheSettings:
|
||||
if self.user is not None and self.password is None:
|
||||
raise ValueError("user is set, password shall be set too.")
|
||||
|
||||
if self.password is not None and self.user is None:
|
||||
raise ValueError("password is set, user shall be set too.")
|
||||
|
||||
return self
|
||||
|
||||
|
||||
class TracingSettings(BaseModel):
|
||||
enable: bool = False
|
||||
|
||||
|
||||
class Settings(BaseSettings):
|
||||
app_name: str
|
||||
|
||||
idfm_api_key: SecretStr
|
||||
clear_static_data: bool
|
||||
|
||||
http: HttpSettings
|
||||
db: DatabaseSettings
|
||||
cache: CacheSettings
|
||||
tracing: TracingSettings
|
||||
|
||||
model_config = SettingsConfigDict(env_prefix="CER__", env_nested_delimiter="__")
|
||||
|
||||
@classmethod
|
||||
def settings_customise_sources(
|
||||
cls,
|
||||
settings_cls: type[BaseSettings],
|
||||
init_settings: PydanticBaseSettingsSource,
|
||||
env_settings: PydanticBaseSettingsSource,
|
||||
dotenv_settings: PydanticBaseSettingsSource,
|
||||
file_secret_settings: PydanticBaseSettingsSource,
|
||||
) -> tuple[PydanticBaseSettingsSource, ...]:
|
||||
return env_settings, init_settings, file_secret_settings
|
89
backend/api_server.py
Executable file
89
backend/api_server.py
Executable file
@@ -0,0 +1,89 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
import uvicorn
|
||||
from contextlib import asynccontextmanager
|
||||
from fastapi import FastAPI, Request
|
||||
from fastapi.middleware.cors import CORSMiddleware
|
||||
from fastapi_cache import FastAPICache
|
||||
from opentelemetry import trace
|
||||
from opentelemetry.exporter.otlp.proto.http.trace_exporter import OTLPSpanExporter
|
||||
from opentelemetry.instrumentation.fastapi import FastAPIInstrumentor
|
||||
from opentelemetry.sdk.resources import Resource, SERVICE_NAME
|
||||
from opentelemetry.sdk.trace import TracerProvider
|
||||
from opentelemetry.sdk.trace.export import BatchSpanProcessor
|
||||
|
||||
from api.db import db
|
||||
from api.dependencies import idfm_interface, redis_backend, settings
|
||||
from api.routers import line, stop
|
||||
|
||||
|
||||
@asynccontextmanager
|
||||
async def lifespan(app: FastAPI):
|
||||
FastAPICache.init(redis_backend, prefix="api", enable=settings.cache.enable)
|
||||
|
||||
await db.connect(settings.db, settings.clear_static_data)
|
||||
if settings.clear_static_data:
|
||||
await idfm_interface.startup()
|
||||
|
||||
yield
|
||||
|
||||
await db.disconnect()
|
||||
|
||||
|
||||
app = FastAPI(lifespan=lifespan)
|
||||
|
||||
app.add_middleware(
|
||||
CORSMiddleware,
|
||||
allow_origins=["http://carrramba.adrien.run", "https://carrramba.adrien.run"],
|
||||
allow_credentials=True,
|
||||
allow_methods=["OPTIONS", "GET"],
|
||||
allow_headers=["*"],
|
||||
)
|
||||
|
||||
# The cache-control header entry is not managed properly by fastapi-cache:
|
||||
# For now, a request with a cache-control set to no-cache
|
||||
# is interpreted as disabling the use of the server cache.
|
||||
# Cf. Improve Cache-Control header parsing and handling
|
||||
# https://github.com/long2ice/fastapi-cache/issues/144 workaround
|
||||
@app.middleware("http")
|
||||
async def fastapi_cache_issue_144_workaround(request: Request, call_next):
|
||||
entries = request.headers.__dict__["_list"]
|
||||
new_entries = [
|
||||
entry for entry in entries if entry[0].decode().lower() != "cache-control"
|
||||
]
|
||||
|
||||
request.headers.__dict__["_list"] = new_entries
|
||||
|
||||
return await call_next(request)
|
||||
|
||||
|
||||
app.include_router(line.router)
|
||||
app.include_router(stop.router)
|
||||
|
||||
|
||||
if settings.tracing.enable:
|
||||
FastAPIInstrumentor.instrument_app(app)
|
||||
|
||||
trace.set_tracer_provider(
|
||||
TracerProvider(resource=Resource.create({SERVICE_NAME: settings.app_name}))
|
||||
)
|
||||
trace.get_tracer_provider().add_span_processor(
|
||||
BatchSpanProcessor(OTLPSpanExporter())
|
||||
)
|
||||
tracer = trace.get_tracer(settings.app_name)
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
http_settings = settings.http
|
||||
|
||||
config = uvicorn.Config(
|
||||
app=app,
|
||||
host=http_settings.host,
|
||||
port=http_settings.port,
|
||||
ssl_certfile=http_settings.cert,
|
||||
proxy_headers=True,
|
||||
)
|
||||
|
||||
server = uvicorn.Server(config)
|
||||
|
||||
server.run()
|
@@ -1,39 +0,0 @@
|
||||
from __future__ import annotations
|
||||
|
||||
from typing import Iterable, Self, TYPE_CHECKING
|
||||
|
||||
from sqlalchemy import select
|
||||
from sqlalchemy.exc import IntegrityError
|
||||
from sqlalchemy.orm import DeclarativeBase
|
||||
|
||||
if TYPE_CHECKING:
|
||||
from .db import Database
|
||||
|
||||
|
||||
class Base(DeclarativeBase):
|
||||
db: Database | None = None
|
||||
|
||||
@classmethod
|
||||
async def add(cls, stops: Self | Iterable[Self]) -> bool:
|
||||
try:
|
||||
if isinstance(stops, Iterable):
|
||||
cls.db.session.add_all(stops) # type: ignore
|
||||
else:
|
||||
cls.db.session.add(stops) # type: ignore
|
||||
await cls.db.session.commit() # type: ignore
|
||||
except (AttributeError, IntegrityError) as err:
|
||||
print(err)
|
||||
return False
|
||||
|
||||
return True
|
||||
|
||||
@classmethod
|
||||
async def get_by_id(cls, id_: int | str) -> Self | None:
|
||||
try:
|
||||
stmt = select(cls).where(cls.id == id_) # type: ignore
|
||||
res = await cls.db.session.execute(stmt) # type: ignore
|
||||
element = res.scalar_one_or_none()
|
||||
except AttributeError as err:
|
||||
print(err)
|
||||
element = None
|
||||
return element
|
@@ -1,48 +0,0 @@
|
||||
from sqlalchemy import text
|
||||
from sqlalchemy.ext.asyncio import (
|
||||
async_sessionmaker,
|
||||
AsyncEngine,
|
||||
AsyncSession,
|
||||
create_async_engine,
|
||||
)
|
||||
|
||||
from .base_class import Base
|
||||
|
||||
|
||||
class Database:
|
||||
def __init__(self) -> None:
|
||||
self._engine: AsyncEngine | None = None
|
||||
self._session_maker: async_sessionmaker[AsyncSession] | None = None
|
||||
self._session: AsyncSession | None = None
|
||||
|
||||
@property
|
||||
def session(self) -> AsyncSession | None:
|
||||
if self._session is None and (session_maker := self._session_maker) is not None:
|
||||
self._session = session_maker()
|
||||
return self._session
|
||||
|
||||
async def connect(self, db_path: str, clear_static_data: bool = False) -> bool:
|
||||
|
||||
# TODO: Preserve UserLastStopSearchResults table from drop.
|
||||
self._engine = create_async_engine(db_path)
|
||||
if self._engine is not None:
|
||||
self._session_maker = async_sessionmaker(
|
||||
self._engine, expire_on_commit=False, class_=AsyncSession
|
||||
)
|
||||
if (session := self.session) is not None:
|
||||
await session.execute(text("CREATE EXTENSION IF NOT EXISTS pg_trgm;"))
|
||||
|
||||
async with self._engine.begin() as conn:
|
||||
if clear_static_data:
|
||||
await conn.run_sync(Base.metadata.drop_all)
|
||||
await conn.run_sync(Base.metadata.create_all)
|
||||
|
||||
return True
|
||||
|
||||
async def disconnect(self) -> None:
|
||||
if self._session is not None:
|
||||
await self._session.close()
|
||||
self._session = None
|
||||
|
||||
if self._engine is not None:
|
||||
await self._engine.dispose()
|
@@ -1,588 +0,0 @@
|
||||
from collections import defaultdict
|
||||
from re import compile as re_compile
|
||||
from time import time
|
||||
from typing import (
|
||||
AsyncIterator,
|
||||
ByteString,
|
||||
Callable,
|
||||
Iterable,
|
||||
List,
|
||||
Type,
|
||||
)
|
||||
|
||||
from aiofiles import open as async_open
|
||||
from aiohttp import ClientSession
|
||||
from msgspec import ValidationError
|
||||
from msgspec.json import Decoder
|
||||
from pyproj import Transformer
|
||||
from shapefile import Reader as ShapeFileReader, ShapeRecord
|
||||
|
||||
from ..db import Database
|
||||
from ..models import ConnectionArea, Line, LinePicto, Stop, StopArea, StopShape
|
||||
from .idfm_types import (
|
||||
ConnectionArea as IdfmConnectionArea,
|
||||
Destinations as IdfmDestinations,
|
||||
IdfmLineState,
|
||||
IdfmResponse,
|
||||
Line as IdfmLine,
|
||||
LinePicto as IdfmPicto,
|
||||
IdfmState,
|
||||
Stop as IdfmStop,
|
||||
StopArea as IdfmStopArea,
|
||||
StopAreaStopAssociation,
|
||||
StopAreaType,
|
||||
StopLineAsso as IdfmStopLineAsso,
|
||||
TransportMode,
|
||||
)
|
||||
from .ratp_types import Picto as RatpPicto
|
||||
|
||||
|
||||
class IdfmInterface:
|
||||
|
||||
IDFM_ROOT_URL = "https://prim.iledefrance-mobilites.fr/marketplace"
|
||||
IDFM_STOP_MON_URL = f"{IDFM_ROOT_URL}/stop-monitoring"
|
||||
|
||||
IDFM_ROOT_URL = "https://data.iledefrance-mobilites.fr/explore/dataset"
|
||||
IDFM_STOPS_URL = (
|
||||
f"{IDFM_ROOT_URL}/arrets/download/?format=json&timezone=Europe/Berlin"
|
||||
)
|
||||
IDFM_PICTO_URL = f"{IDFM_ROOT_URL}/referentiel-des-lignes/files"
|
||||
|
||||
RATP_ROOT_URL = "https://data.ratp.fr/explore/dataset"
|
||||
RATP_PICTO_URL = (
|
||||
f"{RATP_ROOT_URL}"
|
||||
"/pictogrammes-des-lignes-de-metro-rer-tramway-bus-et-noctilien/files"
|
||||
)
|
||||
|
||||
OPERATOR_RE = re_compile(r"[^:]+:Operator::([^:]+):")
|
||||
LINE_RE = re_compile(r"[^:]+:Line::([^:]+):")
|
||||
|
||||
def __init__(self, api_key: str, database: Database) -> None:
|
||||
self._api_key = api_key
|
||||
self._database = database
|
||||
|
||||
self._http_headers = {"Accept": "application/json", "apikey": self._api_key}
|
||||
|
||||
self._epsg2154_epsg3857_transformer = Transformer.from_crs(2154, 3857)
|
||||
|
||||
self._json_stops_decoder = Decoder(type=List[IdfmStop])
|
||||
self._json_stop_areas_decoder = Decoder(type=List[IdfmStopArea])
|
||||
self._json_connection_areas_decoder = Decoder(type=List[IdfmConnectionArea])
|
||||
self._json_lines_decoder = Decoder(type=List[IdfmLine])
|
||||
self._json_stops_lines_assos_decoder = Decoder(type=List[IdfmStopLineAsso])
|
||||
self._json_ratp_pictos_decoder = Decoder(type=List[RatpPicto])
|
||||
self._json_stop_area_stop_asso_decoder = Decoder(
|
||||
type=List[StopAreaStopAssociation]
|
||||
)
|
||||
|
||||
self._response_json_decoder = Decoder(type=IdfmResponse)
|
||||
|
||||
async def startup(self) -> None:
|
||||
BATCH_SIZE = 10000
|
||||
STEPS: tuple[
|
||||
tuple[
|
||||
Type[ConnectionArea] | Type[Stop] | Type[StopArea] | Type[StopShape],
|
||||
Callable,
|
||||
Callable,
|
||||
],
|
||||
...,
|
||||
] = (
|
||||
(
|
||||
StopShape,
|
||||
self._request_stop_shapes,
|
||||
self._format_idfm_stop_shapes,
|
||||
),
|
||||
(
|
||||
ConnectionArea,
|
||||
self._request_idfm_connection_areas,
|
||||
self._format_idfm_connection_areas,
|
||||
),
|
||||
(
|
||||
StopArea,
|
||||
self._request_idfm_stop_areas,
|
||||
self._format_idfm_stop_areas,
|
||||
),
|
||||
(Stop, self._request_idfm_stops, self._format_idfm_stops),
|
||||
)
|
||||
|
||||
for model, get_method, format_method in STEPS:
|
||||
step_begin_ts = time()
|
||||
elements = []
|
||||
|
||||
async for element in get_method():
|
||||
elements.append(element)
|
||||
|
||||
if len(elements) == BATCH_SIZE:
|
||||
await model.add(format_method(*elements))
|
||||
elements.clear()
|
||||
|
||||
if elements:
|
||||
await model.add(format_method(*elements))
|
||||
|
||||
print(f"Add {model.__name__}s: {time() - step_begin_ts}s")
|
||||
|
||||
begin_ts = time()
|
||||
await self._load_lines()
|
||||
print(f"Add Lines and IDFM LinePictos: {time() - begin_ts}s")
|
||||
|
||||
begin_ts = time()
|
||||
await self._load_ratp_pictos(30)
|
||||
print(f"Add RATP LinePictos: {time() - begin_ts}s")
|
||||
|
||||
begin_ts = time()
|
||||
await self._load_lines_stops_assos()
|
||||
print(f"Link Stops to Lines: {time() - begin_ts}s")
|
||||
|
||||
begin_ts = time()
|
||||
await self._load_stop_assos()
|
||||
print(f"Link Stops to StopAreas: {time() - begin_ts}s")
|
||||
|
||||
async def _load_lines(self, batch_size: int = 5000) -> None:
|
||||
lines, pictos = [], []
|
||||
picto_ids = set()
|
||||
async for line in self._request_idfm_lines():
|
||||
if (picto := line.fields.picto) is not None and picto.id_ not in picto_ids:
|
||||
picto_ids.add(picto.id_)
|
||||
pictos.append(picto)
|
||||
|
||||
lines.append(line)
|
||||
if len(lines) == batch_size:
|
||||
await LinePicto.add(IdfmInterface._format_idfm_pictos(*pictos))
|
||||
await Line.add(await self._format_idfm_lines(*lines))
|
||||
lines.clear()
|
||||
pictos.clear()
|
||||
|
||||
if pictos:
|
||||
await LinePicto.add(IdfmInterface._format_idfm_pictos(*pictos))
|
||||
if lines:
|
||||
await Line.add(await self._format_idfm_lines(*lines))
|
||||
|
||||
async def _load_ratp_pictos(self, batch_size: int = 5) -> None:
|
||||
pictos = []
|
||||
|
||||
async for picto in self._request_ratp_pictos():
|
||||
pictos.append(picto)
|
||||
if len(pictos) == batch_size:
|
||||
formatted_pictos = IdfmInterface._format_ratp_pictos(*pictos)
|
||||
await LinePicto.add(map(lambda picto: picto[1], formatted_pictos))
|
||||
await Line.add_pictos(formatted_pictos)
|
||||
pictos.clear()
|
||||
|
||||
if pictos:
|
||||
formatted_pictos = IdfmInterface._format_ratp_pictos(*pictos)
|
||||
await LinePicto.add(map(lambda picto: picto[1], formatted_pictos))
|
||||
await Line.add_pictos(formatted_pictos)
|
||||
|
||||
async def _load_lines_stops_assos(self, batch_size: int = 5000) -> None:
|
||||
total_assos_nb = total_found_nb = 0
|
||||
assos = []
|
||||
async for asso in self._request_idfm_stops_lines_associations():
|
||||
fields = asso.fields
|
||||
try:
|
||||
stop_id = int(fields.stop_id.rsplit(":", 1)[-1])
|
||||
except ValueError as err:
|
||||
print(err)
|
||||
print(f"{fields.stop_id = }")
|
||||
continue
|
||||
|
||||
assos.append((fields.route_long_name, fields.operatorname, stop_id))
|
||||
if len(assos) == batch_size:
|
||||
total_assos_nb += batch_size
|
||||
total_found_nb += await Line.add_stops(assos)
|
||||
assos.clear()
|
||||
|
||||
if assos:
|
||||
total_assos_nb += len(assos)
|
||||
total_found_nb += await Line.add_stops(assos)
|
||||
|
||||
print(f"{total_found_nb} line <-> stop ({total_assos_nb = } found)")
|
||||
|
||||
async def _load_stop_assos(self, batch_size: int = 5000) -> None:
|
||||
total_assos_nb = area_stop_assos_nb = conn_stop_assos_nb = 0
|
||||
area_stop_assos = []
|
||||
connection_stop_assos = []
|
||||
|
||||
async for asso in self._request_idfm_stop_area_stop_associations():
|
||||
fields = asso.fields
|
||||
|
||||
stop_id = int(fields.arrid)
|
||||
|
||||
area_stop_assos.append((int(fields.zdaid), stop_id))
|
||||
connection_stop_assos.append((int(fields.zdcid), stop_id))
|
||||
|
||||
if len(area_stop_assos) == batch_size:
|
||||
total_assos_nb += batch_size
|
||||
|
||||
if (found_nb := await StopArea.add_stops(area_stop_assos)) is not None:
|
||||
area_stop_assos_nb += found_nb
|
||||
area_stop_assos.clear()
|
||||
|
||||
if (
|
||||
found_nb := await ConnectionArea.add_stops(connection_stop_assos)
|
||||
) is not None:
|
||||
conn_stop_assos_nb += found_nb
|
||||
connection_stop_assos.clear()
|
||||
|
||||
if area_stop_assos:
|
||||
total_assos_nb += len(area_stop_assos)
|
||||
if (found_nb := await StopArea.add_stops(area_stop_assos)) is not None:
|
||||
area_stop_assos_nb += found_nb
|
||||
if (
|
||||
found_nb := await ConnectionArea.add_stops(connection_stop_assos)
|
||||
) is not None:
|
||||
conn_stop_assos_nb += found_nb
|
||||
|
||||
print(f"{area_stop_assos_nb} stop area <-> stop ({total_assos_nb = } found)")
|
||||
print(f"{conn_stop_assos_nb} stop area <-> stop ({total_assos_nb = } found)")
|
||||
|
||||
# TODO: This method is synchronous due to the shapefile library.
|
||||
# It's not a blocking issue but it could be nice to find an alternative.
|
||||
async def _request_stop_shapes(self) -> AsyncIterator[ShapeRecord]:
|
||||
# TODO: Use HTTP
|
||||
with ShapeFileReader("./tests/datasets/REF_LDA.zip") as reader:
|
||||
for record in reader.shapeRecords():
|
||||
yield record
|
||||
|
||||
async def _request_idfm_stops(self) -> AsyncIterator[IdfmStop]:
|
||||
# headers = {"Accept": "application/json", "apikey": self._api_key}
|
||||
# async with ClientSession(headers=headers) as session:
|
||||
# async with session.get(self.STOPS_URL) as response:
|
||||
# # print("Status:", response.status)
|
||||
# if response.status == 200:
|
||||
# for point in self._json_stops_decoder.decode(await response.read()):
|
||||
# yield point
|
||||
# TODO: Use HTTP
|
||||
async with async_open("./tests/datasets/stops_dataset.json", "rb") as raw:
|
||||
for element in self._json_stops_decoder.decode(await raw.read()):
|
||||
yield element
|
||||
|
||||
async def _request_idfm_stop_areas(self) -> AsyncIterator[IdfmStopArea]:
|
||||
# TODO: Use HTTP
|
||||
async with async_open("./tests/datasets/zones-d-arrets.json", "rb") as raw:
|
||||
for element in self._json_stop_areas_decoder.decode(await raw.read()):
|
||||
yield element
|
||||
|
||||
async def _request_idfm_connection_areas(self) -> AsyncIterator[IdfmConnectionArea]:
|
||||
async with async_open(
|
||||
"./tests/datasets/zones-de-correspondance.json", "rb"
|
||||
) as raw:
|
||||
for element in self._json_connection_areas_decoder.decode(await raw.read()):
|
||||
yield element
|
||||
|
||||
async def _request_idfm_lines(self) -> AsyncIterator[IdfmLine]:
|
||||
# TODO: Use HTTP
|
||||
async with async_open("./tests/datasets/lines_dataset.json", "rb") as raw:
|
||||
for element in self._json_lines_decoder.decode(await raw.read()):
|
||||
yield element
|
||||
|
||||
async def _request_idfm_stops_lines_associations(
|
||||
self,
|
||||
) -> AsyncIterator[IdfmStopLineAsso]:
|
||||
# TODO: Use HTTP
|
||||
async with async_open("./tests/datasets/arrets-lignes.json", "rb") as raw:
|
||||
for element in self._json_stops_lines_assos_decoder.decode(
|
||||
await raw.read()
|
||||
):
|
||||
yield element
|
||||
|
||||
async def _request_idfm_stop_area_stop_associations(
|
||||
self,
|
||||
) -> AsyncIterator[StopAreaStopAssociation]:
|
||||
# TODO: Use HTTP
|
||||
async with async_open("./tests/datasets/relations.json", "rb") as raw:
|
||||
for element in self._json_stop_area_stop_asso_decoder.decode(
|
||||
await raw.read()
|
||||
):
|
||||
yield element
|
||||
|
||||
async def _request_ratp_pictos(self) -> AsyncIterator[RatpPicto]:
|
||||
# TODO: Use HTTP
|
||||
async with async_open(
|
||||
"./tests/datasets/pictogrammes-des-lignes-de-metro-rer-tramway-bus-et-noctilien.json",
|
||||
"rb",
|
||||
) as fd:
|
||||
for element in self._json_ratp_pictos_decoder.decode(await fd.read()):
|
||||
yield element
|
||||
|
||||
@classmethod
|
||||
def _format_idfm_pictos(cls, *pictos: IdfmPicto) -> Iterable[LinePicto]:
|
||||
ret = []
|
||||
|
||||
for picto in pictos:
|
||||
ret.append(
|
||||
LinePicto(
|
||||
id=picto.id_,
|
||||
mime_type=picto.mimetype,
|
||||
height_px=picto.height,
|
||||
width_px=picto.width,
|
||||
filename=picto.filename,
|
||||
url=f"{cls.IDFM_PICTO_URL}/{picto.id_}/download",
|
||||
thumbnail=picto.thumbnail,
|
||||
format=picto.format,
|
||||
)
|
||||
)
|
||||
|
||||
return ret
|
||||
|
||||
@classmethod
|
||||
def _format_ratp_pictos(cls, *pictos: RatpPicto) -> Iterable[tuple[str, LinePicto]]:
|
||||
ret = []
|
||||
|
||||
for picto in pictos:
|
||||
if (fields := picto.fields.noms_des_fichiers) is not None:
|
||||
ret.append(
|
||||
(
|
||||
picto.fields.indices_commerciaux,
|
||||
LinePicto(
|
||||
id=fields.id_,
|
||||
mime_type=f"image/{fields.format.lower()}",
|
||||
height_px=fields.height,
|
||||
width_px=fields.width,
|
||||
filename=fields.filename,
|
||||
url=f"{cls.RATP_PICTO_URL}/{fields.id_}/download",
|
||||
thumbnail=fields.thumbnail,
|
||||
format=fields.format,
|
||||
),
|
||||
)
|
||||
)
|
||||
|
||||
return ret
|
||||
|
||||
async def _format_idfm_lines(self, *lines: IdfmLine) -> Iterable[Line]:
|
||||
ret = []
|
||||
|
||||
optional_value = IdfmLine.optional_value
|
||||
|
||||
for line in lines:
|
||||
fields = line.fields
|
||||
|
||||
picto_id = fields.picto.id_ if fields.picto is not None else None
|
||||
picto = await LinePicto.get_by_id(picto_id) if picto_id else None
|
||||
|
||||
ret.append(
|
||||
Line(
|
||||
id=fields.id_line,
|
||||
short_name=fields.shortname_line,
|
||||
name=fields.name_line,
|
||||
status=IdfmLineState(fields.status.value),
|
||||
transport_mode=TransportMode(fields.transportmode.value),
|
||||
transport_submode=optional_value(fields.transportsubmode),
|
||||
network_name=optional_value(fields.networkname),
|
||||
group_of_lines_id=optional_value(fields.id_groupoflines),
|
||||
group_of_lines_shortname=optional_value(
|
||||
fields.shortname_groupoflines
|
||||
),
|
||||
colour_web_hexa=fields.colourweb_hexa,
|
||||
text_colour_hexa=fields.textcolourprint_hexa,
|
||||
operator_id=optional_value(fields.operatorref),
|
||||
operator_name=optional_value(fields.operatorname),
|
||||
accessibility=IdfmState(fields.accessibility.value),
|
||||
visual_signs_available=IdfmState(
|
||||
fields.visualsigns_available.value
|
||||
),
|
||||
audible_signs_available=IdfmState(
|
||||
fields.audiblesigns_available.value
|
||||
),
|
||||
picto_id=fields.picto.id_ if fields.picto is not None else None,
|
||||
picto=picto,
|
||||
record_id=line.recordid,
|
||||
record_ts=int(line.record_timestamp.timestamp()),
|
||||
)
|
||||
)
|
||||
|
||||
return ret
|
||||
|
||||
def _format_idfm_stops(self, *stops: IdfmStop) -> Iterable[Stop]:
|
||||
for stop in stops:
|
||||
fields = stop.fields
|
||||
|
||||
try:
|
||||
created_ts = int(fields.arrcreated.timestamp()) # type: ignore
|
||||
except AttributeError:
|
||||
created_ts = None
|
||||
|
||||
epsg3857_point = self._epsg2154_epsg3857_transformer.transform(
|
||||
fields.arrxepsg2154, fields.arryepsg2154
|
||||
)
|
||||
|
||||
yield Stop(
|
||||
id=int(fields.arrid),
|
||||
name=fields.arrname,
|
||||
epsg3857_x=epsg3857_point[0],
|
||||
epsg3857_y=epsg3857_point[1],
|
||||
town_name=fields.arrtown,
|
||||
postal_region=fields.arrpostalregion,
|
||||
transport_mode=TransportMode(fields.arrtype.value),
|
||||
version=fields.arrversion,
|
||||
created_ts=created_ts,
|
||||
changed_ts=int(fields.arrchanged.timestamp()),
|
||||
accessibility=IdfmState(fields.arraccessibility.value),
|
||||
visual_signs_available=IdfmState(fields.arrvisualsigns.value),
|
||||
audible_signs_available=IdfmState(fields.arraudiblesignals.value),
|
||||
record_id=stop.recordid,
|
||||
record_ts=int(stop.record_timestamp.timestamp()),
|
||||
)
|
||||
|
||||
def _format_idfm_stop_areas(self, *stop_areas: IdfmStopArea) -> Iterable[StopArea]:
|
||||
for stop_area in stop_areas:
|
||||
fields = stop_area.fields
|
||||
|
||||
try:
|
||||
created_ts = int(fields.zdacreated.timestamp()) # type: ignore
|
||||
except AttributeError:
|
||||
created_ts = None
|
||||
|
||||
epsg3857_point = self._epsg2154_epsg3857_transformer.transform(
|
||||
fields.zdaxepsg2154, fields.zdayepsg2154
|
||||
)
|
||||
|
||||
yield StopArea(
|
||||
id=int(fields.zdaid),
|
||||
name=fields.zdaname,
|
||||
town_name=fields.zdatown,
|
||||
postal_region=fields.zdapostalregion,
|
||||
epsg3857_x=epsg3857_point[0],
|
||||
epsg3857_y=epsg3857_point[1],
|
||||
type=StopAreaType(fields.zdatype.value),
|
||||
version=fields.zdaversion,
|
||||
created_ts=created_ts,
|
||||
changed_ts=int(fields.zdachanged.timestamp()),
|
||||
)
|
||||
|
||||
def _format_idfm_connection_areas(
|
||||
self,
|
||||
*connection_areas: IdfmConnectionArea,
|
||||
) -> Iterable[ConnectionArea]:
|
||||
for connection_area in connection_areas:
|
||||
|
||||
epsg3857_point = self._epsg2154_epsg3857_transformer.transform(
|
||||
connection_area.zdcxepsg2154, connection_area.zdcyepsg2154
|
||||
)
|
||||
|
||||
yield ConnectionArea(
|
||||
id=int(connection_area.zdcid),
|
||||
name=connection_area.zdcname,
|
||||
town_name=connection_area.zdctown,
|
||||
postal_region=connection_area.zdcpostalregion,
|
||||
epsg3857_x=epsg3857_point[0],
|
||||
epsg3857_y=epsg3857_point[1],
|
||||
transport_mode=StopAreaType(connection_area.zdctype.value),
|
||||
version=connection_area.zdcversion,
|
||||
created_ts=int(connection_area.zdccreated.timestamp()),
|
||||
changed_ts=int(connection_area.zdcchanged.timestamp()),
|
||||
)
|
||||
|
||||
def _format_idfm_stop_shapes(
|
||||
self, *shape_records: ShapeRecord
|
||||
) -> Iterable[StopShape]:
|
||||
for shape_record in shape_records:
|
||||
|
||||
epsg3857_points = [
|
||||
self._epsg2154_epsg3857_transformer.transform(*point)
|
||||
for point in shape_record.shape.points
|
||||
]
|
||||
|
||||
bbox_it = iter(shape_record.shape.bbox)
|
||||
epsg3857_bbox = [
|
||||
self._epsg2154_epsg3857_transformer.transform(*point)
|
||||
for point in zip(bbox_it, bbox_it)
|
||||
]
|
||||
|
||||
yield StopShape(
|
||||
id=shape_record.record[1],
|
||||
type=shape_record.shape.shapeType,
|
||||
epsg3857_bbox=epsg3857_bbox,
|
||||
epsg3857_points=epsg3857_points,
|
||||
)
|
||||
|
||||
async def render_line_picto(self, line: Line) -> tuple[None | str, None | str]:
|
||||
begin_ts = time()
|
||||
line_picto_path = line_picto_format = None
|
||||
target = f"/tmp/{line.id}_repr"
|
||||
|
||||
picto = line.picto
|
||||
if picto is not None:
|
||||
if (picto_data := await self._get_line_picto(line)) is not None:
|
||||
async with async_open(target, "wb") as fd:
|
||||
await fd.write(bytes(picto_data))
|
||||
line_picto_path = target
|
||||
line_picto_format = picto.mime_type
|
||||
|
||||
print(f"render_line_picto: {time() - begin_ts}")
|
||||
return (line_picto_path, line_picto_format)
|
||||
|
||||
async def _get_line_picto(self, line: Line) -> ByteString | None:
|
||||
print("---------------------------------------------------------------------")
|
||||
begin_ts = time()
|
||||
data = None
|
||||
|
||||
picto = line.picto
|
||||
if picto is not None and picto.url is not None:
|
||||
headers = (
|
||||
self._http_headers if picto.url.startswith(self.IDFM_ROOT_URL) else None
|
||||
)
|
||||
session_begin_ts = time()
|
||||
async with ClientSession(headers=headers) as session:
|
||||
session_creation_ts = time()
|
||||
print(f"Session creation {session_creation_ts - session_begin_ts}")
|
||||
async with session.get(picto.url) as response:
|
||||
get_end_ts = time()
|
||||
print(f"GET {get_end_ts - session_creation_ts}")
|
||||
data = await response.read()
|
||||
print(f"read {time() - get_end_ts}")
|
||||
|
||||
print(f"render_line_picto: {time() - begin_ts}")
|
||||
print("---------------------------------------------------------------------")
|
||||
return data
|
||||
|
||||
async def get_next_passages(self, stop_point_id: int) -> IdfmResponse | None:
|
||||
ret = None
|
||||
params = {"MonitoringRef": f"STIF:StopPoint:Q:{stop_point_id}:"}
|
||||
async with ClientSession(headers=self._http_headers) as session:
|
||||
async with session.get(self.IDFM_STOP_MON_URL, params=params) as response:
|
||||
if response.status == 200:
|
||||
data = await response.read()
|
||||
try:
|
||||
ret = self._response_json_decoder.decode(data)
|
||||
except ValidationError as err:
|
||||
print(err)
|
||||
|
||||
return ret
|
||||
|
||||
async def get_destinations(self, stop_id: int) -> IdfmDestinations | None:
|
||||
begin_ts = time()
|
||||
|
||||
destinations: IdfmDestinations = defaultdict(set)
|
||||
|
||||
if (stop := await Stop.get_by_id(stop_id)) is not None:
|
||||
expected_stop_ids = {stop.id}
|
||||
|
||||
elif (stop_area := await StopArea.get_by_id(stop_id)) is not None:
|
||||
expected_stop_ids = {stop.id for stop in stop_area.stops}
|
||||
|
||||
else:
|
||||
return None
|
||||
|
||||
if (res := await self.get_next_passages(stop_id)) is not None:
|
||||
|
||||
for delivery in res.Siri.ServiceDelivery.StopMonitoringDelivery:
|
||||
if delivery.Status == IdfmState.true:
|
||||
for stop_visit in delivery.MonitoredStopVisit:
|
||||
|
||||
monitoring_ref = stop_visit.MonitoringRef.value
|
||||
try:
|
||||
monitored_stop_id = int(monitoring_ref.split(":")[-2])
|
||||
except (IndexError, ValueError):
|
||||
print(f"Unable to get stop id from {monitoring_ref}")
|
||||
continue
|
||||
|
||||
journey = stop_visit.MonitoredVehicleJourney
|
||||
if (
|
||||
dst_names := journey.DestinationName
|
||||
) and monitored_stop_id in expected_stop_ids:
|
||||
|
||||
line_id = journey.LineRef.value.split(":")[-2]
|
||||
destinations[line_id].add(dst_names[0].value)
|
||||
|
||||
print(f"get_next_passages: {time() - begin_ts}")
|
||||
return destinations
|
@@ -1,25 +0,0 @@
|
||||
from datetime import datetime
|
||||
from typing import Optional
|
||||
|
||||
from msgspec import Struct
|
||||
|
||||
|
||||
class PictoFieldsFile(Struct, rename={"id_": "id"}):
|
||||
id_: str
|
||||
height: int
|
||||
width: int
|
||||
filename: str
|
||||
thumbnail: bool
|
||||
format: str
|
||||
|
||||
|
||||
class PictoFields(Struct):
|
||||
indices_commerciaux: str
|
||||
noms_des_fichiers: Optional[PictoFieldsFile] = None
|
||||
|
||||
|
||||
class Picto(Struct):
|
||||
datasetid: str
|
||||
recordid: str
|
||||
fields: PictoFields
|
||||
record_timestamp: datetime
|
@@ -1,258 +0,0 @@
|
||||
from __future__ import annotations
|
||||
|
||||
from typing import Iterable, Sequence, TYPE_CHECKING
|
||||
|
||||
from sqlalchemy import (
|
||||
BigInteger,
|
||||
Column,
|
||||
Enum,
|
||||
Float,
|
||||
ForeignKey,
|
||||
Integer,
|
||||
JSON,
|
||||
select,
|
||||
String,
|
||||
Table,
|
||||
)
|
||||
from sqlalchemy.orm import (
|
||||
mapped_column,
|
||||
Mapped,
|
||||
relationship,
|
||||
selectinload,
|
||||
with_polymorphic,
|
||||
)
|
||||
from sqlalchemy.schema import Index
|
||||
|
||||
from ..db import Base, db
|
||||
from ..idfm_interface.idfm_types import TransportMode, IdfmState, StopAreaType
|
||||
|
||||
if TYPE_CHECKING:
|
||||
from .line import Line
|
||||
|
||||
|
||||
stop_area_stop_association_table = Table(
|
||||
"stop_area_stop_association_table",
|
||||
Base.metadata,
|
||||
Column("stop_id", ForeignKey("_stops.id")),
|
||||
Column("stop_area_id", ForeignKey("stop_areas.id")),
|
||||
)
|
||||
|
||||
|
||||
class _Stop(Base):
|
||||
|
||||
db = db
|
||||
|
||||
id = mapped_column(BigInteger, primary_key=True)
|
||||
kind = mapped_column(String)
|
||||
|
||||
name = mapped_column(String, nullable=False, index=True)
|
||||
town_name = mapped_column(String, nullable=False)
|
||||
postal_region = mapped_column(String, nullable=False)
|
||||
epsg3857_x = mapped_column(Float, nullable=False)
|
||||
epsg3857_y = mapped_column(Float, nullable=False)
|
||||
|
||||
version = mapped_column(String, nullable=False)
|
||||
created_ts = mapped_column(BigInteger)
|
||||
changed_ts = mapped_column(BigInteger, nullable=False)
|
||||
|
||||
lines: Mapped[list[Line]] = relationship(
|
||||
"Line",
|
||||
secondary="line_stop_association_table",
|
||||
back_populates="stops",
|
||||
lazy="selectin",
|
||||
)
|
||||
areas: Mapped[list["StopArea"]] = relationship(
|
||||
"StopArea", secondary=stop_area_stop_association_table, back_populates="stops"
|
||||
)
|
||||
connection_area_id: Mapped[int] = mapped_column(
|
||||
ForeignKey("connection_areas.id"), nullable=True
|
||||
)
|
||||
connection_area: Mapped["ConnectionArea"] = relationship(
|
||||
back_populates="stops", lazy="selectin"
|
||||
)
|
||||
|
||||
__tablename__ = "_stops"
|
||||
__mapper_args__ = {"polymorphic_identity": "_stops", "polymorphic_on": kind}
|
||||
__table_args__ = (
|
||||
# To optimize the ilike requests
|
||||
Index(
|
||||
"name_idx_gin",
|
||||
name,
|
||||
postgresql_ops={"name": "gin_trgm_ops"},
|
||||
postgresql_using="gin",
|
||||
),
|
||||
)
|
||||
|
||||
# TODO: Test https://www.cybertec-postgresql.com/en/postgresql-more-performance-for-like-and-ilike-statements/
|
||||
# TODO: Should be able to remove with_polymorphic ?
|
||||
@classmethod
|
||||
async def get_by_name(cls, name: str) -> Sequence[type[_Stop]] | None:
|
||||
session = cls.db.session
|
||||
if session is None:
|
||||
return None
|
||||
|
||||
stop_stop_area = with_polymorphic(_Stop, [Stop, StopArea])
|
||||
stmt = (
|
||||
select(stop_stop_area)
|
||||
.where(stop_stop_area.name.ilike(f"%{name}%"))
|
||||
.options(
|
||||
selectinload(stop_stop_area.areas),
|
||||
selectinload(stop_stop_area.lines),
|
||||
)
|
||||
)
|
||||
|
||||
res = await session.execute(stmt)
|
||||
stops = res.scalars().all()
|
||||
|
||||
return stops
|
||||
|
||||
|
||||
class Stop(_Stop):
|
||||
|
||||
id = mapped_column(BigInteger, ForeignKey("_stops.id"), primary_key=True)
|
||||
|
||||
transport_mode = mapped_column(Enum(TransportMode), nullable=False)
|
||||
accessibility = mapped_column(Enum(IdfmState), nullable=False)
|
||||
visual_signs_available = mapped_column(Enum(IdfmState), nullable=False)
|
||||
audible_signs_available = mapped_column(Enum(IdfmState), nullable=False)
|
||||
|
||||
record_id = mapped_column(String, nullable=False)
|
||||
record_ts = mapped_column(BigInteger, nullable=False)
|
||||
|
||||
__tablename__ = "stops"
|
||||
__mapper_args__ = {"polymorphic_identity": "stops", "polymorphic_load": "inline"}
|
||||
|
||||
|
||||
class StopArea(_Stop):
|
||||
|
||||
id = mapped_column(BigInteger, ForeignKey("_stops.id"), primary_key=True)
|
||||
|
||||
type = mapped_column(Enum(StopAreaType), nullable=False)
|
||||
|
||||
stops: Mapped[list["Stop"]] = relationship(
|
||||
"Stop",
|
||||
secondary=stop_area_stop_association_table,
|
||||
back_populates="areas",
|
||||
lazy="selectin",
|
||||
)
|
||||
|
||||
__tablename__ = "stop_areas"
|
||||
__mapper_args__ = {
|
||||
"polymorphic_identity": "stop_areas",
|
||||
"polymorphic_load": "inline",
|
||||
}
|
||||
|
||||
@classmethod
|
||||
async def add_stops(
|
||||
cls, stop_area_to_stop_ids: Iterable[tuple[int, int]]
|
||||
) -> int | None:
|
||||
session = cls.db.session
|
||||
if session is None:
|
||||
return None
|
||||
|
||||
stop_area_ids, stop_ids = set(), set()
|
||||
for stop_area_id, stop_id in stop_area_to_stop_ids:
|
||||
stop_area_ids.add(stop_area_id)
|
||||
stop_ids.add(stop_id)
|
||||
|
||||
stop_areas_res = await session.scalars(
|
||||
select(StopArea)
|
||||
.where(StopArea.id.in_(stop_area_ids))
|
||||
.options(selectinload(StopArea.stops))
|
||||
)
|
||||
stop_areas: dict[int, StopArea] = {
|
||||
stop_area.id: stop_area for stop_area in stop_areas_res.all()
|
||||
}
|
||||
|
||||
stop_res = await session.execute(select(Stop).where(Stop.id.in_(stop_ids)))
|
||||
stops: dict[int, Stop] = {stop.id: stop for stop in stop_res.scalars()}
|
||||
|
||||
found = 0
|
||||
for stop_area_id, stop_id in stop_area_to_stop_ids:
|
||||
if (stop_area := stop_areas.get(stop_area_id)) is not None:
|
||||
if (stop := stops.get(stop_id)) is not None:
|
||||
stop_area.stops.append(stop)
|
||||
found += 1
|
||||
else:
|
||||
print(f"No stop found for {stop_id} id")
|
||||
else:
|
||||
print(f"No stop area found for {stop_area_id}")
|
||||
|
||||
await session.commit()
|
||||
|
||||
return found
|
||||
|
||||
|
||||
class StopShape(Base):
|
||||
|
||||
db = db
|
||||
|
||||
id = mapped_column(BigInteger, primary_key=True) # Same id than ConnectionArea
|
||||
type = mapped_column(Integer, nullable=False)
|
||||
epsg3857_bbox = mapped_column(JSON)
|
||||
epsg3857_points = mapped_column(JSON)
|
||||
|
||||
__tablename__ = "stop_shapes"
|
||||
|
||||
|
||||
class ConnectionArea(Base):
|
||||
|
||||
db = db
|
||||
|
||||
id = mapped_column(BigInteger, primary_key=True)
|
||||
|
||||
name = mapped_column(String, nullable=False)
|
||||
town_name = mapped_column(String, nullable=False)
|
||||
postal_region = mapped_column(String, nullable=False)
|
||||
epsg3857_x = mapped_column(Float, nullable=False)
|
||||
epsg3857_y = mapped_column(Float, nullable=False)
|
||||
transport_mode = mapped_column(Enum(StopAreaType), nullable=False)
|
||||
|
||||
version = mapped_column(String, nullable=False)
|
||||
created_ts = mapped_column(BigInteger)
|
||||
changed_ts = mapped_column(BigInteger, nullable=False)
|
||||
|
||||
stops: Mapped[list["_Stop"]] = relationship(back_populates="connection_area")
|
||||
|
||||
__tablename__ = "connection_areas"
|
||||
|
||||
# TODO: Merge with StopArea.add_stops
|
||||
@classmethod
|
||||
async def add_stops(
|
||||
cls, conn_area_to_stop_ids: Iterable[tuple[int, int]]
|
||||
) -> int | None:
|
||||
session = cls.db.session
|
||||
if session is None:
|
||||
return None
|
||||
|
||||
conn_area_ids, stop_ids = set(), set()
|
||||
for conn_area_id, stop_id in conn_area_to_stop_ids:
|
||||
conn_area_ids.add(conn_area_id)
|
||||
stop_ids.add(stop_id)
|
||||
|
||||
conn_area_res = await session.execute(
|
||||
select(ConnectionArea)
|
||||
.where(ConnectionArea.id.in_(conn_area_ids))
|
||||
.options(selectinload(ConnectionArea.stops))
|
||||
)
|
||||
conn_areas: dict[int, ConnectionArea] = {
|
||||
conn.id: conn for conn in conn_area_res.scalars()
|
||||
}
|
||||
|
||||
stop_res = await session.execute(select(_Stop).where(_Stop.id.in_(stop_ids)))
|
||||
stops: dict[int, _Stop] = {stop.id: stop for stop in stop_res.scalars()}
|
||||
|
||||
found = 0
|
||||
for conn_area_id, stop_id in conn_area_to_stop_ids:
|
||||
if (conn_area := conn_areas.get(conn_area_id)) is not None:
|
||||
if (stop := stops.get(stop_id)) is not None:
|
||||
conn_area.stops.append(stop)
|
||||
found += 1
|
||||
else:
|
||||
print(f"No stop found for {stop_id} id")
|
||||
else:
|
||||
print(f"No connection area found for {conn_area_id}")
|
||||
|
||||
await session.commit()
|
||||
|
||||
return found
|
@@ -1,25 +0,0 @@
|
||||
from sqlalchemy import Column, ForeignKey, String, Table
|
||||
from sqlalchemy.orm import Mapped, mapped_column, relationship
|
||||
|
||||
from ..db import Base, db
|
||||
from .stop import _Stop
|
||||
|
||||
user_last_stop_search_stops_associations_table = Table(
|
||||
"user_last_stop_search_stops_associations_table",
|
||||
Base.metadata,
|
||||
Column("user_mxid", ForeignKey("user_last_stop_search_results.user_mxid")),
|
||||
Column("stop_id", ForeignKey("_stops.id")),
|
||||
)
|
||||
|
||||
|
||||
class UserLastStopSearchResults(Base):
|
||||
|
||||
db = db
|
||||
|
||||
__tablename__ = "user_last_stop_search_results"
|
||||
|
||||
user_mxid = mapped_column(String, primary_key=True)
|
||||
request_content = mapped_column(String, nullable=False)
|
||||
stops: Mapped[_Stop] = relationship(
|
||||
_Stop, secondary=user_last_stop_search_stops_associations_table
|
||||
)
|
21
backend/config.local.yaml
Normal file
21
backend/config.local.yaml
Normal file
@@ -0,0 +1,21 @@
|
||||
app_name: carrramba-encore-rate
|
||||
clear_static_data: false
|
||||
|
||||
http:
|
||||
host: 127.0.0.1
|
||||
port: 8080
|
||||
cert: ./config/cert.pem
|
||||
|
||||
db:
|
||||
name: carrramba-encore-rate
|
||||
host: 127.0.0.1
|
||||
port: 5432
|
||||
driver: postgresql+psycopg
|
||||
user: cer
|
||||
password: cer_password
|
||||
|
||||
cache:
|
||||
enable: true
|
||||
|
||||
tracing:
|
||||
enable: false
|
21
backend/config.sample.yaml
Normal file
21
backend/config.sample.yaml
Normal file
@@ -0,0 +1,21 @@
|
||||
app_name: carrramba-encore-rate
|
||||
clear_static_data: false
|
||||
|
||||
http:
|
||||
host: 0.0.0.0
|
||||
port: 8080
|
||||
# cert: ./config/cert.pem
|
||||
|
||||
db:
|
||||
name: carrramba-encore-rate
|
||||
host: postgres
|
||||
port: 5432
|
||||
user: cer
|
||||
|
||||
cache:
|
||||
enable: true
|
||||
host: redis
|
||||
# TODO: Add user credentials
|
||||
|
||||
tracing:
|
||||
enable: false
|
575
backend/db_updater.py
Executable file
575
backend/db_updater.py
Executable file
@@ -0,0 +1,575 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
from asyncio import run, gather
|
||||
from logging import getLogger, INFO, Handler as LoggingHandler, NOTSET
|
||||
from itertools import islice
|
||||
from time import time
|
||||
from os import environ
|
||||
from typing import Callable, Iterable, List, Type
|
||||
|
||||
from aiofiles.tempfile import NamedTemporaryFile
|
||||
from aiohttp import ClientSession
|
||||
from msgspec import ValidationError
|
||||
from msgspec.json import Decoder
|
||||
from pyproj import Transformer
|
||||
from shapefile import Reader as ShapeFileReader, ShapeRecord # type: ignore
|
||||
from tqdm import tqdm
|
||||
from yaml import safe_load
|
||||
|
||||
from api.db import Base, db, Database
|
||||
from api.models import ConnectionArea, Line, LinePicto, Stop, StopArea, StopShape
|
||||
from api.idfm_interface.idfm_types import (
|
||||
ConnectionArea as IdfmConnectionArea,
|
||||
IdfmLineState,
|
||||
Line as IdfmLine,
|
||||
LinePicto as IdfmPicto,
|
||||
IdfmState,
|
||||
Stop as IdfmStop,
|
||||
StopArea as IdfmStopArea,
|
||||
StopAreaStopAssociation,
|
||||
StopAreaType,
|
||||
StopLineAsso as IdfmStopLineAsso,
|
||||
TransportMode,
|
||||
)
|
||||
from api.idfm_interface.ratp_types import Picto as RatpPicto
|
||||
from api.settings import Settings
|
||||
|
||||
|
||||
CONFIG_PATH = environ.get("CONFIG_PATH", "./config.sample.yaml")
|
||||
BATCH_SIZE = 1000
|
||||
|
||||
|
||||
IDFM_ROOT_URL = "https://data.iledefrance-mobilites.fr/explore/dataset"
|
||||
IDFM_CONNECTION_AREAS_URL = (
|
||||
f"{IDFM_ROOT_URL}/zones-de-correspondance/download/?format=json"
|
||||
)
|
||||
IDFM_LINES_URL = f"{IDFM_ROOT_URL}/referentiel-des-lignes/download/?format=json"
|
||||
IDFM_PICTO_URL = f"{IDFM_ROOT_URL}/referentiel-des-lignes/files"
|
||||
IDFM_STOP_AREAS_URL = f"{IDFM_ROOT_URL}/zones-d-arrets/download/?format=json"
|
||||
IDFM_STOP_SHAPES_URL = "https://eu.ftp.opendatasoft.com/stif/Reflex/REF_ArR.zip"
|
||||
IDFM_STOP_AREA_SHAPES_URL = "https://eu.ftp.opendatasoft.com/stif/Reflex/REF_ZdA.zip"
|
||||
|
||||
IDFM_STOP_STOP_AREAS_ASSOS_URL = f"{IDFM_ROOT_URL}/relations/download/?format=json"
|
||||
IDFM_STOPS_LINES_ASSOS_URL = f"{IDFM_ROOT_URL}/arrets-lignes/download/?format=json"
|
||||
IDFM_STOPS_URL = f"{IDFM_ROOT_URL}/arrets/download/?format=json"
|
||||
|
||||
RATP_ROOT_URL = "https://data.ratp.fr/api/explore/v2.1/catalog/datasets"
|
||||
RATP_PICTOS_URL = (
|
||||
f"{RATP_ROOT_URL}"
|
||||
"/pictogrammes-des-lignes-de-metro-rer-tramway-bus-et-noctilien/exports/json?lang=fr"
|
||||
)
|
||||
|
||||
|
||||
# From https://stackoverflow.com/a/38739634
|
||||
class TqdmLoggingHandler(LoggingHandler):
|
||||
def __init__(self, level=NOTSET):
|
||||
super().__init__(level)
|
||||
|
||||
def emit(self, record):
|
||||
try:
|
||||
msg = self.format(record)
|
||||
tqdm.write(msg)
|
||||
self.flush()
|
||||
except Exception:
|
||||
self.handleError(record)
|
||||
|
||||
|
||||
logger = getLogger(__name__)
|
||||
logger.setLevel(INFO)
|
||||
logger.addHandler(TqdmLoggingHandler())
|
||||
|
||||
|
||||
epsg2154_epsg3857_transformer = Transformer.from_crs(2154, 3857)
|
||||
|
||||
json_stops_decoder = Decoder(type=List[IdfmStop])
|
||||
json_stop_areas_decoder = Decoder(type=List[IdfmStopArea])
|
||||
json_connection_areas_decoder = Decoder(type=List[IdfmConnectionArea])
|
||||
json_lines_decoder = Decoder(type=List[IdfmLine])
|
||||
json_stops_lines_assos_decoder = Decoder(type=List[IdfmStopLineAsso])
|
||||
json_ratp_pictos_decoder = Decoder(type=List[RatpPicto])
|
||||
json_stop_area_stop_asso_decoder = Decoder(type=List[StopAreaStopAssociation])
|
||||
|
||||
|
||||
def format_idfm_pictos(*pictos: IdfmPicto) -> Iterable[LinePicto]:
|
||||
ret = []
|
||||
|
||||
for picto in pictos:
|
||||
ret.append(
|
||||
LinePicto(
|
||||
id=picto.id_,
|
||||
mime_type=picto.mimetype,
|
||||
height_px=picto.height,
|
||||
width_px=picto.width,
|
||||
filename=picto.filename,
|
||||
url=f"{IDFM_PICTO_URL}/{picto.id_}/download",
|
||||
thumbnail=picto.thumbnail,
|
||||
format=picto.format,
|
||||
)
|
||||
)
|
||||
|
||||
return ret
|
||||
|
||||
|
||||
def format_ratp_pictos(*pictos: RatpPicto) -> Iterable[tuple[str, LinePicto]]:
|
||||
ret = []
|
||||
|
||||
for picto in pictos:
|
||||
if (fields := picto.noms_des_fichiers) is not None:
|
||||
ret.append(
|
||||
(
|
||||
picto.indices_commerciaux,
|
||||
LinePicto(
|
||||
id=fields.id_,
|
||||
mime_type=f"image/{fields.format.lower()}",
|
||||
height_px=fields.height,
|
||||
width_px=fields.width,
|
||||
filename=fields.filename,
|
||||
url=f"{RATP_PICTOS_URL}/{fields.id_}/download",
|
||||
thumbnail=fields.thumbnail,
|
||||
format=fields.format,
|
||||
),
|
||||
)
|
||||
)
|
||||
|
||||
return ret
|
||||
|
||||
|
||||
def format_idfm_lines(*lines: IdfmLine) -> Iterable[Line]:
|
||||
ret = []
|
||||
|
||||
optional_value = IdfmLine.optional_value
|
||||
|
||||
for line in lines:
|
||||
fields = line.fields
|
||||
|
||||
line_id = fields.id_line
|
||||
try:
|
||||
formatted_line_id = int(line_id[1:] if line_id[0] == "C" else line_id)
|
||||
except ValueError:
|
||||
logger.warning("Unable to format %s line id.", line_id)
|
||||
continue
|
||||
|
||||
try:
|
||||
operator_id = int(fields.operatorref) # type: ignore
|
||||
except (ValueError, TypeError):
|
||||
logger.warning("Unable to format %s operator id.", fields.operatorref)
|
||||
operator_id = 0
|
||||
|
||||
ret.append(
|
||||
Line(
|
||||
id=formatted_line_id,
|
||||
short_name=fields.shortname_line,
|
||||
name=fields.name_line,
|
||||
status=IdfmLineState(fields.status.value),
|
||||
transport_mode=TransportMode(fields.transportmode.value),
|
||||
transport_submode=optional_value(fields.transportsubmode),
|
||||
network_name=optional_value(fields.networkname),
|
||||
group_of_lines_id=optional_value(fields.id_groupoflines),
|
||||
group_of_lines_shortname=optional_value(fields.shortname_groupoflines),
|
||||
colour_web_hexa=fields.colourweb_hexa,
|
||||
text_colour_hexa=fields.textcolourprint_hexa,
|
||||
operator_id=operator_id,
|
||||
operator_name=optional_value(fields.operatorname),
|
||||
accessibility=IdfmState(fields.accessibility.value),
|
||||
visual_signs_available=IdfmState(fields.visualsigns_available.value),
|
||||
audible_signs_available=IdfmState(fields.audiblesigns_available.value),
|
||||
picto_id=fields.picto.id_ if fields.picto is not None else None,
|
||||
record_id=line.recordid,
|
||||
record_ts=int(line.record_timestamp.timestamp()),
|
||||
)
|
||||
)
|
||||
|
||||
return ret
|
||||
|
||||
|
||||
def format_idfm_stops(*stops: IdfmStop) -> Iterable[Stop]:
|
||||
for stop in stops:
|
||||
fields = stop.fields
|
||||
|
||||
try:
|
||||
created_ts = int(fields.arrcreated.timestamp()) # type: ignore
|
||||
except AttributeError:
|
||||
created_ts = None
|
||||
|
||||
epsg3857_point = epsg2154_epsg3857_transformer.transform(
|
||||
fields.arrxepsg2154, fields.arryepsg2154
|
||||
)
|
||||
|
||||
try:
|
||||
postal_region = int(fields.arrpostalregion)
|
||||
except ValueError:
|
||||
logger.warning("Unable to format %s postal region.", fields.arrpostalregion)
|
||||
continue
|
||||
|
||||
yield Stop(
|
||||
id=int(fields.arrid),
|
||||
name=fields.arrname,
|
||||
epsg3857_x=epsg3857_point[0],
|
||||
epsg3857_y=epsg3857_point[1],
|
||||
town_name=fields.arrtown,
|
||||
postal_region=postal_region,
|
||||
transport_mode=TransportMode(fields.arrtype.value),
|
||||
version=fields.arrversion,
|
||||
created_ts=created_ts,
|
||||
changed_ts=int(fields.arrchanged.timestamp()),
|
||||
accessibility=IdfmState(fields.arraccessibility.value),
|
||||
visual_signs_available=IdfmState(fields.arrvisualsigns.value),
|
||||
audible_signs_available=IdfmState(fields.arraudiblesignals.value),
|
||||
record_id=stop.recordid,
|
||||
record_ts=int(stop.record_timestamp.timestamp()),
|
||||
)
|
||||
|
||||
|
||||
def format_idfm_stop_areas(*stop_areas: IdfmStopArea) -> Iterable[StopArea]:
|
||||
for stop_area in stop_areas:
|
||||
fields = stop_area.fields
|
||||
|
||||
try:
|
||||
created_ts = int(fields.zdacreated.timestamp()) # type: ignore
|
||||
except AttributeError:
|
||||
created_ts = None
|
||||
|
||||
epsg3857_point = epsg2154_epsg3857_transformer.transform(
|
||||
fields.zdaxepsg2154, fields.zdayepsg2154
|
||||
)
|
||||
|
||||
yield StopArea(
|
||||
id=int(fields.zdaid),
|
||||
name=fields.zdaname,
|
||||
town_name=fields.zdatown,
|
||||
postal_region=fields.zdapostalregion,
|
||||
epsg3857_x=epsg3857_point[0],
|
||||
epsg3857_y=epsg3857_point[1],
|
||||
type=StopAreaType(fields.zdatype.value),
|
||||
version=fields.zdaversion,
|
||||
created_ts=created_ts,
|
||||
changed_ts=int(fields.zdachanged.timestamp()),
|
||||
)
|
||||
|
||||
|
||||
def format_idfm_connection_areas(
|
||||
*connection_areas: IdfmConnectionArea,
|
||||
) -> Iterable[ConnectionArea]:
|
||||
for connection_area in connection_areas:
|
||||
fields = connection_area.fields
|
||||
|
||||
epsg3857_point = epsg2154_epsg3857_transformer.transform(
|
||||
fields.zdcxepsg2154, fields.zdcyepsg2154
|
||||
)
|
||||
|
||||
yield ConnectionArea(
|
||||
id=int(fields.zdcid),
|
||||
name=fields.zdcname,
|
||||
town_name=fields.zdctown,
|
||||
postal_region=fields.zdcpostalregion,
|
||||
epsg3857_x=epsg3857_point[0],
|
||||
epsg3857_y=epsg3857_point[1],
|
||||
transport_mode=StopAreaType(fields.zdctype.value),
|
||||
version=fields.zdcversion,
|
||||
created_ts=int(fields.zdccreated.timestamp()),
|
||||
changed_ts=int(fields.zdcchanged.timestamp()),
|
||||
)
|
||||
|
||||
|
||||
def format_idfm_stop_shapes(*shape_records: ShapeRecord) -> Iterable[StopShape]:
|
||||
for shape_record in shape_records:
|
||||
|
||||
epsg3857_points = [
|
||||
epsg2154_epsg3857_transformer.transform(*point)
|
||||
for point in shape_record.shape.points
|
||||
]
|
||||
|
||||
try:
|
||||
bbox_it = iter(shape_record.shape.bbox)
|
||||
epsg3857_bbox = [
|
||||
epsg2154_epsg3857_transformer.transform(*point)
|
||||
for point in zip(bbox_it, bbox_it)
|
||||
]
|
||||
except AttributeError:
|
||||
# Handle stop shapes for which no bbox is provided
|
||||
epsg3857_bbox = []
|
||||
|
||||
yield StopShape(
|
||||
id=shape_record.record[1],
|
||||
type=shape_record.shape.shapeType,
|
||||
epsg3857_bbox=epsg3857_bbox,
|
||||
epsg3857_points=epsg3857_points,
|
||||
)
|
||||
|
||||
|
||||
async def http_get(url: str) -> str | None:
|
||||
chunks = []
|
||||
|
||||
headers = {"Accept": "application/json"}
|
||||
async with ClientSession(headers=headers) as session:
|
||||
|
||||
async with session.get(url) as response:
|
||||
size = int(response.headers.get("content-length", 0)) or None
|
||||
|
||||
progress_bar = tqdm(desc=f"Downloading {url}", total=size)
|
||||
|
||||
if response.status == 200:
|
||||
async for chunk in response.content.iter_chunked(1024 * 1024):
|
||||
chunks.append(chunk.decode())
|
||||
progress_bar.update(len(chunk))
|
||||
|
||||
else:
|
||||
return None
|
||||
|
||||
return "".join(chunks)
|
||||
|
||||
|
||||
async def http_request(
|
||||
url: str, decode: Callable, format_method: Callable, model: Type[Base]
|
||||
) -> bool:
|
||||
elements = []
|
||||
|
||||
data = await http_get(url)
|
||||
if data is None:
|
||||
return False
|
||||
|
||||
try:
|
||||
for element in decode(data):
|
||||
elements.append(element)
|
||||
|
||||
if len(elements) == BATCH_SIZE:
|
||||
await model.add(format_method(*elements))
|
||||
elements.clear()
|
||||
|
||||
if elements:
|
||||
await model.add(format_method(*elements))
|
||||
|
||||
except ValidationError as err:
|
||||
logger.warning(err)
|
||||
return False
|
||||
|
||||
return True
|
||||
|
||||
|
||||
async def load_idfm_stops() -> bool:
|
||||
return await http_request(
|
||||
IDFM_STOPS_URL, json_stops_decoder.decode, format_idfm_stops, Stop
|
||||
)
|
||||
|
||||
|
||||
async def load_idfm_stop_areas() -> bool:
|
||||
return await http_request(
|
||||
IDFM_STOP_AREAS_URL,
|
||||
json_stop_areas_decoder.decode,
|
||||
format_idfm_stop_areas,
|
||||
StopArea,
|
||||
)
|
||||
|
||||
|
||||
async def load_idfm_connection_areas() -> bool:
|
||||
return await http_request(
|
||||
IDFM_CONNECTION_AREAS_URL,
|
||||
json_connection_areas_decoder.decode,
|
||||
format_idfm_connection_areas,
|
||||
ConnectionArea,
|
||||
)
|
||||
|
||||
|
||||
async def load_idfm_stop_shapes(url: str) -> None:
|
||||
async with ClientSession(headers={"Accept": "application/zip"}) as session:
|
||||
async with session.get(url) as response:
|
||||
size = int(response.headers.get("content-length", 0)) or None
|
||||
dl_progress_bar = tqdm(desc=f"Downloading {url}", total=size)
|
||||
|
||||
if response.status == 200:
|
||||
async with NamedTemporaryFile(suffix=".zip") as tmp_file:
|
||||
async for chunk in response.content.iter_chunked(1024 * 1024):
|
||||
await tmp_file.write(chunk)
|
||||
dl_progress_bar.update(len(chunk))
|
||||
|
||||
with ShapeFileReader(tmp_file.name) as reader:
|
||||
step_begin_ts = time()
|
||||
|
||||
shapes = reader.shapeRecords()
|
||||
shapes_len = len(shapes)
|
||||
|
||||
db_progress_bar = tqdm(
|
||||
desc=f"Filling db with {shapes_len} StopShapes",
|
||||
total=shapes_len,
|
||||
)
|
||||
|
||||
begin, end, finished = 0, BATCH_SIZE, False
|
||||
while not finished:
|
||||
elements = islice(shapes, begin, end)
|
||||
formatteds = list(format_idfm_stop_shapes(*elements))
|
||||
await StopShape.add(formatteds)
|
||||
|
||||
begin = end
|
||||
end = begin + BATCH_SIZE
|
||||
finished = begin > len(shapes)
|
||||
db_progress_bar.update(BATCH_SIZE)
|
||||
|
||||
logger.info(
|
||||
f"Add {StopShape.__name__}s: {time() - step_begin_ts}s"
|
||||
)
|
||||
|
||||
|
||||
async def load_idfm_lines() -> None:
|
||||
data = await http_get(IDFM_LINES_URL)
|
||||
if data is None:
|
||||
return None
|
||||
|
||||
lines, pictos = [], []
|
||||
picto_ids = set()
|
||||
|
||||
for line in json_lines_decoder.decode(data):
|
||||
|
||||
if (picto := line.fields.picto) is not None and picto.id_ not in picto_ids:
|
||||
picto_ids.add(picto.id_)
|
||||
pictos.append(picto)
|
||||
|
||||
lines.append(line)
|
||||
|
||||
if len(lines) == BATCH_SIZE:
|
||||
await LinePicto.add(list(format_idfm_pictos(*pictos)))
|
||||
await Line.add(list(format_idfm_lines(*lines)))
|
||||
lines.clear()
|
||||
pictos.clear()
|
||||
|
||||
if pictos:
|
||||
await LinePicto.add(list(format_idfm_pictos(*pictos)))
|
||||
if lines:
|
||||
await Line.add(list(format_idfm_lines(*lines)))
|
||||
|
||||
|
||||
async def load_ratp_pictos(batch_size: int = 5) -> None:
|
||||
data = await http_get(RATP_PICTOS_URL)
|
||||
if data is None:
|
||||
return None
|
||||
|
||||
pictos = []
|
||||
for picto in json_ratp_pictos_decoder.decode(data):
|
||||
pictos.append(picto)
|
||||
|
||||
if len(pictos) == batch_size:
|
||||
formatteds = format_ratp_pictos(*pictos)
|
||||
await LinePicto.add([picto[1] for picto in formatteds])
|
||||
await Line.add_pictos(formatteds)
|
||||
pictos.clear()
|
||||
|
||||
if pictos:
|
||||
formatteds = format_ratp_pictos(*pictos)
|
||||
await LinePicto.add([picto[1] for picto in formatteds])
|
||||
await Line.add_pictos(formatteds)
|
||||
|
||||
|
||||
async def load_lines_stops_assos(batch_size: int = 5000) -> None:
|
||||
data = await http_get(IDFM_STOPS_LINES_ASSOS_URL)
|
||||
if data is None:
|
||||
return None
|
||||
|
||||
total_assos_nb = total_found_nb = 0
|
||||
assos = []
|
||||
for asso in json_stops_lines_assos_decoder.decode(data):
|
||||
fields = asso.fields
|
||||
try:
|
||||
stop_id = int(fields.stop_id.rsplit(":", 1)[-1])
|
||||
except ValueError as err:
|
||||
logger.error(err)
|
||||
logger.error(f"{fields.stop_id = }")
|
||||
continue
|
||||
|
||||
assos.append((fields.route_long_name, fields.operatorname, stop_id))
|
||||
if len(assos) == batch_size:
|
||||
total_assos_nb += batch_size
|
||||
total_found_nb += await Line.add_stops(assos)
|
||||
assos.clear()
|
||||
|
||||
if assos:
|
||||
total_assos_nb += len(assos)
|
||||
total_found_nb += await Line.add_stops(assos)
|
||||
|
||||
logger.info(f"{total_found_nb} line <-> stop ({total_assos_nb = } found)")
|
||||
|
||||
|
||||
async def load_stop_assos(batch_size: int = 5000) -> None:
|
||||
data = await http_get(IDFM_STOP_STOP_AREAS_ASSOS_URL)
|
||||
if data is None:
|
||||
return None
|
||||
|
||||
total_assos_nb = area_stop_assos_nb = conn_stop_assos_nb = 0
|
||||
area_stop_assos = []
|
||||
connection_stop_assos = []
|
||||
|
||||
for asso in json_stop_area_stop_asso_decoder.decode(data):
|
||||
fields = asso.fields
|
||||
|
||||
stop_id = int(fields.arrid)
|
||||
|
||||
area_stop_assos.append((int(fields.zdaid), stop_id))
|
||||
connection_stop_assos.append((int(fields.zdcid), stop_id))
|
||||
|
||||
if len(area_stop_assos) == batch_size:
|
||||
total_assos_nb += batch_size
|
||||
|
||||
if (found_nb := await StopArea.add_stops(area_stop_assos)) is not None:
|
||||
area_stop_assos_nb += found_nb
|
||||
area_stop_assos.clear()
|
||||
|
||||
if (
|
||||
found_nb := await ConnectionArea.add_stops(connection_stop_assos)
|
||||
) is not None:
|
||||
conn_stop_assos_nb += found_nb
|
||||
connection_stop_assos.clear()
|
||||
|
||||
if area_stop_assos:
|
||||
total_assos_nb += len(area_stop_assos)
|
||||
if (found_nb := await StopArea.add_stops(area_stop_assos)) is not None:
|
||||
area_stop_assos_nb += found_nb
|
||||
if (
|
||||
found_nb := await ConnectionArea.add_stops(connection_stop_assos)
|
||||
) is not None:
|
||||
conn_stop_assos_nb += found_nb
|
||||
|
||||
logger.info(f"{area_stop_assos_nb} stop area <-> stop ({total_assos_nb = } found)")
|
||||
logger.info(f"{conn_stop_assos_nb} stop area <-> stop ({total_assos_nb = } found)")
|
||||
|
||||
|
||||
async def prepare(db: Database) -> None:
|
||||
await load_idfm_lines()
|
||||
|
||||
await gather(
|
||||
*(
|
||||
load_idfm_stops(),
|
||||
load_idfm_stop_areas(),
|
||||
load_idfm_connection_areas(),
|
||||
load_ratp_pictos(),
|
||||
)
|
||||
)
|
||||
|
||||
await gather(
|
||||
*(
|
||||
load_idfm_stop_shapes(IDFM_STOP_SHAPES_URL),
|
||||
load_idfm_stop_shapes(IDFM_STOP_AREA_SHAPES_URL),
|
||||
load_lines_stops_assos(),
|
||||
load_stop_assos(),
|
||||
)
|
||||
)
|
||||
|
||||
|
||||
def load_settings(path: str) -> Settings:
|
||||
with open(path, "r") as config_file:
|
||||
config = safe_load(config_file)
|
||||
|
||||
return Settings(**config)
|
||||
|
||||
|
||||
async def main() -> None:
|
||||
settings = load_settings(CONFIG_PATH)
|
||||
|
||||
await db.connect(settings.db, True)
|
||||
|
||||
begin_ts = time()
|
||||
await prepare(db)
|
||||
logger.info(f"Elapsed time: {time() - begin_ts}s")
|
||||
|
||||
await db.disconnect()
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
run(main())
|
@@ -1,19 +0,0 @@
|
||||
version: '3.7'
|
||||
|
||||
services:
|
||||
|
||||
database:
|
||||
image: postgres:15.1-alpine
|
||||
restart: always
|
||||
environment:
|
||||
- POSTGRES_USER=postgres
|
||||
- POSTGRES_PASSWORD=postgres
|
||||
logging:
|
||||
options:
|
||||
max-size: 10m
|
||||
max-file: "3"
|
||||
ports:
|
||||
- '127.0.0.1:5438:5432'
|
||||
volumes:
|
||||
- ./docker/database/docker-entrypoint-initdb.d:/docker-entrypoint-initdb.d
|
||||
- ./docker/database/data:/var/lib/postgresql/data
|
249
backend/main.py
249
backend/main.py
@@ -1,249 +0,0 @@
|
||||
from collections import defaultdict
|
||||
from datetime import datetime
|
||||
from os import environ, EX_USAGE
|
||||
from typing import Sequence
|
||||
|
||||
from fastapi import FastAPI, HTTPException
|
||||
from fastapi.middleware.cors import CORSMiddleware
|
||||
from fastapi.staticfiles import StaticFiles
|
||||
from rich import print
|
||||
|
||||
from backend.db import db
|
||||
from backend.idfm_interface import Destinations as IdfmDestinations, IdfmInterface
|
||||
from backend.models import Line, Stop, StopArea, StopShape
|
||||
from backend.schemas import (
|
||||
Line as LineSchema,
|
||||
TransportMode,
|
||||
NextPassage as NextPassageSchema,
|
||||
NextPassages as NextPassagesSchema,
|
||||
Stop as StopSchema,
|
||||
StopArea as StopAreaSchema,
|
||||
StopShape as StopShapeSchema,
|
||||
)
|
||||
|
||||
API_KEY = environ.get("API_KEY")
|
||||
if API_KEY is None:
|
||||
print('No "API_KEY" environment variable set... abort.')
|
||||
exit(EX_USAGE)
|
||||
|
||||
# TODO: Remove postgresql+asyncpg from environ variable
|
||||
DB_PATH = "postgresql+asyncpg://cer_user:cer_password@127.0.0.1:5438/cer_db"
|
||||
|
||||
|
||||
app = FastAPI()
|
||||
app.add_middleware(
|
||||
CORSMiddleware,
|
||||
allow_origins=[
|
||||
"https://localhost:4443",
|
||||
"https://localhost:3000",
|
||||
],
|
||||
allow_credentials=True,
|
||||
allow_methods=["*"],
|
||||
allow_headers=["*"],
|
||||
)
|
||||
|
||||
idfm_interface = IdfmInterface(API_KEY, db)
|
||||
|
||||
|
||||
@app.on_event("startup")
|
||||
async def startup():
|
||||
await db.connect(DB_PATH, clear_static_data=True)
|
||||
await idfm_interface.startup()
|
||||
# await db.connect(DB_PATH, clear_static_data=False)
|
||||
print("Connected")
|
||||
|
||||
|
||||
@app.on_event("shutdown")
|
||||
async def shutdown():
|
||||
await db.disconnect()
|
||||
|
||||
|
||||
STATIC_ROOT = "../frontend/"
|
||||
app.mount("/widget", StaticFiles(directory=STATIC_ROOT, html=True), name="widget")
|
||||
|
||||
|
||||
def optional_datetime_to_ts(dt: datetime | None) -> int | None:
|
||||
return int(dt.timestamp()) if dt else None
|
||||
|
||||
|
||||
@app.get("/line/{line_id}", response_model=LineSchema)
|
||||
async def get_line(line_id: str) -> LineSchema:
|
||||
line: Line | None = await Line.get_by_id(line_id)
|
||||
|
||||
if line is None:
|
||||
raise HTTPException(status_code=404, detail=f'Line "{line_id}" not found')
|
||||
|
||||
return LineSchema(
|
||||
id=line.id,
|
||||
shortName=line.short_name,
|
||||
name=line.name,
|
||||
status=line.status,
|
||||
transportMode=TransportMode.from_idfm_transport_mode(
|
||||
line.transport_mode, line.transport_submode
|
||||
),
|
||||
backColorHexa=line.colour_web_hexa,
|
||||
foreColorHexa=line.text_colour_hexa,
|
||||
operatorId=line.operator_id,
|
||||
accessibility=line.accessibility,
|
||||
visualSignsAvailable=line.visual_signs_available,
|
||||
audibleSignsAvailable=line.audible_signs_available,
|
||||
stopIds=[stop.id for stop in line.stops],
|
||||
)
|
||||
|
||||
|
||||
def _format_stop(stop: Stop) -> StopSchema:
|
||||
return StopSchema(
|
||||
id=stop.id,
|
||||
name=stop.name,
|
||||
town=stop.town_name,
|
||||
epsg3857_x=stop.epsg3857_x,
|
||||
epsg3857_y=stop.epsg3857_y,
|
||||
lines=[line.id for line in stop.lines],
|
||||
)
|
||||
|
||||
|
||||
@app.get("/stop/")
|
||||
async def get_stop(
|
||||
name: str = "", limit: int = 10
|
||||
) -> Sequence[StopAreaSchema | StopSchema]:
|
||||
# TODO: Add limit support
|
||||
|
||||
formatted: list[StopAreaSchema | StopSchema] = []
|
||||
matching_stops = await Stop.get_by_name(name)
|
||||
# print(matching_stops, flush=True)
|
||||
|
||||
stop_areas: dict[int, StopArea] = {}
|
||||
stops: dict[int, Stop] = {}
|
||||
for stop in matching_stops:
|
||||
# print(f"{stop.__dict__ = }", flush=True)
|
||||
dst = stop_areas if isinstance(stop, StopArea) else stops
|
||||
dst[stop.id] = stop
|
||||
|
||||
for stop_area in stop_areas.values():
|
||||
|
||||
formatted_stops = []
|
||||
for stop in stop_area.stops:
|
||||
formatted_stops.append(_format_stop(stop))
|
||||
try:
|
||||
del stops[stop.id]
|
||||
except KeyError as err:
|
||||
print(err)
|
||||
|
||||
formatted.append(
|
||||
StopAreaSchema(
|
||||
id=stop_area.id,
|
||||
name=stop_area.name,
|
||||
town=stop_area.town_name,
|
||||
type=stop_area.type,
|
||||
lines=[line.id for line in stop_area.lines],
|
||||
stops=formatted_stops,
|
||||
)
|
||||
)
|
||||
|
||||
# print(f"{stops = }", flush=True)
|
||||
formatted.extend(_format_stop(stop) for stop in stops.values())
|
||||
|
||||
return formatted
|
||||
|
||||
|
||||
# TODO: Cache response for 30 secs ?
|
||||
@app.get("/stop/nextPassages/{stop_id}")
|
||||
async def get_next_passages(stop_id: str) -> NextPassagesSchema | None:
|
||||
res = await idfm_interface.get_next_passages(stop_id)
|
||||
if res is None:
|
||||
return None
|
||||
|
||||
service_delivery = res.Siri.ServiceDelivery
|
||||
stop_monitoring_deliveries = service_delivery.StopMonitoringDelivery
|
||||
|
||||
by_line_by_dst_passages: dict[
|
||||
str, dict[str, list[NextPassageSchema]]
|
||||
] = defaultdict(lambda: defaultdict(list))
|
||||
|
||||
for delivery in stop_monitoring_deliveries:
|
||||
for stop_visit in delivery.MonitoredStopVisit:
|
||||
|
||||
journey = stop_visit.MonitoredVehicleJourney
|
||||
|
||||
# re.match will return None if the given journey.LineRef.value is not valid.
|
||||
try:
|
||||
line_id = IdfmInterface.LINE_RE.match(journey.LineRef.value).group(1)
|
||||
except AttributeError as exc:
|
||||
raise HTTPException(
|
||||
status_code=404, detail=f'Line "{journey.LineRef.value}" not found'
|
||||
) from exc
|
||||
|
||||
call = journey.MonitoredCall
|
||||
|
||||
dst_names = call.DestinationDisplay
|
||||
dsts = [dst.value for dst in dst_names] if dst_names else []
|
||||
arrivalPlatformName = (
|
||||
call.ArrivalPlatformName.value if call.ArrivalPlatformName else None
|
||||
)
|
||||
|
||||
next_passage = NextPassageSchema(
|
||||
line=line_id,
|
||||
operator=journey.OperatorRef.value,
|
||||
destinations=dsts,
|
||||
atStop=call.VehicleAtStop,
|
||||
aimedArrivalTs=optional_datetime_to_ts(call.AimedArrivalTime),
|
||||
expectedArrivalTs=optional_datetime_to_ts(call.ExpectedArrivalTime),
|
||||
arrivalPlatformName=arrivalPlatformName,
|
||||
aimedDepartTs=optional_datetime_to_ts(call.AimedDepartureTime),
|
||||
expectedDepartTs=optional_datetime_to_ts(call.ExpectedDepartureTime),
|
||||
arrivalStatus=call.ArrivalStatus.value,
|
||||
departStatus=call.DepartureStatus.value,
|
||||
)
|
||||
|
||||
by_line_passages = by_line_by_dst_passages[line_id]
|
||||
# TODO: by_line_passages[dst].extend(dsts) instead ?
|
||||
for dst in dsts:
|
||||
by_line_passages[dst].append(next_passage)
|
||||
|
||||
return NextPassagesSchema(
|
||||
ts=service_delivery.ResponseTimestamp.timestamp(),
|
||||
passages=by_line_by_dst_passages,
|
||||
)
|
||||
|
||||
|
||||
@app.get("/stop/{stop_id}/destinations")
|
||||
async def get_stop_destinations(
|
||||
stop_id: int,
|
||||
) -> IdfmDestinations | None:
|
||||
destinations = await idfm_interface.get_destinations(stop_id)
|
||||
|
||||
return destinations
|
||||
|
||||
|
||||
# TODO: Rename endpoint -> /stop/{stop_id}/shape
|
||||
@app.get("/stop_shape/{stop_id}")
|
||||
async def get_stop_shape(stop_id: int) -> StopShapeSchema | None:
|
||||
connection_area = None
|
||||
|
||||
if (stop := await Stop.get_by_id(stop_id)) is not None:
|
||||
connection_area = stop.connection_area
|
||||
|
||||
elif (stop_area := await StopArea.get_by_id(stop_id)) is not None:
|
||||
connection_areas = {stop.connection_area for stop in stop_area.stops}
|
||||
connection_areas_len = len(connection_areas)
|
||||
if connection_areas_len == 1:
|
||||
connection_area = connection_areas.pop()
|
||||
|
||||
else:
|
||||
prefix = "More than one" if connection_areas_len else "No"
|
||||
msg = f"{prefix} connection area has been found for stop area #{stop_id}"
|
||||
raise HTTPException(status_code=500, detail=msg)
|
||||
|
||||
if (
|
||||
connection_area is not None
|
||||
and (shape := await StopShape.get_by_id(connection_area.id)) is not None
|
||||
):
|
||||
return StopShapeSchema(
|
||||
id=shape.id,
|
||||
type=shape.type,
|
||||
epsg3857_bbox=shape.epsg3857_bbox,
|
||||
epsg3857_points=shape.epsg3857_points,
|
||||
)
|
||||
|
||||
msg = f"No shape found for stop {stop_id}"
|
||||
raise HTTPException(status_code=404, detail=msg)
|
@@ -4,25 +4,52 @@ version = "0.1.0"
|
||||
description = ""
|
||||
authors = ["Adrien SUEUR <me@adrien.run>"]
|
||||
readme = "README.md"
|
||||
packages = [{include = "backend"}]
|
||||
|
||||
[tool.poetry.dependencies]
|
||||
python = "^3.11"
|
||||
aiohttp = "^3.8.3"
|
||||
rich = "^12.6.0"
|
||||
aiofiles = "^22.1.0"
|
||||
sqlalchemy = {extras = ["asyncio"], version = "^2.0.1"}
|
||||
fastapi = "^0.88.0"
|
||||
fastapi = "^0.103.0"
|
||||
uvicorn = "^0.20.0"
|
||||
asyncpg = "^0.27.0"
|
||||
msgspec = "^0.12.0"
|
||||
pyshp = "^2.3.1"
|
||||
opentelemetry-instrumentation-fastapi = "^0.38b0"
|
||||
sqlalchemy-utils = "^0.41.1"
|
||||
opentelemetry-instrumentation-logging = "^0.38b0"
|
||||
opentelemetry-sdk = "^1.17.0"
|
||||
opentelemetry-api = "^1.17.0"
|
||||
opentelemetry-exporter-otlp-proto-http = "^1.17.0"
|
||||
opentelemetry-instrumentation-sqlalchemy = "^0.38b0"
|
||||
sqlalchemy = "^2.0.12"
|
||||
psycopg = "^3.1.9"
|
||||
pyyaml = "^6.0"
|
||||
fastapi-cache2 = {extras = ["redis"], version = "^0.2.1"}
|
||||
pydantic-settings = "^2.0.3"
|
||||
|
||||
[tool.poetry.group.db_updater.dependencies]
|
||||
aiofiles = "^22.1.0"
|
||||
aiohttp = "^3.8.3"
|
||||
fastapi = "^0.103.0"
|
||||
msgspec = "^0.12.0"
|
||||
opentelemetry-instrumentation-fastapi = "^0.38b0"
|
||||
opentelemetry-instrumentation-sqlalchemy = "^0.38b0"
|
||||
opentelemetry-sdk = "^1.17.0"
|
||||
opentelemetry-api = "^1.17.0"
|
||||
psycopg = "^3.1.9"
|
||||
pyproj = "^3.5.0"
|
||||
pyshp = "^2.3.1"
|
||||
python = "^3.11"
|
||||
pyyaml = "^6.0"
|
||||
sqlalchemy = "^2.0.12"
|
||||
sqlalchemy-utils = "^0.41.1"
|
||||
tqdm = "^4.65.0"
|
||||
pydantic-settings = "^2.0.3"
|
||||
|
||||
[build-system]
|
||||
requires = ["poetry-core"]
|
||||
build-backend = "poetry.core.masonry.api"
|
||||
|
||||
|
||||
|
||||
[tool.poetry.group.dev.dependencies]
|
||||
pylsp-mypy = "^0.6.2"
|
||||
mccabe = "^0.7.0"
|
||||
@@ -39,8 +66,11 @@ autopep8 = "^2.0.1"
|
||||
pyflakes = "^3.0.1"
|
||||
yapf = "^0.32.0"
|
||||
whatthepatch = "^1.0.4"
|
||||
sqlalchemy = {extras = ["mypy"], version = "^2.0.1"}
|
||||
mypy = "^1.0.0"
|
||||
icecream = "^2.1.3"
|
||||
types-sqlalchemy-utils = "^1.0.1"
|
||||
types-pyyaml = "^6.0.12.9"
|
||||
types-tqdm = "^4.65.0.1"
|
||||
|
||||
[tool.mypy]
|
||||
plugins = "sqlalchemy.ext.mypy.plugin"
|
||||
|
99
docker-compose.yml
Normal file
99
docker-compose.yml
Normal file
@@ -0,0 +1,99 @@
|
||||
version: '3.7'
|
||||
|
||||
services:
|
||||
|
||||
database:
|
||||
image: postgres:15.1-alpine
|
||||
restart: always
|
||||
environment:
|
||||
- POSTGRES_USER=postgres
|
||||
- POSTGRES_PASSWORD=postgres
|
||||
logging:
|
||||
options:
|
||||
max-size: 10m
|
||||
max-file: "3"
|
||||
ports:
|
||||
- "127.0.0.1:5432:5432"
|
||||
volumes:
|
||||
- ./backend/docker/database/docker-entrypoint-initdb.d:/docker-entrypoint-initdb.d
|
||||
- ./backend/docker/database/data:/var/lib/postgresql/data
|
||||
|
||||
redis:
|
||||
image: redis:latest
|
||||
restart: always
|
||||
command: redis-server --loglevel warning
|
||||
ports:
|
||||
- "127.0.0.1:6379:6379"
|
||||
|
||||
jaeger-agent:
|
||||
image: jaegertracing/jaeger-agent:latest
|
||||
command:
|
||||
- "--reporter.grpc.host-port=jaeger-collector:14250"
|
||||
ports:
|
||||
- "127.0.0.1:5775:5775/udp"
|
||||
- "127.0.0.1:6831:6831/udp"
|
||||
- "127.0.0.1:6832:6832/udp"
|
||||
- "127.0.0.1:5778:5778"
|
||||
restart: on-failure
|
||||
depends_on:
|
||||
- jaeger-collector
|
||||
|
||||
jaeger-collector:
|
||||
image: jaegertracing/jaeger-collector:latest
|
||||
command:
|
||||
- "--cassandra.keyspace=jaeger_v1_dc1"
|
||||
- "--cassandra.servers=cassandra"
|
||||
- "--collector.zipkin.host-port=9411"
|
||||
- "--sampling.initial-sampling-probability=.5"
|
||||
- "--sampling.target-samples-per-second=.01"
|
||||
environment:
|
||||
- SAMPLING_CONFIG_TYPE=adaptive
|
||||
- COLLECTOR_OTLP_ENABLED=true
|
||||
ports:
|
||||
- "127.0.0.1:4317:4317"
|
||||
- "127.0.0.1:4318:4318"
|
||||
restart: on-failure
|
||||
depends_on:
|
||||
- cassandra-schema
|
||||
|
||||
cassandra:
|
||||
image: cassandra:latest
|
||||
|
||||
cassandra-schema:
|
||||
image: jaegertracing/jaeger-cassandra-schema:latest
|
||||
depends_on:
|
||||
- cassandra
|
||||
|
||||
jaeger-query:
|
||||
image: jaegertracing/jaeger-query:latest
|
||||
command:
|
||||
- "--cassandra.keyspace=jaeger_v1_dc1"
|
||||
- "--cassandra.servers=cassandra"
|
||||
ports:
|
||||
- "127.0.0.1:16686:16686"
|
||||
restart: on-failure
|
||||
depends_on:
|
||||
- cassandra-schema
|
||||
|
||||
carrramba-encore-rate-api:
|
||||
build:
|
||||
context: ./backend/
|
||||
dockerfile: Dockerfile.backend
|
||||
environment:
|
||||
- CONFIG_PATH=./config.local.yaml
|
||||
- IDFM_API_KEY=set_your_idfm_key_here
|
||||
ports:
|
||||
- "127.0.0.1:8080:8080"
|
||||
|
||||
carrramba-encore-rate-frontend:
|
||||
build:
|
||||
context: ./frontend/
|
||||
ports:
|
||||
- "127.0.0.1:80:8081"
|
||||
|
||||
carrramba-encore-rate-db-updater:
|
||||
build:
|
||||
context: ./backend/
|
||||
dockerfile: Dockerfile.db_updater
|
||||
environment:
|
||||
- CONFIG_PATH=./config.local.yaml
|
4
frontend/Dockerfile
Normal file
4
frontend/Dockerfile
Normal file
@@ -0,0 +1,4 @@
|
||||
# pull the latest official nginx image
|
||||
FROM nginx:mainline-alpine-slim
|
||||
|
||||
COPY dist /usr/share/nginx/html
|
@@ -1,4 +1,4 @@
|
||||
import { Component } from 'solid-js';
|
||||
import { Component, createSignal } from 'solid-js';
|
||||
import { IVisibilityActionRequest, MatrixCapabilities, WidgetApi, WidgetApiToWidgetAction } from 'matrix-widget-api';
|
||||
import { HopeProvider } from "@hope-ui/solid";
|
||||
|
||||
@@ -6,9 +6,10 @@ import { BusinessDataProvider } from './businessData';
|
||||
import { AppContextProvider } from './appContext';
|
||||
|
||||
import { PassagesDisplay } from './passagesDisplay';
|
||||
import { StopsSearchMenu } from './stopsSearchMenu';
|
||||
import { StopsSearchMenu } from './stopsSearchMenu/stopsSearchMenu';
|
||||
|
||||
import "./App.scss";
|
||||
import { onCleanup, onMount } from 'solid-js';
|
||||
|
||||
|
||||
function parseFragment() {
|
||||
@@ -43,6 +44,33 @@ const App: Component = () => {
|
||||
api.transport.reply(ev.detail, {});
|
||||
});
|
||||
|
||||
createSignal({
|
||||
height: window.innerHeight,
|
||||
width: window.innerWidth
|
||||
});
|
||||
|
||||
const onResize = () => {
|
||||
const body = document.body;
|
||||
|
||||
if (window.innerWidth * 9 / 16 < window.innerHeight) {
|
||||
body.style['height'] = 'auto';
|
||||
body.style['width'] = '100vw';
|
||||
}
|
||||
else {
|
||||
body.style['height'] = '100vh';
|
||||
body.style['width'] = 'auto';
|
||||
}
|
||||
};
|
||||
|
||||
onMount(() => {
|
||||
window.addEventListener('resize', onResize);
|
||||
onResize();
|
||||
});
|
||||
|
||||
onCleanup(() => {
|
||||
window.removeEventListener('resize', onResize);
|
||||
})
|
||||
|
||||
return (
|
||||
<BusinessDataProvider>
|
||||
<AppContextProvider>
|
||||
|
@@ -21,15 +21,14 @@ export interface BusinessDataStore {
|
||||
getStop: (stopId: number) => Stop | undefined;
|
||||
searchStopByName: (name: string) => Promise<Stops>;
|
||||
|
||||
getStopDestinations: (stopId: number) => Promise<StopDestinations>;
|
||||
getStopDestinations: (stopId: number) => Promise<StopDestinations | undefined>;
|
||||
getStopShape: (stopId: number) => Promise<StopShape | undefined>;
|
||||
};
|
||||
|
||||
export const BusinessDataContext = createContext<BusinessDataStore>();
|
||||
|
||||
export function BusinessDataProvider(props: { children: JSX.Element }) {
|
||||
|
||||
const [serverUrl] = createSignal<string>("https://localhost:4443");
|
||||
const [serverUrl] = createSignal<string>("https://carrramba.adrien.run/api");
|
||||
|
||||
type Store = {
|
||||
lines: Lines;
|
||||
@@ -44,11 +43,20 @@ export function BusinessDataProvider(props: { children: JSX.Element }) {
|
||||
let line = store.lines[lineId];
|
||||
if (line === undefined) {
|
||||
console.log(`${lineId} not found... fetch it from backend.`);
|
||||
const data = await fetch(`${serverUrl()}/line/${lineId}`, {
|
||||
|
||||
const response = await fetch(`${serverUrl()}/line/${lineId}`, {
|
||||
headers: { 'Content-Type': 'application/json' }
|
||||
});
|
||||
line = await data.json();
|
||||
setStore('lines', lineId, line);
|
||||
|
||||
const json = await response.json();
|
||||
|
||||
if (response.ok) {
|
||||
setStore('lines', lineId, json);
|
||||
line = json;
|
||||
}
|
||||
else {
|
||||
console.warn(`No line found for ${lineId} line id:`, json);
|
||||
}
|
||||
}
|
||||
return line;
|
||||
}
|
||||
@@ -91,12 +99,19 @@ export function BusinessDataProvider(props: { children: JSX.Element }) {
|
||||
}
|
||||
|
||||
const refreshPassages = async (stopId: number): Promise<void> => {
|
||||
const httpOptions = { headers: { "Content-Type": "application/json" } };
|
||||
console.log(`Fetching data for ${stopId}`);
|
||||
const data = await fetch(`${serverUrl()}/stop/nextPassages/${stopId}`, httpOptions);
|
||||
const response = await data.json();
|
||||
_cleanupPassages(response.passages);
|
||||
addPassages(response.passages);
|
||||
const httpOptions = { headers: { "Content-Type": "application/json" } };
|
||||
const response = await fetch(`${serverUrl()}/stop/${stopId}/nextPassages`, httpOptions);
|
||||
|
||||
const json = await response.json();
|
||||
|
||||
if (response.ok) {
|
||||
_cleanupPassages(json.passages);
|
||||
addPassages(json.passages);
|
||||
}
|
||||
else {
|
||||
console.warn(`No passage found for ${stopId} stop:`, json);
|
||||
}
|
||||
}
|
||||
|
||||
const addPassages = (passages: Passages): void => {
|
||||
@@ -155,39 +170,58 @@ ${linePassagesDestination.length} here... refresh all them.`);
|
||||
}
|
||||
|
||||
const searchStopByName = async (name: string): Promise<Stops> => {
|
||||
const data = await fetch(`${serverUrl()}/stop/?name=${name}`, {
|
||||
const byIdStops: Stops = {};
|
||||
|
||||
const response = await fetch(`${serverUrl()}/stop/?name=${name}`, {
|
||||
headers: { 'Content-Type': 'application/json' }
|
||||
});
|
||||
const stops = await data.json();
|
||||
|
||||
const byIdStops: Stops = {};
|
||||
for (const stop of stops) {
|
||||
byIdStops[stop.id] = stop;
|
||||
setStore('stops', stop.id, stop);
|
||||
for (const innerStop of stop.stops) {
|
||||
setStore('stops', innerStop.id, innerStop);
|
||||
const json = await response.json();
|
||||
|
||||
if (response.ok) {
|
||||
for (const stop of json) {
|
||||
byIdStops[stop.id] = stop;
|
||||
setStore('stops', stop.id, stop);
|
||||
|
||||
if (stop.stops !== undefined) {
|
||||
for (const innerStop of stop.stops) {
|
||||
setStore('stops', innerStop.id, innerStop);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
else {
|
||||
console.warn(`No stop found for '${name}' query:`, json);
|
||||
}
|
||||
|
||||
return byIdStops;
|
||||
}
|
||||
|
||||
const getStopDestinations = async (stopId: number): Promise<StopDestinations> => {
|
||||
const data = await fetch(`${serverUrl()}/stop/${stopId}/destinations`, {
|
||||
const getStopDestinations = async (stopId: number): Promise<StopDestinations | undefined> => {
|
||||
const response = await fetch(`${serverUrl()}/stop/${stopId}/destinations`, {
|
||||
headers: { 'Content-Type': 'application/json' }
|
||||
});
|
||||
const response = await data.json();
|
||||
return response;
|
||||
const destinations = response.ok ? await response.json() : undefined;
|
||||
return destinations;
|
||||
}
|
||||
|
||||
const getStopShape = async (stopId: number): Promise<StopShape | undefined> => {
|
||||
let shape = store.stopShapes[stopId];
|
||||
if (shape === undefined) {
|
||||
console.log(`No shape found for ${stopId} stop... fetch it from backend.`);
|
||||
const data = await fetch(`${serverUrl()}/stop_shape/${stopId}`, {
|
||||
const response = await fetch(`${serverUrl()}/stop/${stopId}/shape`, {
|
||||
headers: { 'Content-Type': 'application/json' }
|
||||
});
|
||||
shape = await data.json();
|
||||
setStore('stopShapes', stopId, shape);
|
||||
|
||||
const json = await response.json();
|
||||
|
||||
if (response.ok) {
|
||||
setStore('stopShapes', stopId, json);
|
||||
shape = json;
|
||||
}
|
||||
else {
|
||||
console.warn(`No shape found for ${stopId} stop:`, json);
|
||||
}
|
||||
}
|
||||
return shape;
|
||||
}
|
||||
|
@@ -13,10 +13,8 @@
|
||||
src: url(/public/fonts/IDFVoyageur-Medium.otf);
|
||||
}
|
||||
|
||||
body {
|
||||
html, body {
|
||||
aspect-ratio: 16/9;
|
||||
width: 100vw;
|
||||
height: none;
|
||||
|
||||
margin: 0;
|
||||
|
||||
|
@@ -116,7 +116,9 @@ export const Map: ParentComponent<{}> = () => {
|
||||
const foundStopIds = new Set();
|
||||
for (const foundStop of stops) {
|
||||
foundStopIds.add(foundStop.id);
|
||||
foundStop.stops.forEach(s => foundStopIds.add(s.id));
|
||||
if (foundStop.stops !== undefined) {
|
||||
foundStop.stops.forEach(s => foundStopIds.add(s.id));
|
||||
}
|
||||
}
|
||||
|
||||
for (const [stopIdStr, feature] of Object.entries(displayedFeatures)) {
|
||||
|
@@ -8,7 +8,7 @@ export enum TrafficStatus {
|
||||
|
||||
export class Passage {
|
||||
line: number;
|
||||
operator: string;
|
||||
operator: number;
|
||||
destinations: string[];
|
||||
atStop: boolean;
|
||||
aimedArrivalTs: number;
|
||||
@@ -19,7 +19,7 @@ export class Passage {
|
||||
arrivalStatus: string;
|
||||
departStatus: string;
|
||||
|
||||
constructor(line: number, operator: string, destinations: string[], atStop: boolean, aimedArrivalTs: number,
|
||||
constructor(line: number, operator: number, destinations: string[], atStop: boolean, aimedArrivalTs: number,
|
||||
expectedArrivalTs: number, arrivalPlatformName: string, aimedDepartTs: number, expectedDepartTs: number,
|
||||
arrivalStatus: string, departStatus: string) {
|
||||
this.line = line;
|
||||
@@ -45,9 +45,9 @@ export class Stop {
|
||||
epsg3857_x: number;
|
||||
epsg3857_y: number;
|
||||
stops: Stop[];
|
||||
lines: string[];
|
||||
lines: number[];
|
||||
|
||||
constructor(id: number, name: string, town: string, epsg3857_x: number, epsg3857_y: number, stops: Stop[], lines: string[]) {
|
||||
constructor(id: number, name: string, town: string, epsg3857_x: number, epsg3857_y: number, stops: Stop[], lines: number[]) {
|
||||
this.id = id;
|
||||
this.name = name;
|
||||
this.town = town;
|
||||
@@ -82,7 +82,7 @@ export class StopShape {
|
||||
export type StopShapes = Record<number, StopShape>;
|
||||
|
||||
export class Line {
|
||||
id: string;
|
||||
id: number;
|
||||
shortName: string;
|
||||
name: string;
|
||||
status: string; // TODO: Use an enum
|
||||
@@ -95,7 +95,7 @@ export class Line {
|
||||
audibleSignsAvailable: string; // TODO: Use an enum
|
||||
stopIds: number[];
|
||||
|
||||
constructor(id: string, shortName: string, name: string, status: string, transportMode: string, backColorHexa: string,
|
||||
constructor(id: number, shortName: string, name: string, status: string, transportMode: string, backColorHexa: string,
|
||||
foreColorHexa: string, operatorId: number, accessibility: boolean, visualSignsAvailable: string,
|
||||
audibleSignsAvailable: string, stopIds: number[]) {
|
||||
this.id = id;
|
||||
|
@@ -26,7 +26,7 @@ export const TransportModeWeights: Record<string, number> = {
|
||||
export function getTransportModeSrc(mode: string, color: boolean = true): string | undefined {
|
||||
let ret = undefined;
|
||||
if (validTransportModes.includes(mode)) {
|
||||
return `/carrramba-encore-rate/public/symbole_${mode}_${color ? "" : "support_fonce_"}RVB.svg`;
|
||||
return `/symbole_${mode}_${color ? "" : "support_fonce_"}RVB.svg`;
|
||||
}
|
||||
return ret;
|
||||
}
|
||||
|
BIN
medias/presentation.mp4
(Stored with Git LFS)
Normal file
BIN
medias/presentation.mp4
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
medias/presentation.png
(Stored with Git LFS)
Normal file
BIN
medias/presentation.png
(Stored with Git LFS)
Normal file
Binary file not shown.
Reference in New Issue
Block a user