Skip to content

catenarytransit/catenary-backend

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4,095 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Architecture

Catenary Backend is a distributed system comprised of microservices operating in Kubernetes. The system is designed for fault tolerance, high-avaliability, and native execution speed in x86-64 using the Rust systems programming language.

  • Maple: GTFS Downloader and ingestion engine into postgres
  • Edelweiss: Routing execution engine (in progress, see also our routing testbed
  • Avens: OSM Preprocessor and graph generator for routing (in progress)
  • Alpenrose: Distributed system to ingest GTFS-rt and other realtime data (Rose des Alpes); successor to Kactus.
  • Aspen: Processing of realtime data and dynamic insertion into other engines. Submodule Pando is used for distribution management
  • Linnaea: Visualisation of the graphs for debugging and research paper purposes
  • Gentian: Transit graph generation runner (in progress)
  • Harebell: Map tile geometry generator creating line ordering optimised graph maps (LOOM) MVT files.
  • Spruce: Websocket server for frontend to stream data to and from backend, including realtime locations, stop times (not started yet)
  • Birch: HTTP API server
  • OSM Station Import: Imports railway stations from OpenStreetMap PBF files for GTFS stop association

The kubernetes configuration is generated using Helm templates. See Helm's documentation for further information on that.

The code is heavily commented, go to each folder in src for more information.

Citations

This project is built off the work of researchers in Germany and many other places, vielen dank!

Submodules maintained

Dependent binaries maintained

  • Pfaedle-rs: Shape matching in rust using low memory.

Agency specific submodules

  • Amtrak GTFS rt: Conversion of proprietary realtime data from amtrak's website into gtfs-rt.
  • Chicago GTFS Rt: conversion of proprietary Chicago CTA data to GTFS realtime
  • Rtc Québec GTFS RT: conversion of proprietary app RTC Nomade to GTFS realtime
  • Via Rail GTFS RT: Conversion of Via Rail tracking to GTFS Realtime.

Install Dependencies

sudo apt install -y postgresql-common
sudo /usr/share/postgresql-common/pgdg/apt.postgresql.org.sh
sudo apt install libprotoc-dev protobuf-compiler build-essential gcc pkg-config libssl-dev unzip wget cmake openssl libpq-dev

install coin cbc integer linear programming solver.

sudo apt-get install coinor-cbc coinor-libcbc-dev

Install postgres

sudo apt install postgresql-18 postgresql-18-postgis-3 postgresql-contrib postgresql

You may also use an external database if you prefer.

Enable PostGis

CREATE EXTENSION postgis;

For Contributors

Good commit messages are required to contribute to this project.

Installation of Postgres

See https://www.postgresql.org/download

PostGIS is also required like

sudo apt install postgresql-16-postgis-3

See https://trac.osgeo.org/postgis/wiki/UsersWikiPostGIS3UbuntuPGSQLApt for more instructions

SQL notes

We've switched to diesel for our queries. Read the diesel documentation to learn how to use it. https://diesel.rs/guides/getting-started.html

Lib PQ is also required to install the diesel cli. Only postgres is required. Example

sudo apt-get install libpq-dev
cargo install diesel_cli --no-default-features --features postgres

Common Database debugging

Is Postgis not installing? This page may be helpful: https://trac.osgeo.org/postgis/wiki/UsersWikiPostGIS3UbuntuPGSQLApt

Updating transitland submodules

git submodule update --rebase --remote

Style Guide

Code should be formatted with cargo fmt and be well documented. The following cargo clippy rules are enforced.

#![deny(
    clippy::mutable_key_type,
    clippy::map_entry,
    clippy::boxed_local,
    clippy::let_unit_value,
    clippy::redundant_allocation,
    clippy::bool_comparison,
    clippy::bind_instead_of_map,
    clippy::vec_box,
    clippy::while_let_loop,
    clippy::useless_asref,
    clippy::repeat_once,
    clippy::deref_addrof,
    clippy::suspicious_map,
    clippy::arc_with_non_send_sync,
    clippy::single_char_pattern,
    clippy::for_kv_map,
    clippy::let_unit_value,
    clippy::let_and_return,
    clippy::iter_nth,
    clippy::iter_cloned_collect,
    clippy::bytes_nth,
    clippy::deprecated_clippy_cfg_attr,
    clippy::match_result_ok,
    clippy::cmp_owned,
    clippy::cmp_null,
    clippy::op_ref,
    clippy::useless_vec,
    clippy::module_inception
)]

Future potential submodule names

  • Truffle reachability analysis
  • Chinese plum (needs to pick better syonym for it)

OSM Station Association

The OSM Station Import system associates GTFS railway stops with OpenStreetMap stations. This provides additional metadata like multilingual names, UIC references, and station relationships.

Step 1: Import OSM Stations

First, obtain a pre-filtered PBF file containing railway stations (or use one from Geofabrik filtered with osmium):

cargo run --bin osmstationimport -- --file /path/to/railstations.osm.pbf

multiple imports can be done sequentially, and will overwrite old data.

cargo run --bin osmstationimport -- --file railstations-europe-latest.osm.pbf
cargo run --bin osmstationimport -- --file railstations-north-america-latest.osm.pbf
cargo run --bin osmstationimport -- --file railstations-asia-latest.osm.pbf

The importer:

  • Computes SHA256 hash to skip duplicate imports
  • Extracts rail, tram, and subway stations
  • Parses multilingual names (name:en, name:de, etc.)
  • Stores in gtfs.osm_stations table with spatial indexing

Step 2: Run Maple Import

When Maple processes GTFS feeds, it automatically matches stops to OSM stations for rail/tram/subway routes:

cargo run --bin maple -- --transitland /path/to/transitland-atlas

See the Maple readme for more info

Routing Engine Setup

The routing engine provides point-to-point transit directions using three main microservices:

  1. Avens: OSM preprocessor to build walking/biking graphs.
  2. Gentian: Timetable compiler to create a zero-copy transit timetable.
  3. Edelweiss: The routing execution engine.

Step 1: Preprocess OSM Graphs with Avens

Avens takes OpenStreetMap PBF files and compiles them into a serialized routing graph format with an R-Tree spatial index.

To compile a batch of regions using a configuration file (recommended for managing multiple regions):

cargo run --bin avens --release -- batch --config avens_regions.toml

Example avens_regions.toml:

[global]
tmp_dir = "/tmp/avens_raw"
output_dir = "/path/to/graphs"

[[regions]]
name = "germany"
url = "https://download.geofabrik.de/europe/germany-latest.osm.pbf"

[[regions]]
name = "switzerland"
url = "https://download.geofabrik.de/europe/switzerland-latest.osm.pbf"

Procedure for Managing and Updating OSM PBFs: To manage your region endpoints, simply update the [[regions]] list in your TOML config file. Re-running the batch command will fetch the latest full PBFs from the listed URLs (overwriting old temp files) and regenerate the routing graphs in the output_dir seamlessly.

Alternatively, you can pull directly from a URL and compile a single region:

cargo run --bin avens --release -- pull --url https://download.geofabrik.de/europe/germany-latest.osm.pbf --output-dir /path/to/graphs/germany

To extract a locally downloaded PBF file:

cargo run --bin avens --release -- extract --input /path/to/germany-latest.osm.pbf --output-dir /path/to/graphs/germany

You can view statistics about a compiled graph directory using:

cargo run --bin avens --release -- info --graph-dir /path/to/graphs/germany

Step 2: Compile Timetables with Gentian

Gentian builds a RAPTOR-compatible zero-copy transit routing timetable from data imported into PostgreSQL (e.g., via Maple). It integrates with the Avens graphs to accurately compute transfer footpaths.

cargo run --bin gentian --release -- --output /path/to/timetable.bin --avens-graphs /path/to/graphs

Step 3: Run the Edelweiss Routing Engine

Edelweiss provides the actual routing service via a Tarpc RPC server (listening on 0.0.0.0:9090). It uses the Avens graphs for OSM routing and dynamically loads timetables compiled by Gentian.

GRAPH_DIR=/path/to/graphs cargo run --bin edelweiss --release

Sponsor this project

Packages

 
 
 

Contributors