Skip to content

Commit 81138e2

Browse files
committed
DuckDB rewrite: adapt project metadata
1 parent f3cde47 commit 81138e2

File tree

9 files changed

+49
-50
lines changed

9 files changed

+49
-50
lines changed

Dockerfile

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
11
FROM node:alpine
2-
LABEL org.opencontainers.image.title="gtfs-via-postgres"
3-
LABEL org.opencontainers.image.description="Process GTFS using PostgreSQL."
2+
LABEL org.opencontainers.image.title="gtfs-via-duckdb"
3+
LABEL org.opencontainers.image.description="Analyze GTFS datasets using DuckDB."
44
LABEL org.opencontainers.image.authors="Jannis R <mail@jannisr.de>"
5-
LABEL org.opencontainers.image.documentation="https://github.com/public-transport/gtfs-via-postgres"
6-
LABEL org.opencontainers.image.source="https://github.com/public-transport/gtfs-via-postgres"
7-
LABEL org.opencontainers.image.revision="4.0.0"
5+
LABEL org.opencontainers.image.documentation="https://github.com/public-transport/gtfs-via-duckdb"
6+
LABEL org.opencontainers.image.source="https://github.com/public-transport/gtfs-via-duckdb"
7+
LABEL org.opencontainers.image.revision="5.0.0"
88
LABEL org.opencontainers.image.licenses="(Apache-2.0 AND Prosperity-3.0.0)"
99

1010
WORKDIR /app

LICENSE-APACHE

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -176,7 +176,7 @@
176176

177177
END OF TERMS AND CONDITIONS
178178

179-
Copyright 2020 gtfs-via-postgres contributors
179+
Copyright 2020 gtfs-via-postgres & gtfs-via-duckdb contributors
180180

181181
Licensed under the Apache License, Version 2.0 (the "License");
182182
you may not use this file except in compliance with the License.

LICENSE-PROSPERITY.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
Contributor: Jannis R
44

5-
Source Code: https://github.com/public-transport/gtfs-via-postgres
5+
Source Code: https://github.com/public-transport/gtfs-via-duckdb
66

77
## Purpose
88

docs/analysis/active-trips-by-hour.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
Do you want to know how many trips are running at a specific point in time?
44

5-
`gtfs-via-postgres` optionally provides a **(materialized) view `stats_active_trips_by_hour` to answer this. Use the `--stats-active-trips-by-hour` flag to enable it**:
5+
`gtfs-via-duckdb` optionally provides a **(materialized) view `stats_active_trips_by_hour` to answer this. Use the `--stats-active-trips-by-hour` flag to enable it**:
66

77
- If you run `gtfs-to-sql` with `--stats-active-trips-by-hour=view`, `stats_active_trips_by_hour` will be a "regular" non-materialized view. Use this option if you want to import the GTFS data quickly, and if you only query `stats_active_trips_by_hour` rarely or in time-uncritical scenarios.
88
- If you pass `--stats-active-trips-by-hour=materialized-view`, the `stats_active_trips_by_hour` view will [be materialized](https://www.postgresql.org/docs/14/rules-materializedviews.html). Use this option if you need fast queries, and if you can tolerate significantly longer import times (a minute for small feeds, many hours for large feeds).
Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
# analysing a GTFS dataset by route ID, stop ID and/or hour
22

3-
With the `--stats-by-route-and-stop-and-hour` option, `gtfs-via-postgres` provides a view `stats_by_agency_route_stop_hour`. Just like [`stats_by_route_id_and_date`](feed-by-route-and-date.md), it aggregates all arrivals by `agency_id`, `route_id`, `stop_id` and `effective_hour`.
3+
With the `--stats-by-route-and-stop-and-hour` option, `gtfs-via-duckdb` provides a view `stats_by_agency_route_stop_hour`. Just like [`stats_by_route_id_and_date`](feed-by-route-and-date.md), it aggregates all arrivals by `agency_id`, `route_id`, `stop_id` and `effective_hour`.
44

55
Note: As a materialized view, `stats_by_agency_route_stop_hour` takes up a significant amount of space, e.g. 13GB with the 2023-05-02 VBB GTFS feed.

docs/analysis/feed-by-route-date.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ Are you trying to answer a question like those below?
66
- Has the number of stop time events decreased, compared to the last dataset version?
77
- Do specific routes stop running during certain time periods?
88

9-
`gtfs-via-postgres` optionally provides a **(materialized) view `stats_by_route_date` to help with such SQL queries. Use the `--stats-by-route-date` flag to enable it** in the generated SQL:
9+
`gtfs-via-duckdb` optionally provides a **(materialized) view `stats_by_route_date` to help with such SQL queries. Use the `--stats-by-route-date` flag to enable it** in the generated SQL:
1010

1111
- If you run `gtfs-to-sql` with `--stats-by-route-date=view`, `stats_by_route_date` will be a "regular" non-materialized view. Use this option if you want to import the GTFS data quickly, and if you only query `stats_by_route_date` rarely or in time-uncritical scenarios.
1212
- If you pass `--stats-by-route-date=materialized-view`, the `stats_by_route_date` view will [be materialized](https://www.postgresql.org/docs/14/rules-materializedviews.html). Use this option if you need fast queries, and if you can tolerate significantly longer import times (3m for the 64mb 2023-03-05 SNCB/NMBS GTFS feed, 1h15m for the 540mb 2023-02-27 VBB GTFS feed).

docs/multiple-datasets.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5,13 +5,13 @@ Using `gtfs-via-postgres`, you can import more than one dataset into a single Po
55
As an example, let's compare two datasets from [Paris](https://en.wikipedia.org/wiki/Île-de-France_Mobilités) and [Berlin](https://en.wikipedia.org/wiki/Verkehrsverbund_Berlin-Brandenburg).
66

77
```shell
8-
wget -U 'gtfs-via-postgres demo' -O paris.gtfs.zip 'https://eu.ftp.opendatasoft.com/stif/GTFS/IDFM-gtfs.zip'
8+
wget -U 'gtfs-via-duckdb demo' -O paris.gtfs.zip 'https://eu.ftp.opendatasoft.com/stif/GTFS/IDFM-gtfs.zip'
99
unzip -d paris.gtfs paris.gtfs.zip
1010
gtfs-to-sql --require-dependencies \
1111
--schema paris -- paris.gtfs/*.txt \
1212
| sponge | psql -b
1313

14-
wget -U 'gtfs-via-postgres demo' -O berlin.gtfs.zip 'https://www.vbb.de/vbbgtfs'
14+
wget -U 'gtfs-via-duckdb demo' -O berlin.gtfs.zip 'https://www.vbb.de/vbbgtfs'
1515
unzip -d berlin.gtfs berlin.gtfs.zip
1616
gtfs-to-sql --require-dependencies \
1717
--schema berlin -- berlin.gtfs/*.txt \

package.json

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
{
2-
"name": "gtfs-via-postgres",
3-
"description": "Process GTFS using PostgreSQL.",
4-
"version": "4.10.4",
2+
"name": "gtfs-via-duckdb",
3+
"description": "Analyze GTFS datasets using DuckDB.",
4+
"version": "5.0.0",
55
"main": "lib/index.js",
66
"bin": {
77
"gtfs-to-sql": "cli.js"
@@ -22,8 +22,8 @@
2222
"public transport",
2323
"transit",
2424
"convert",
25-
"postgres",
26-
"postgresql",
25+
"duckdb",
26+
"data analysis",
2727
"sql"
2828
],
2929
"author": "Jannis R <mail@jannisr.de>",
@@ -32,12 +32,12 @@
3232
"Magnus Burton <magnus.burton@me.com>",
3333
"smohiudd <saadiq@developmentseed.org>"
3434
],
35-
"homepage": "https://github.com/public-transport/gtfs-via-postgres/tree/4.10.4",
35+
"homepage": "https://github.com/public-transport/gtfs-via-duckdb/tree/5.0.0",
3636
"repository": {
3737
"type": "git",
38-
"url": "git+https://github.com/public-transport/gtfs-via-postgres.git"
38+
"url": "git+https://github.com/public-transport/gtfs-via-duckdb.git"
3939
},
40-
"bugs": "https://github.com/public-transport/gtfs-via-postgres/issues",
40+
"bugs": "https://github.com/public-transport/gtfs-via-duckdb/issues",
4141
"license": "(Apache-2.0 AND Prosperity-3.0.0)",
4242
"funding": [
4343
{

0 commit comments

Comments
 (0)