Compare commits
3 commits
a5349123ce
...
96f918445c
| Author | SHA1 | Date | |
|---|---|---|---|
| 96f918445c | |||
| 6891e3b206 | |||
| f37fd68e07 |
2 changed files with 98 additions and 28 deletions
|
|
@ -6,6 +6,7 @@ import os
|
||||||
import sqlite3
|
import sqlite3
|
||||||
from datetime import date, datetime, timedelta, timezone
|
from datetime import date, datetime, timedelta, timezone
|
||||||
from itertools import islice
|
from itertools import islice
|
||||||
|
from zoneinfo import ZoneInfo
|
||||||
|
|
||||||
import httpx
|
import httpx
|
||||||
import pyarrow as pa
|
import pyarrow as pa
|
||||||
|
|
@ -107,18 +108,33 @@ def chunked(iterable, size):
|
||||||
|
|
||||||
|
|
||||||
async def positions_job():
|
async def positions_job():
|
||||||
ts = datetime.now(timezone.utc)
|
fetch_ts = datetime.now(timezone.utc) # when *we* fetched
|
||||||
|
|
||||||
client = get_async_client()
|
client = get_async_client()
|
||||||
try:
|
try:
|
||||||
data = await fetch_positions(client)
|
data = await fetch_positions(client)
|
||||||
# first element is timestamp string
|
# 2) API‐reported timestamp
|
||||||
ts_str, *positions = data
|
server_ts_str, *positions = data
|
||||||
ts = datetime.fromisoformat(ts_str)
|
# server_ts is local to Warsaw
|
||||||
# Build Arrow table
|
server_ts = datetime.fromisoformat(server_ts_str).replace(
|
||||||
table = pa.Table.from_pylist(positions)
|
tzinfo=ZoneInfo("Europe/Warsaw")
|
||||||
# write locally only
|
)
|
||||||
write_positions(table, ts)
|
|
||||||
logger.info("wrote positions shard for %s", ts)
|
# 3) build Arrow table enrich each dict with both timestamps
|
||||||
|
enriched = []
|
||||||
|
for p in positions:
|
||||||
|
enriched.append(
|
||||||
|
{
|
||||||
|
**p,
|
||||||
|
"server_ts": server_ts,
|
||||||
|
"fetched_ts": fetch_ts,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
table = pa.Table.from_pylist(enriched)
|
||||||
|
|
||||||
|
# 4) pass both timestamps through
|
||||||
|
write_positions(table, server_ts)
|
||||||
|
logger.info("wrote positions shard for %s", server_ts)
|
||||||
except Exception:
|
except Exception:
|
||||||
# record error details
|
# record error details
|
||||||
logger.exception("job=positions_job")
|
logger.exception("job=positions_job")
|
||||||
|
|
@ -129,7 +145,7 @@ async def positions_job():
|
||||||
for cid in {p["c"] for p in positions}:
|
for cid in {p["c"] for p in positions}:
|
||||||
if should_fetch_course(cid) and cid not in in_flight_courses:
|
if should_fetch_course(cid) and cid not in in_flight_courses:
|
||||||
in_flight_courses.add(cid)
|
in_flight_courses.add(cid)
|
||||||
await course_queue.put((cid, ts, 0))
|
await course_queue.put((cid, server_ts, 0))
|
||||||
|
|
||||||
|
|
||||||
async def fetch_and_store_course(course_id: int, ts: datetime, client=None):
|
async def fetch_and_store_course(course_id: int, ts: datetime, client=None):
|
||||||
|
|
@ -142,15 +158,21 @@ async def fetch_and_store_course(course_id: int, ts: datetime, client=None):
|
||||||
recs = await fetch_course_posts(client, [course_id])
|
recs = await fetch_course_posts(client, [course_id])
|
||||||
# now do your I/O; catch+log only errors here
|
# now do your I/O; catch+log only errors here
|
||||||
for rec in recs:
|
for rec in recs:
|
||||||
rows = [
|
rows = []
|
||||||
|
for seq, stop in enumerate(rec.get("r", []), start=1):
|
||||||
|
# stop["t"] is like "1900-01-01 05:53:00"
|
||||||
|
# parse out the time portion
|
||||||
|
t = datetime.strptime(stop["t"], "%Y-%m-%d %H:%M:%S").time()
|
||||||
|
rows.append(
|
||||||
{
|
{
|
||||||
"fetch_date": ts,
|
"fetch_date": ts, # UTC fetch timestamp
|
||||||
"course_id": rec["c"],
|
"course_id": rec["c"],
|
||||||
|
"stop_seq": seq, # preserve order!
|
||||||
"stop_id": stop["s"],
|
"stop_id": stop["s"],
|
||||||
"dep_time": stop["t"],
|
"dep_time": t, # real datetime.time
|
||||||
|
"dep_zone": "Europe/Warsaw",
|
||||||
}
|
}
|
||||||
for stop in rec.get("r", [])
|
)
|
||||||
]
|
|
||||||
table = pa.Table.from_pylist(rows)
|
table = pa.Table.from_pylist(rows)
|
||||||
write_course_posts(table, rec["c"], ts)
|
write_course_posts(table, rec["c"], ts)
|
||||||
record_course_fetched(rec["c"], ts)
|
record_course_fetched(rec["c"], ts)
|
||||||
|
|
@ -174,7 +196,23 @@ async def daily_snapshot_job():
|
||||||
try:
|
try:
|
||||||
# stops
|
# stops
|
||||||
posts = await fetch_posts(client)
|
posts = await fetch_posts(client)
|
||||||
table = pa.Table.from_pylist(posts)
|
rows = []
|
||||||
|
for grp in posts:
|
||||||
|
# grp: { "n": name, "s": group_id, "t": group_type, "p": [ {s,x,y,t}, … ] }
|
||||||
|
for stop in grp.get("p", []):
|
||||||
|
rows.append(
|
||||||
|
{
|
||||||
|
"fetch_date": ts, # utc fetch timestamp
|
||||||
|
"group_name": grp["n"],
|
||||||
|
"group_id": grp["s"],
|
||||||
|
"group_type": grp["t"],
|
||||||
|
"stop_id": stop["s"],
|
||||||
|
"stop_type": stop["t"],
|
||||||
|
"x": stop["x"],
|
||||||
|
"y": stop["y"],
|
||||||
|
}
|
||||||
|
)
|
||||||
|
table = pa.Table.from_pylist(rows)
|
||||||
fp = write_stops(table, ts)
|
fp = write_stops(table, ts)
|
||||||
push_to_nas(fp)
|
push_to_nas(fp)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -24,10 +24,14 @@ def _ensure_dir(path: str):
|
||||||
os.makedirs(path, exist_ok=True)
|
os.makedirs(path, exist_ok=True)
|
||||||
|
|
||||||
|
|
||||||
def write_positions(table: pa.Table, ts: datetime) -> str:
|
def write_positions(
|
||||||
|
table: pa.Table,
|
||||||
|
server_ts: datetime,
|
||||||
|
) -> str:
|
||||||
"""
|
"""
|
||||||
Convert x/y to a Point geometry, drop the originals,
|
Convert x/y to a Point geometry, drop the originals,
|
||||||
and write out a GeoParquet.
|
and write out a GeoParquet, carrying through
|
||||||
|
the server_ts and fetched_ts columns.
|
||||||
"""
|
"""
|
||||||
# 1) Extract x/y as numpy arrays (zero_copy_only=False to ensure a NumPy copy)
|
# 1) Extract x/y as numpy arrays (zero_copy_only=False to ensure a NumPy copy)
|
||||||
xs = table.column("x").to_numpy()
|
xs = table.column("x").to_numpy()
|
||||||
|
|
@ -44,11 +48,13 @@ def write_positions(table: pa.Table, ts: datetime) -> str:
|
||||||
# 4) drop old coords & append 'geometry'
|
# 4) drop old coords & append 'geometry'
|
||||||
table = table.drop(["x", "y"]).append_column("geometry", geom)
|
table = table.drop(["x", "y"]).append_column("geometry", geom)
|
||||||
|
|
||||||
subdir = ts.strftime("positions/%Y/%m/%d/%H")
|
# (server_ts and fetched_ts are already columns in `table`)
|
||||||
|
|
||||||
|
subdir = server_ts.strftime("positions/%Y/%m/%d/%H")
|
||||||
local_dir = os.path.join(LOCAL_DATA_DIR, subdir)
|
local_dir = os.path.join(LOCAL_DATA_DIR, subdir)
|
||||||
_ensure_dir(local_dir)
|
_ensure_dir(local_dir)
|
||||||
|
|
||||||
filename = f"{ts.strftime('%M%S')}_{HOST_ID}.parquet"
|
filename = f"{server_ts.strftime('%M%S')}_{HOST_ID}.parquet"
|
||||||
local_path = os.path.join(local_dir, filename)
|
local_path = os.path.join(local_dir, filename)
|
||||||
pq.write_table(table, local_path, compression="snappy")
|
pq.write_table(table, local_path, compression="snappy")
|
||||||
|
|
||||||
|
|
@ -72,21 +78,47 @@ def write_course_posts(table: pa.Table, course_id: int, ts: datetime) -> str:
|
||||||
|
|
||||||
def write_stops(table: pa.Table, snapshot_dt: datetime) -> str:
|
def write_stops(table: pa.Table, snapshot_dt: datetime) -> str:
|
||||||
"""
|
"""
|
||||||
Write the daily snapshot of stops (getPosts).
|
Take a flat table with columns
|
||||||
|
fetch_date, group_name, group_id, group_type,
|
||||||
|
stop_id, stop_type, x, y
|
||||||
|
and convert x/y → a WKB Point geometry,
|
||||||
|
drop the raw coords, and write GeoParquet.
|
||||||
"""
|
"""
|
||||||
|
# 1) pull x,y into numpy
|
||||||
|
xs = table.column("x").to_numpy()
|
||||||
|
ys = table.column("y").to_numpy()
|
||||||
|
|
||||||
|
# 2) build PointType with CRS
|
||||||
|
pt_builder = ga.point().with_crs("EPSG:4326")
|
||||||
|
|
||||||
|
# 3) build WKB geometry
|
||||||
|
geom = ga.as_wkb(pt_builder.from_geobuffers(None, xs, ys))
|
||||||
|
|
||||||
|
# 4) drop coords & append geometry column
|
||||||
|
table = table.drop(["x", "y"]).append_column("geometry", geom)
|
||||||
|
|
||||||
|
# 5) write out
|
||||||
subdir = snapshot_dt.strftime("stops/%Y-%m-%d")
|
subdir = snapshot_dt.strftime("stops/%Y-%m-%d")
|
||||||
local_dir = os.path.join(LOCAL_DATA_DIR, subdir)
|
local_dir = os.path.join(LOCAL_DATA_DIR, subdir)
|
||||||
_ensure_dir(local_dir)
|
_ensure_dir(local_dir)
|
||||||
|
|
||||||
local_path = os.path.join(local_dir, f"stops_{HOST_ID}.parquet")
|
filename = f"stops_{HOST_ID}.parquet"
|
||||||
pq.write_table(table, local_path, compression="snappy")
|
out_path = os.path.join(local_dir, filename)
|
||||||
return local_path
|
pq.write_table(table, out_path, compression="snappy")
|
||||||
|
return out_path
|
||||||
|
|
||||||
|
|
||||||
def write_posts_lines(table: pa.Table, snapshot_dt: datetime) -> str:
|
def write_posts_lines(table: pa.Table, snapshot_dt: datetime) -> str:
|
||||||
"""
|
"""
|
||||||
Write the daily stop→lines mapping (getPostsLines).
|
Write the daily stop→lines mapping (getPostsLines).
|
||||||
"""
|
"""
|
||||||
|
# ensure every row gets a fetch_date
|
||||||
|
if "fetch_date" not in table.column_names:
|
||||||
|
table = table.append_column(
|
||||||
|
"fetch_date",
|
||||||
|
pa.array([snapshot_dt] * table.num_rows, type=pa.timestamp("ns", tz="UTC")),
|
||||||
|
)
|
||||||
|
|
||||||
subdir = snapshot_dt.strftime("stops_lines/%Y-%m-%d")
|
subdir = snapshot_dt.strftime("stops_lines/%Y-%m-%d")
|
||||||
local_dir = os.path.join(LOCAL_DATA_DIR, subdir)
|
local_dir = os.path.join(LOCAL_DATA_DIR, subdir)
|
||||||
_ensure_dir(local_dir)
|
_ensure_dir(local_dir)
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue