feat: switch to JSON-based log ingestion for higher reliability

This commit is contained in:
2026-04-07 15:47:29 -04:00
parent 6ed92d080f
commit 5f637b5272
24 changed files with 1862 additions and 22 deletions

View File

@@ -1,36 +1,36 @@
import asyncio
import os
import logging
import json
from typing import Any
from pathlib import Path
from decnet.correlation.parser import parse_line
from decnet.web.repository import BaseRepository
logger = logging.getLogger("decnet.web.ingester")
async def log_ingestion_worker(repo: BaseRepository) -> None:
"""
Background task that tails the DECNET_INGEST_LOG_FILE and
inserts parsed LogEvents into the SQLite repository.
Background task that tails the DECNET_INGEST_LOG_FILE.json and
inserts structured JSON logs into the SQLite repository.
"""
log_file_path_str = os.environ.get("DECNET_INGEST_LOG_FILE")
if not log_file_path_str:
base_log_file = os.environ.get("DECNET_INGEST_LOG_FILE")
if not base_log_file:
logger.warning("DECNET_INGEST_LOG_FILE not set. Log ingestion disabled.")
return
log_path = Path(log_file_path_str)
json_log_path = Path(base_log_file).with_suffix(".json")
position = 0
logger.info(f"Starting log ingestion from {log_path}")
logger.info(f"Starting JSON log ingestion from {json_log_path}")
while True:
try:
if not log_path.exists():
if not json_log_path.exists():
await asyncio.sleep(2)
continue
stat = log_path.stat()
stat = json_log_path.stat()
if stat.st_size < position:
# File rotated or truncated
position = 0
@@ -40,26 +40,26 @@ async def log_ingestion_worker(repo: BaseRepository) -> None:
await asyncio.sleep(1)
continue
with open(log_path, "r", encoding="utf-8", errors="replace") as f:
with open(json_log_path, "r", encoding="utf-8", errors="replace") as f:
f.seek(position)
while True:
line = f.readline()
if not line:
break # EOF reached
event = parse_line(line)
if event:
log_data = {
"timestamp": event.timestamp.strftime("%Y-%m-%d %H:%M:%S"),
"decky": event.decky,
"service": event.service,
"event_type": event.event_type,
"attacker_ip": event.attacker_ip or "Unknown",
"raw_line": event.raw
}
if not line.endswith('\n'):
# Partial line read, don't process yet, don't advance position
break
try:
log_data = json.loads(line.strip())
await repo.add_log(log_data)
position = f.tell()
except json.JSONDecodeError:
logger.error(f"Failed to decode JSON log line: {line}")
continue
# Update position after successful line read
position = f.tell()
except Exception as e:
logger.error(f"Error in log ingestion worker: {e}")