midle point. error in constr.

This commit is contained in:
Kalzu Rekku
2025-06-10 23:18:48 +03:00
parent c4fed415dc
commit 11a565732d

View File

@ -1,4 +1,3 @@
# app/main.py
import os
import uuid
@ -16,16 +15,60 @@ import uuid as uuid_lib
from collections import deque
from pythonjsonlogger import jsonlogger
import sys
from .database import RRDDatabase
# --- Service Configuration ---
# Generate a unique Service UUID on startup, or get it from an environment variable
SERVICE_UUID = os.environ.get("SERVICE_UUID", str(uuid.uuid4()))
database = RRDDatabase()
# --- Logging Configuration ---
# Get the root logger
logger = logging.getLogger()
logger.setLevel(logging.INFO)
# Custom handler to capture logs
class BufferHandler(logging.Handler):
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
# Instantiate the formatter once for efficiency
self.formatter = jsonlogger.JsonFormatter()
def emit(self, record):
try:
# Format the record as JSON string, then parse to dict
log_entry = json.loads(self.formatter.format(record))
log_buffer.add_log(log_entry)
except Exception as e:
# Log the error to stderr, to avoid recursion or filling the buffer with errors
print(f"Error in BufferHandler: Could not process log record: {e}", file=sys.stderr)
# Optionally, you could log record.msg or record.exc_info here for more context
class LogBuffer:
def __init__(self, maxlen=1000):
self.buffer = deque(maxlen=maxlen)
def add_log(self, record):
# Ensure 'asctime' is present or handle its absence
timestamp = record.get('asctime') or datetime.utcnow().isoformat()
self.buffer.append({
'timestamp': timestamp,
'level': record.get('levelname'),
'message': record.get('message'),
'extra': {k: v for k, v in record.items()
if k not in ['asctime', 'levelname', 'message', 'name', 'lineno', 'filename', 'pathname', 'funcName', 'process', 'processName', 'thread', 'threadName']}
# Added more common LogRecord attributes to exclude from 'extra'
})
def get_logs(self, limit=100):
return list(self.buffer)[-limit:]
# Create global log buffer
log_buffer = LogBuffer()
# Use a handler that streams to stdout
logHandler = logging.StreamHandler()
@ -36,45 +79,14 @@ formatter = jsonlogger.JsonFormatter(
)
logHandler.setFormatter(formatter)
# Add the handler to the root logger
# Add handlers to the root logger
# Avoid adding handlers multiple times in a uvicorn environment
if not logger.handlers:
logger.addHandler(logHandler)
class LogBuffer:
def __init__(self, maxlen=1000):
self.buffer = deque(maxlen=maxlen)
def add_log(self, record):
self.buffer.append({
'timestamp': record.get('asctime'),
'level': record.get('levelname'),
'message': record.get('message'),
'extra': {k: v for k, v in record.items()
if k not in ['asctime', 'levelname', 'message', 'name']}
})
def get_logs(self, limit=100):
return list(self.buffer)[-limit:]
# Create global log buffer
log_buffer = LogBuffer()
# Add buffer handler to logger
# Add buffer handler to logger ONLY ONCE
buffer_handler = BufferHandler()
logger.addHandler(buffer_handler)
# Custom handler to capture logs
class BufferHandler(logging.Handler):
def emit(self, record):
try:
# Format the record as JSON
formatter = jsonlogger.JsonFormatter()
log_entry = json.loads(formatter.format(record))
log_buffer.add_log(log_entry)
except Exception:
pass
# --- FastAPI Application ---
app = FastAPI(
@ -175,9 +187,6 @@ async def update_node_status(
}
)
# In a real implementation, you would now update the RRD database here.
# e.g., database.update(node_uuid, status_update)
# For now, we simulate the logic.
if service_uuid != SERVICE_UUID:
logger.warning(
"Node sent status to wrong service UUID",
@ -185,7 +194,30 @@ async def update_node_status(
)
return {"error": "Service UUID mismatch", "peers": []}
# Auto-discovery logic: update our list of known nodes
# Update RRD database with system metrics
try:
database.update_system_metrics(
node_uuid=node_uuid,
timestamp=status_update.timestamp,
uptime_seconds=status_update.status.uptime_seconds,
load_avg=status_update.status.load_avg,
memory_usage_percent=status_update.status.memory_usage_percent
)
# Update ping metrics
for target_uuid, latency in status_update.pings.pings.items():
database.update_ping_metrics(
node_uuid=node_uuid,
target_uuid=target_uuid,
timestamp=status_update.timestamp,
latency_ms=latency
)
except Exception as e:
logger.error(f"Database update failed: {e}")
# Continue processing even if DB update fails
# Auto-discovery logic
if node_uuid not in known_nodes_db:
logger.info(f"New node discovered: {node_uuid}")
# A real app would need a strategy to handle node addresses