A FastAPI gateway for local LLMs that adds intelligent web research, multilingual recency/how-to detection, time-anchored guidance, context injection, and OpenAI-compatible SSE streaming. Turn any local model into a recency-aware, context-enhanced assistant instantly.
multilingual python streaming aurelia proxy gateway sse non-commercial rag fastapi source-available context-injection llm local-llm web-research openai-compatible time-anchoring
-
Updated
Nov 20, 2025 - Python