fix: Better error logging + increase LLM timeout to 120s for RAG check

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
This commit is contained in:
Benjamin Admin
2026-05-06 14:33:58 +02:00
parent 510f513811
commit 7b8440191e
@@ -194,11 +194,12 @@ async def _verify_control_with_llm(
)
try:
async with httpx.AsyncClient(timeout=60.0) as client:
async with httpx.AsyncClient(timeout=120.0) as client:
resp = await client.post(f"{OLLAMA_URL}/api/generate", json={
"model": OLLAMA_MODEL,
"prompt": prompt,
"stream": False,
"options": {"num_predict": 200}, # Limit response length
})
if resp.status_code != 200:
@@ -229,5 +230,5 @@ async def _verify_control_with_llm(
}
except Exception as e:
logger.warning("LLM verification failed: %s", e)
logger.warning("LLM verification failed: %s %s", type(e).__name__, e)
return None