feat(control-pipeline): add LLM dedup endpoint for borderline review queue
POST /v1/canonical/generate/llm-dedup uses local Ollama (qwen3.5:35b-a3b) to verify borderline duplicate matches (score 0.85-0.91). More accurate than embedding similarity for compliance controls with subtle scope differences (e.g. "documented" vs "implemented"). Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
This commit is contained in:
@@ -1402,6 +1402,187 @@ async def get_batch_dedup_status(dedup_id: str):
|
||||
return status
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# LLM DEDUP REVIEW — Local LLM verifies borderline duplicates
|
||||
# =============================================================================
|
||||
|
||||
class LLMDedupRequest(BaseModel):
|
||||
dry_run: bool = True
|
||||
limit: int = 100
|
||||
min_score: float = 0.85 # Only review entries >= this score
|
||||
max_score: float = 0.91 # Only review entries < this score (0.91+ already handled)
|
||||
model: str = "qwen3.5:35b-a3b"
|
||||
|
||||
|
||||
_llm_dedup_status: dict = {}
|
||||
|
||||
|
||||
async def _run_llm_dedup(req: LLMDedupRequest, job_id: str):
|
||||
"""Use local LLM to verify borderline dedup matches."""
|
||||
import httpx
|
||||
import os
|
||||
|
||||
OLLAMA_URL = os.getenv("OLLAMA_URL", "http://host.docker.internal:11434")
|
||||
|
||||
db = SessionLocal()
|
||||
try:
|
||||
# Load review entries in the score band
|
||||
limit_clause = f"LIMIT {req.limit}" if req.limit > 0 else ""
|
||||
rows = db.execute(text(f"""
|
||||
SELECT r.id, r.candidate_control_id, r.candidate_title,
|
||||
r.matched_control_id, r.similarity_score,
|
||||
c1.objective as candidate_objective,
|
||||
c1.requirements::text as candidate_requirements,
|
||||
c2.title as matched_title,
|
||||
c2.objective as matched_objective,
|
||||
c2.requirements::text as matched_requirements
|
||||
FROM compliance.control_dedup_reviews r
|
||||
LEFT JOIN compliance.canonical_controls c1 ON c1.control_id = r.candidate_control_id
|
||||
LEFT JOIN compliance.canonical_controls c2 ON c2.control_id = r.matched_control_id
|
||||
WHERE r.dedup_stage = 'batch_dedup'
|
||||
AND r.similarity_score >= :min_score
|
||||
AND r.similarity_score < :max_score
|
||||
ORDER BY r.similarity_score DESC
|
||||
{limit_clause}
|
||||
"""), {"min_score": req.min_score, "max_score": req.max_score}).fetchall()
|
||||
|
||||
total = len(rows)
|
||||
duplicates = 0
|
||||
different = 0
|
||||
errors = 0
|
||||
results = []
|
||||
|
||||
_llm_dedup_status[job_id] = {
|
||||
"status": "running", "total": total, "processed": 0,
|
||||
"duplicates": 0, "different": 0, "errors": 0,
|
||||
"dry_run": req.dry_run,
|
||||
}
|
||||
|
||||
for i, row in enumerate(rows):
|
||||
try:
|
||||
# Build comparison prompt
|
||||
candidate_ctx = row.candidate_title or ""
|
||||
if row.candidate_objective:
|
||||
candidate_ctx += f"\nObjective: {row.candidate_objective[:300]}"
|
||||
if row.candidate_requirements and row.candidate_requirements not in ("[]", "null"):
|
||||
candidate_ctx += f"\nRequirements: {row.candidate_requirements[:300]}"
|
||||
|
||||
matched_ctx = row.matched_title or ""
|
||||
if row.matched_objective:
|
||||
matched_ctx += f"\nObjective: {row.matched_objective[:300]}"
|
||||
if row.matched_requirements and row.matched_requirements not in ("[]", "null"):
|
||||
matched_ctx += f"\nRequirements: {row.matched_requirements[:300]}"
|
||||
|
||||
prompt = f"Control A ({row.candidate_control_id}):\n{candidate_ctx}\n\nControl B ({row.matched_control_id}):\n{matched_ctx}\n\nSind diese Controls Duplikate?"
|
||||
|
||||
async with httpx.AsyncClient(timeout=30.0) as client:
|
||||
resp = await client.post(
|
||||
f"{OLLAMA_URL}/api/chat",
|
||||
json={
|
||||
"model": req.model,
|
||||
"stream": False,
|
||||
"messages": [
|
||||
{"role": "system", "content": "Du bist ein Compliance-Experte. Vergleiche zwei Controls und entscheide: DUPLIKAT (gleiche Anforderung, nur anders formuliert) oder VERSCHIEDEN (unterschiedlicher Scope/Inhalt). Antworte NUR mit einem JSON: {\"verdict\": \"DUPLIKAT\" oder \"VERSCHIEDEN\", \"reason\": \"kurze Begruendung\"}"},
|
||||
{"role": "user", "content": prompt},
|
||||
],
|
||||
},
|
||||
)
|
||||
|
||||
if resp.status_code != 200:
|
||||
errors += 1
|
||||
continue
|
||||
|
||||
content = resp.json().get("message", {}).get("content", "")
|
||||
# Parse verdict from response
|
||||
parsed = _parse_llm_json(content)
|
||||
if not parsed:
|
||||
errors += 1
|
||||
continue
|
||||
|
||||
verdict = parsed.get("verdict", "").upper()
|
||||
reason = parsed.get("reason", "")
|
||||
|
||||
if "DUPLIKAT" in verdict:
|
||||
duplicates += 1
|
||||
if not req.dry_run:
|
||||
# Mark as duplicate
|
||||
db.execute(text("""
|
||||
UPDATE compliance.canonical_controls
|
||||
SET release_state = 'duplicate',
|
||||
merged_into_uuid = (SELECT id FROM compliance.canonical_controls WHERE control_id = :matched LIMIT 1),
|
||||
updated_at = NOW()
|
||||
WHERE control_id = :candidate AND release_state = 'draft'
|
||||
"""), {"candidate": row.candidate_control_id, "matched": row.matched_control_id})
|
||||
else:
|
||||
different += 1
|
||||
|
||||
results.append({
|
||||
"candidate": row.candidate_control_id,
|
||||
"matched": row.matched_control_id,
|
||||
"score": float(row.similarity_score),
|
||||
"verdict": verdict,
|
||||
"reason": reason[:100],
|
||||
})
|
||||
|
||||
except Exception as e:
|
||||
errors += 1
|
||||
logger.warning("LLM dedup error for %s: %s", row.candidate_control_id, e)
|
||||
|
||||
if not req.dry_run and (i + 1) % 50 == 0:
|
||||
db.commit()
|
||||
|
||||
_llm_dedup_status[job_id] = {
|
||||
"status": "running", "total": total, "processed": i + 1,
|
||||
"duplicates": duplicates, "different": different, "errors": errors,
|
||||
"dry_run": req.dry_run,
|
||||
}
|
||||
|
||||
if not req.dry_run:
|
||||
db.commit()
|
||||
|
||||
_llm_dedup_status[job_id] = {
|
||||
"status": "completed", "total": total, "processed": total,
|
||||
"duplicates": duplicates, "different": different, "errors": errors,
|
||||
"dry_run": req.dry_run, "results": results[-50:],
|
||||
}
|
||||
logger.info("LLM dedup %s completed: %d dup, %d diff, %d err out of %d",
|
||||
job_id, duplicates, different, errors, total)
|
||||
|
||||
except Exception as e:
|
||||
logger.error("LLM dedup %s failed: %s", job_id, e)
|
||||
_llm_dedup_status[job_id] = {"status": "failed", "error": str(e)}
|
||||
finally:
|
||||
db.close()
|
||||
|
||||
|
||||
@router.post("/generate/llm-dedup")
|
||||
async def start_llm_dedup(req: LLMDedupRequest):
|
||||
"""Use local LLM to verify borderline dedup matches from the review queue.
|
||||
|
||||
Sends each candidate+matched pair to the local Ollama LLM for a
|
||||
DUPLIKAT/VERSCHIEDEN verdict. Much more accurate than embedding similarity.
|
||||
Default is dry_run=True (preview only, no DB changes).
|
||||
"""
|
||||
import uuid
|
||||
job_id = str(uuid.uuid4())[:8]
|
||||
_llm_dedup_status[job_id] = {"status": "starting"}
|
||||
asyncio.create_task(_run_llm_dedup(req, job_id))
|
||||
return {
|
||||
"status": "running",
|
||||
"job_id": job_id,
|
||||
"message": f"LLM dedup started. Poll /generate/llm-dedup-status/{job_id}",
|
||||
}
|
||||
|
||||
|
||||
@router.get("/generate/llm-dedup-status/{job_id}")
|
||||
async def get_llm_dedup_status(job_id: str):
|
||||
"""Get status of an LLM dedup job."""
|
||||
status = _llm_dedup_status.get(job_id)
|
||||
if not status:
|
||||
raise HTTPException(status_code=404, detail="LLM dedup job not found")
|
||||
return status
|
||||
|
||||
|
||||
# =============================================================================
|
||||
# ANCHOR BACKFILL
|
||||
# =============================================================================
|
||||
|
||||
Reference in New Issue
Block a user