Phase 1 — Python (klausur-service): 5 monoliths → 36 files - dsfa_corpus_ingestion.py (1,828 LOC → 5 files) - cv_ocr_engines.py (2,102 LOC → 7 files) - cv_layout.py (3,653 LOC → 10 files) - vocab_worksheet_api.py (2,783 LOC → 8 files) - grid_build_core.py (1,958 LOC → 6 files) Phase 2 — Go (edu-search-service, school-service): 8 monoliths → 19 files - staff_crawler.go (1,402 → 4), policy/store.go (1,168 → 3) - policy_handlers.go (700 → 2), repository.go (684 → 2) - search.go (592 → 2), ai_extraction_handlers.go (554 → 2) - seed_data.go (591 → 2), grade_service.go (646 → 2) Phase 3 — TypeScript (admin-lehrer): 45 monoliths → 220+ files - sdk/types.ts (2,108 → 16 domain files) - ai/rag/page.tsx (2,686 → 14 files) - 22 page.tsx files split into _components/ + _hooks/ - 11 component files split into sub-components - 10 SDK data catalogs added to loc-exceptions - Deleted dead backup index_original.ts (4,899 LOC) All original public APIs preserved via re-export facades. Zero new errors: Python imports verified, Go builds clean, TypeScript tsc --noEmit shows only pre-existing errors. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
94 lines
2.7 KiB
TypeScript
94 lines
2.7 KiB
TypeScript
import { useState, useCallback } from 'react'
|
|
import type { ChunkDetail, TraceabilityResult } from './types'
|
|
import { API_PROXY } from './types'
|
|
|
|
export function useQualitySearch() {
|
|
const [searchQuery, setSearchQuery] = useState('')
|
|
const [searchResults, setSearchResults] = useState<ChunkDetail[]>([])
|
|
const [searching, setSearching] = useState(false)
|
|
const [selectedRegulation, setSelectedRegulation] = useState<string>('')
|
|
const [topK, setTopK] = useState(10)
|
|
|
|
const [selectedChunk, setSelectedChunk] = useState<ChunkDetail | null>(null)
|
|
const [traceability, setTraceability] = useState<TraceabilityResult | null>(null)
|
|
const [loadingTrace, setLoadingTrace] = useState(false)
|
|
|
|
const handleSearch = useCallback(async () => {
|
|
if (!searchQuery.trim()) return
|
|
|
|
setSearching(true)
|
|
setSearchResults([])
|
|
setSelectedChunk(null)
|
|
setTraceability(null)
|
|
|
|
try {
|
|
let url = `${API_PROXY}?action=search&query=${encodeURIComponent(searchQuery)}&top_k=${topK}`
|
|
if (selectedRegulation) {
|
|
url += `®ulations=${encodeURIComponent(selectedRegulation)}`
|
|
}
|
|
|
|
const res = await fetch(url)
|
|
if (res.ok) {
|
|
const data = await res.json()
|
|
setSearchResults(data.results || [])
|
|
}
|
|
} catch (error) {
|
|
console.error('Search failed:', error)
|
|
} finally {
|
|
setSearching(false)
|
|
}
|
|
}, [searchQuery, selectedRegulation, topK])
|
|
|
|
const loadTraceability = useCallback(async (chunk: ChunkDetail) => {
|
|
setSelectedChunk(chunk)
|
|
setLoadingTrace(true)
|
|
|
|
try {
|
|
const res = await fetch(
|
|
`${API_PROXY}?action=traceability&chunk_id=${encodeURIComponent(chunk.id || chunk.regulation_code + '_' + chunk.chunk_index)}®ulation=${encodeURIComponent(chunk.regulation_code)}`
|
|
)
|
|
|
|
if (res.ok) {
|
|
const data = await res.json()
|
|
setTraceability({
|
|
chunk,
|
|
requirements: data.requirements || [],
|
|
controls: data.controls || [],
|
|
})
|
|
} else {
|
|
setTraceability({ chunk, requirements: [], controls: [] })
|
|
}
|
|
} catch (error) {
|
|
console.error('Failed to load traceability:', error)
|
|
setTraceability({ chunk, requirements: [], controls: [] })
|
|
} finally {
|
|
setLoadingTrace(false)
|
|
}
|
|
}, [])
|
|
|
|
const handleSampleQuery = (query: string, reg: string) => {
|
|
setSearchQuery(query)
|
|
setSelectedRegulation(reg)
|
|
setTimeout(() => {
|
|
handleSearch()
|
|
}, 100)
|
|
}
|
|
|
|
return {
|
|
searchQuery,
|
|
setSearchQuery,
|
|
searchResults,
|
|
searching,
|
|
selectedRegulation,
|
|
setSelectedRegulation,
|
|
topK,
|
|
setTopK,
|
|
selectedChunk,
|
|
traceability,
|
|
loadingTrace,
|
|
handleSearch,
|
|
loadTraceability,
|
|
handleSampleQuery,
|
|
}
|
|
}
|