This repository has been archived on 2026-02-15. You can view files and clone it. You cannot open issues or pull requests or push a commit.
Files
breakpilot-pwa/ai-compliance-sdk/internal/llm/ollama_adapter.go
Benjamin Admin 21a844cb8a fix: Restore all files lost during destructive rebase
A previous `git pull --rebase origin main` dropped 177 local commits,
losing 3400+ files across admin-v2, backend, studio-v2, website,
klausur-service, and many other services. The partial restore attempt
(660295e2) only recovered some files.

This commit restores all missing files from pre-rebase ref 98933f5e
while preserving post-rebase additions (night-scheduler, night-mode UI,
NightModeWidget dashboard integration).

Restored features include:
- AI Module Sidebar (FAB), OCR Labeling, OCR Compare
- GPU Dashboard, RAG Pipeline, Magic Help
- Klausur-Korrektur (8 files), Abitur-Archiv (5+ files)
- Companion, Zeugnisse-Crawler, Screen Flow
- Full backend, studio-v2, website, klausur-service
- All compliance SDKs, agent-core, voice-service
- CI/CD configs, documentation, scripts

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-02-09 09:51:32 +01:00

351 lines
8.7 KiB
Go

package llm
import (
"bytes"
"context"
"encoding/json"
"fmt"
"io"
"net/http"
"time"
"github.com/google/uuid"
)
// OllamaAdapter implements the Provider interface for Ollama
type OllamaAdapter struct {
baseURL string
defaultModel string
httpClient *http.Client
}
// NewOllamaAdapter creates a new Ollama adapter
func NewOllamaAdapter(baseURL, defaultModel string) *OllamaAdapter {
return &OllamaAdapter{
baseURL: baseURL,
defaultModel: defaultModel,
httpClient: &http.Client{
Timeout: 5 * time.Minute, // LLM requests can be slow
},
}
}
// Name returns the provider name
func (o *OllamaAdapter) Name() string {
return ProviderOllama
}
// IsAvailable checks if Ollama is reachable
func (o *OllamaAdapter) IsAvailable(ctx context.Context) bool {
req, err := http.NewRequestWithContext(ctx, "GET", o.baseURL+"/api/tags", nil)
if err != nil {
return false
}
ctx, cancel := context.WithTimeout(ctx, 5*time.Second)
defer cancel()
req = req.WithContext(ctx)
resp, err := o.httpClient.Do(req)
if err != nil {
return false
}
defer resp.Body.Close()
return resp.StatusCode == http.StatusOK
}
// ListModels returns available Ollama models
func (o *OllamaAdapter) ListModels(ctx context.Context) ([]Model, error) {
req, err := http.NewRequestWithContext(ctx, "GET", o.baseURL+"/api/tags", nil)
if err != nil {
return nil, err
}
resp, err := o.httpClient.Do(req)
if err != nil {
return nil, fmt.Errorf("failed to list models: %w", err)
}
defer resp.Body.Close()
if resp.StatusCode != http.StatusOK {
return nil, fmt.Errorf("unexpected status: %d", resp.StatusCode)
}
var result struct {
Models []struct {
Name string `json:"name"`
ModifiedAt string `json:"modified_at"`
Size int64 `json:"size"`
Details struct {
Format string `json:"format"`
Family string `json:"family"`
ParameterSize string `json:"parameter_size"`
QuantizationLevel string `json:"quantization_level"`
} `json:"details"`
} `json:"models"`
}
if err := json.NewDecoder(resp.Body).Decode(&result); err != nil {
return nil, fmt.Errorf("failed to decode response: %w", err)
}
models := make([]Model, len(result.Models))
for i, m := range result.Models {
models[i] = Model{
ID: m.Name,
Name: m.Name,
Provider: ProviderOllama,
Description: fmt.Sprintf("%s (%s)", m.Details.Family, m.Details.ParameterSize),
ContextSize: 4096, // Default, actual varies by model
Capabilities: []string{"chat", "completion"},
}
}
return models, nil
}
// Complete performs text completion
func (o *OllamaAdapter) Complete(ctx context.Context, req *CompletionRequest) (*CompletionResponse, error) {
model := req.Model
if model == "" {
model = o.defaultModel
}
start := time.Now()
ollamaReq := map[string]any{
"model": model,
"prompt": req.Prompt,
"stream": false,
}
if req.MaxTokens > 0 {
if ollamaReq["options"] == nil {
ollamaReq["options"] = make(map[string]any)
}
ollamaReq["options"].(map[string]any)["num_predict"] = req.MaxTokens
}
if req.Temperature > 0 {
if ollamaReq["options"] == nil {
ollamaReq["options"] = make(map[string]any)
}
ollamaReq["options"].(map[string]any)["temperature"] = req.Temperature
}
body, err := json.Marshal(ollamaReq)
if err != nil {
return nil, err
}
httpReq, err := http.NewRequestWithContext(ctx, "POST", o.baseURL+"/api/generate", bytes.NewReader(body))
if err != nil {
return nil, err
}
httpReq.Header.Set("Content-Type", "application/json")
resp, err := o.httpClient.Do(httpReq)
if err != nil {
return nil, fmt.Errorf("ollama request failed: %w", err)
}
defer resp.Body.Close()
if resp.StatusCode != http.StatusOK {
bodyBytes, _ := io.ReadAll(resp.Body)
return nil, fmt.Errorf("ollama error: %s", string(bodyBytes))
}
var result struct {
Model string `json:"model"`
Response string `json:"response"`
Done bool `json:"done"`
TotalDuration int64 `json:"total_duration"`
LoadDuration int64 `json:"load_duration"`
PromptEvalCount int `json:"prompt_eval_count"`
PromptEvalDuration int64 `json:"prompt_eval_duration"`
EvalCount int `json:"eval_count"`
EvalDuration int64 `json:"eval_duration"`
}
if err := json.NewDecoder(resp.Body).Decode(&result); err != nil {
return nil, fmt.Errorf("failed to decode response: %w", err)
}
duration := time.Since(start)
return &CompletionResponse{
ID: uuid.New().String(),
Model: result.Model,
Provider: ProviderOllama,
Text: result.Response,
FinishReason: "stop",
Usage: UsageStats{
PromptTokens: result.PromptEvalCount,
CompletionTokens: result.EvalCount,
TotalTokens: result.PromptEvalCount + result.EvalCount,
},
Duration: duration,
}, nil
}
// Chat performs chat completion
func (o *OllamaAdapter) Chat(ctx context.Context, req *ChatRequest) (*ChatResponse, error) {
model := req.Model
if model == "" {
model = o.defaultModel
}
start := time.Now()
// Convert messages to Ollama format
messages := make([]map[string]string, len(req.Messages))
for i, m := range req.Messages {
messages[i] = map[string]string{
"role": m.Role,
"content": m.Content,
}
}
ollamaReq := map[string]any{
"model": model,
"messages": messages,
"stream": false,
}
if req.MaxTokens > 0 {
if ollamaReq["options"] == nil {
ollamaReq["options"] = make(map[string]any)
}
ollamaReq["options"].(map[string]any)["num_predict"] = req.MaxTokens
}
if req.Temperature > 0 {
if ollamaReq["options"] == nil {
ollamaReq["options"] = make(map[string]any)
}
ollamaReq["options"].(map[string]any)["temperature"] = req.Temperature
}
body, err := json.Marshal(ollamaReq)
if err != nil {
return nil, err
}
httpReq, err := http.NewRequestWithContext(ctx, "POST", o.baseURL+"/api/chat", bytes.NewReader(body))
if err != nil {
return nil, err
}
httpReq.Header.Set("Content-Type", "application/json")
resp, err := o.httpClient.Do(httpReq)
if err != nil {
return nil, fmt.Errorf("ollama chat request failed: %w", err)
}
defer resp.Body.Close()
if resp.StatusCode != http.StatusOK {
bodyBytes, _ := io.ReadAll(resp.Body)
return nil, fmt.Errorf("ollama chat error: %s", string(bodyBytes))
}
var result struct {
Model string `json:"model"`
Message struct {
Role string `json:"role"`
Content string `json:"content"`
} `json:"message"`
Done bool `json:"done"`
TotalDuration int64 `json:"total_duration"`
PromptEvalCount int `json:"prompt_eval_count"`
EvalCount int `json:"eval_count"`
}
if err := json.NewDecoder(resp.Body).Decode(&result); err != nil {
return nil, fmt.Errorf("failed to decode chat response: %w", err)
}
duration := time.Since(start)
return &ChatResponse{
ID: uuid.New().String(),
Model: result.Model,
Provider: ProviderOllama,
Message: Message{
Role: result.Message.Role,
Content: result.Message.Content,
},
FinishReason: "stop",
Usage: UsageStats{
PromptTokens: result.PromptEvalCount,
CompletionTokens: result.EvalCount,
TotalTokens: result.PromptEvalCount + result.EvalCount,
},
Duration: duration,
}, nil
}
// Embed creates embeddings
func (o *OllamaAdapter) Embed(ctx context.Context, req *EmbedRequest) (*EmbedResponse, error) {
model := req.Model
if model == "" {
model = "nomic-embed-text" // Default embedding model
}
start := time.Now()
var embeddings [][]float64
for _, input := range req.Input {
ollamaReq := map[string]any{
"model": model,
"prompt": input,
}
body, err := json.Marshal(ollamaReq)
if err != nil {
return nil, err
}
httpReq, err := http.NewRequestWithContext(ctx, "POST", o.baseURL+"/api/embeddings", bytes.NewReader(body))
if err != nil {
return nil, err
}
httpReq.Header.Set("Content-Type", "application/json")
resp, err := o.httpClient.Do(httpReq)
if err != nil {
return nil, fmt.Errorf("ollama embedding request failed: %w", err)
}
defer resp.Body.Close()
if resp.StatusCode != http.StatusOK {
bodyBytes, _ := io.ReadAll(resp.Body)
return nil, fmt.Errorf("ollama embedding error: %s", string(bodyBytes))
}
var result struct {
Embedding []float64 `json:"embedding"`
}
if err := json.NewDecoder(resp.Body).Decode(&result); err != nil {
return nil, fmt.Errorf("failed to decode embedding response: %w", err)
}
embeddings = append(embeddings, result.Embedding)
}
duration := time.Since(start)
return &EmbedResponse{
ID: uuid.New().String(),
Model: model,
Provider: ProviderOllama,
Embeddings: embeddings,
Usage: UsageStats{
TotalTokens: len(req.Input) * 256, // Approximate
},
Duration: duration,
}, nil
}