feat: replaced ollama with litellm (#18)
Some checks failed
CI / Format (push) Successful in 3s
CI / Clippy (push) Successful in 2m53s
CI / Security Audit (push) Successful in 1m42s
CI / Tests (push) Failing after 3m59s
CI / Deploy (push) Has been skipped
CI / E2E Tests (push) Has been skipped

Co-authored-by: Sharang Parnerkar <parnerkarsharang@gmail.com>
Reviewed-on: #18
This commit was merged in pull request #18.
This commit is contained in:
2026-02-26 17:52:47 +00:00
parent 0deaaca848
commit fe4f8e84ae
28 changed files with 1107 additions and 500 deletions

View File

@@ -1,5 +1,5 @@
# CERTifAI LibreChat Configuration
# Ollama backend for self-hosted LLM inference.
# LiteLLM proxy for unified multi-provider LLM access.
version: 1.2.8
cache: true
@@ -19,22 +19,16 @@ interface:
endpoints:
custom:
- name: "Ollama"
apiKey: "ollama"
baseURL: "https://mac-mini-von-benjamin-2:11434/v1/"
- name: "LiteLLM"
apiKey: "${LITELLM_API_KEY}"
baseURL: "https://llm-dev.meghsakha.com/v1/"
models:
default:
- "llama3.1:8b"
- "qwen3:30b-a3b"
- "Qwen3-Coder-30B-A3B-Instruct"
fetch: true
titleConvo: true
titleModel: "current_model"
summarize: false
summaryModel: "current_model"
forcePrompt: false
modelDisplayLabel: "CERTifAI Ollama"
dropParams:
- stop
- user
- frequency_penalty
- presence_penalty
modelDisplayLabel: "CERTifAI LiteLLM"