platform/openbao/README.md ingress hosts (line 108) had `bao.<domain>` while
the same file's ClusterSecretStore example (line 127) used the canonical
`bao.<location-code>.<sovereign-domain>` form. Pass 7's active-active fix
addressed the body but missed the ingress placeholder. Aligned with the
canonical form.
platform/librechat/README.md OAuth callback (line 154) had
`chat.ai-hub.<domain>/oauth/openid/callback` — same Application-endpoint
shape Pass 25 fixed in llm-gateway. Pass 22 marked the file clean and Pass
29 fixed the Keycloak issuer line but didn't re-sweep. Per NAMING §5.2
Application endpoints are `{app}.{environment}.{sovereign-domain}`. Fixed.
docs/GLOSSARY.md verified clean — single-source-of-truth has held across
the loop (Pass 6/7/14/20/22/26/27 all consistent with current GLOSSARY).
Validation log Pass 31 entry includes meta-note: third file (librechat)
that needed re-opening after a "clean" mark — banner scans miss YAML-block
drift. Future passes should default to a full placeholder-shape grep on
every file touched.
|
||
|---|---|---|
| .. | ||
| README.md | ||
LibreChat
Open-source chat UI with multi-model support and file uploads. Application Blueprint (see docs/PLATFORM-TECH-STACK.md §4.6). Default end-user chat surface in bp-cortex — fronts the LLM Gateway and routes through NeMo Guardrails for safety.
Status: Accepted | Updated: 2026-04-27
Overview
LibreChat provides a ChatGPT-like interface supporting multiple AI backends, file uploads, and customizable agent presets.
flowchart LR
subgraph LibreChat["LibreChat"]
UI[Chat UI]
Presets[Agent Presets]
Files[File Handling]
end
subgraph Backends["AI Backends"]
OpenAI[OpenAI API]
Custom[Custom Endpoints]
RAG[RAG Service]
end
subgraph Storage["Storage"]
FerretDB[FerretDB]
FileStore[File Storage]
end
User[User] --> UI
UI --> Presets
UI --> Files
Presets --> Backends
Files --> FileStore
UI --> FerretDB
Why LibreChat?
| Feature | Benefit |
|---|---|
| Multi-model | Switch between AI backends |
| Agent presets | Pre-configured assistants |
| File uploads | Document analysis |
| Conversation history | Persistent chat storage |
| SSO integration | Enterprise authentication |
Configuration
Helm Values
librechat:
replicas: 2
config:
endpoints:
custom:
- name: "AI Hub"
apiKey: "${RAG_SERVICE_API_KEY}"
baseURL: "http://rag-service.ai-hub.svc:8000/v1"
models:
default: ["deep-thinker", "quick-thinker", "compliance-advisor"]
titleModel: "quick-thinker"
dropParams: ["stop", "user"]
registration:
socialLogins: ["openid"]
fileConfig:
endpoints:
custom:
fileLimit: 10
fileSizeLimit: 50 # MB
supportedMimeTypes:
- "application/pdf"
- "application/vnd.openxmlformats-officedocument.wordprocessingml.document"
- "text/plain"
ferretdb:
enabled: true
# FerretDB provides MongoDB wire protocol compatibility
# backed by CNPG PostgreSQL
auth:
rootPassword: "" # From ESO
persistence:
size: 10Gi
Agent Presets
Deep Thinker
{
"name": "Deep Thinker",
"model": "deep-thinker",
"description": "Complex reasoning with visible chain-of-thought",
"systemPrompt": "You are a thoughtful analyst. Think step by step and show your reasoning.",
"temperature": 0.7,
"maxTokens": 8192
}
Quick Thinker
{
"name": "Quick Thinker",
"model": "quick-thinker",
"description": "Fast responses for simple queries",
"systemPrompt": "You are a helpful assistant. Be concise and direct.",
"temperature": 0.3,
"maxTokens": 2048
}
Compliance Advisor
{
"name": "Compliance Advisor",
"model": "compliance-advisor",
"description": "Regulatory knowledge with citations",
"systemPrompt": "You are a compliance expert. Always cite your sources with document references.",
"temperature": 0.1,
"maxTokens": 4096
}
SSO Configuration
Azure AD OIDC
socialLogins:
- openid
openidConfig:
issuer: "https://login.microsoftonline.com/${TENANT_ID}/v2.0"
clientId: "${CLIENT_ID}"
clientSecret: "${CLIENT_SECRET}"
scope: ["openid", "profile", "email"]
callbackURL: "https://chat.<env>.<sovereign-domain>/oauth/openid/callback"
Keycloak
openidConfig:
issuer: "https://keycloak.<location-code>.<sovereign-domain>/realms/<org>"
clientId: "librechat"
clientSecret: "" # From ESO
scope: ["openid", "profile", "email"]
File Upload Flow
sequenceDiagram
participant User
participant LibreChat
participant RAG as RAG Service
participant Milvus
User->>LibreChat: Upload PDF
LibreChat->>RAG: POST /ingest/file
RAG->>RAG: Parse & chunk
RAG->>Milvus: Store vectors (ephemeral)
RAG-->>LibreChat: file_id
User->>LibreChat: Ask question about file
LibreChat->>RAG: Query with file_id context
RAG->>Milvus: Search ephemeral partition
RAG-->>LibreChat: Response with citations
Environment Variables
| Variable | Purpose |
|---|---|
MONGO_URI |
FerretDB connection string (MongoDB wire protocol) |
OPENID_CLIENT_ID |
SSO client ID |
OPENID_CLIENT_SECRET |
SSO client secret |
CREDS_KEY |
Encryption key for credentials |
CREDS_IV |
Encryption IV |
JWT_SECRET |
JWT signing secret |
Custom Endpoints
endpoints:
custom:
- name: "RAG Service"
baseURL: "http://rag-service.ai-hub.svc:8000/v1"
apiKey: "${API_KEY}"
models:
default:
- deep-thinker
- quick-thinker
- compliance-advisor
- aiops-advisor
- dev-advisor
- internet-search
Monitoring
| Metric | Description |
|---|---|
| Active users | Concurrent chat sessions |
| Message count | Total messages sent |
| File uploads | Documents processed |
| Response time | Backend latency |
Consequences
Positive:
- ChatGPT-like experience
- Multi-model switching
- File upload support
- Enterprise SSO
- Customizable presets
Negative:
- Requires FerretDB (MongoDB wire protocol on CNPG)
- Complex configuration
- UI customization limited
Part of OpenOva