Model Prism
Multi-tenant LLM gateway with intelligent routing and cost control. Drop-in replacement for the OpenAI API — for your entire organization.
Everything you need
Model Prism bundles all critical features of a production-ready LLM gateway into a single, easily deployable service.
Intelligent Auto-Routing
Classifier-based routing automatically directs each request to the optimal model — based on complexity, context, and configured cost tiers. Rule sets and fallbacks for maximum control.
Multi-Tenant & RBAC
Full tenant isolation: Every team, every customer gets their own API keys, quotas, and permissions. RBAC at tenant and model level, LDAP/SSO integration.
Real-Time Cost Control
Token-accurate cost tracking per tenant, model, and time period. Budget alerts, automatic throttling on overage, and detailed analytics dashboards.
OpenAI-Compatible API
Drop-in replacement for the OpenAI API — no code changes required. Supports Chat Completions, Embeddings, and Function Calling. Compatible with every OpenAI SDK.
Prompt Logging & Audit Trail
Complete audit log of all LLM requests — prompt, response, model, tokens, cost, and timestamp. Exportable for compliance and debugging.
Air-gapped & Scalable
Fully offline with OFFLINE=true — no outbound connections, no telemetry. Stateless architecture for horizontal scaling, Kubernetes-ready with Helm chart.
Up and running in 5 minutes
Start Model Prism instantly with Docker Compose. No database setup, no complex configuration.
docker pull ghcr.io/ai-ohara-systems/model-prism:latest # 1. Clone the repository
git clone https://github.com/ai-ohara-systems/model-prism
cd model-prism
# 2. Start
docker compose up -d
# → Setup wizard: http://localhost:3000
# 3. Test — drop-in for OpenAI API
curl http://localhost:3000/api/api/v1/chat/completions \
-H "Authorization: Bearer omp-your-api-key" \
-H "Content-Type: application/json" \
-d '{"model": "auto", "messages": [{"role": "user", "content": "Hello!"}]}' Supported Providers
Model Prism connects to all major LLM providers through unified adapters.
+ any OpenAI-compatible endpoint (vLLM, LM Studio, LocalAI, ...)
Ready to deploy Model Prism?
Source Available (ELv2), production-ready. Fully air-gapped with OFFLINE=true — no outbound connections. Enterprise license available for advanced features.
Model Prism on GitHub