Documentation Index
Fetch the complete documentation index at: https://docs.llmgrid.ai/llms.txt
Use this file to discover all available pages before exploring further.
What is LLMGrid?
LLMGrid is an enterprise‑ready AI gateway and orchestration platform that provides a single control plane for using large language models, tools, and agents. It centralizes access, governance, routing, safety, and observability—without requiring application rewrites. LLMGrid exposes an OpenAI‑compatible API, so existing SDKs and frameworks work by simply pointing to the LLMGrid proxy.
Core Capabilities
OpenAI‑Compatible Proxy
- Drop‑in replacement for OpenAI SDKs
- Minimal code changes (update
base_urland API key) - Supports chat, embeddings, streaming, tools, and function calling
Model & Traffic Routing
- Route requests across models and providers
- Configure fallbacks and retry strategies
- Use aliases to keep application code stable during model changes
- Optimize for availability, latency, or efficiency
Governance & Safety
- Virtual Keys for scoped access and limits
- Guardrails for input/output and tool enforcement
- Budgets and rate limits to prevent overuse
- Tags for routing, attribution, and segmentation
Agents, Tools & Retrieval
- Register Agents with skills and capabilities
- Attach Search Tools for live retrieval and grounding
- Manage Vector Stores for RAG workflows
- Secure tool execution with pre‑execution checks
Observability & Analytics
- Request and audit logs
- Usage and cost analytics by model, key, team, tag, or agent
- Cache analytics and health checks
- End‑to‑end visibility for debugging and audits
Performance & Efficiency
- Response Caching (Redis‑backed) to reduce latency and repeat calls
- Semantic caching for similarity‑based reuse
- Centralized cost tracking and discounts
How LLMGrid Fits In
LLMGrid sits between your applications and AI capabilities: Your App ↓ LLMGrid (Auth • Routing • Guardrails • Observability) ↓ Models • Tools • Vector Stores • Search This architecture lets teams evolve models and controls independently of application code.Who Should Use LLMGrid?
- Platform teams needing centralized governance
- Developers shipping AI features quickly
- Security & compliance teams enforcing policies
- FinOps teams monitoring usage and cost
- Enterprises running multi‑model, multi‑tool AI workloads
Getting Started
- Create a Virtual Key
- Point your OpenAI SDK to the LLMGrid proxy
- Configure models and routing
- Add guardrails, budgets, and observability
- Iterate safely as usage grows
Related Sections
- API Reference – OpenAI‑compatible endpoints
- Models – Configure available models
- Router Settings – Control routing and fallbacks
- Guardrails – Enforce safety and compliance
- Usage & Logs – Observe and analyze traffic
- Security & Compliance – Enterprise controls and governance

