footer{text-align:center;padding:2rem;font-size:0.75rem;font-family:var(--font-mono);color:var(--cream-muted);border-top:1px solid var(--bg3)} footer a{color:var(--cream-muted)}footer a:hover{color:var(--cream)} .sig{font-family:var(--font-serif);font-style:italic;color:var(--leather-light);font-size:0.9rem}
Point your app at Stockyard and route requests to OpenAI, Anthropic, Google, Ollama, and 12 more providers through a single stable endpoint. No Redis, no Postgres, no Docker.
That request just passed through 76 middleware modules. Traced, logged, audited — automatically.
Cost, latency, token count, model, and provider — recorded automatically. Query traces through the API or browse them in the dashboard.
Free — LookoutEvery request gets a hash-chained ledger entry. You can prove what was sent, what came back, and that nothing was altered.
Free — BrandRate limiting, token budgets, content filtering, PII redaction, prompt injection detection — all configurable per-route, all running in the middleware chain.
Free — 76 modulesIf a provider goes down, Stockyard fails over to the next one automatically. Model-aware routing sends Claude requests to Anthropic first, GPT requests to OpenAI first.
Free — ChuteMap stable names like "fast" or "smart" to real models underneath. Swap providers later without touching app code.
Free — Alias APIStockyard ships with 150 tools and 76 modules, but you don't need to configure any of them. Install it, point your app at it, and the proxy works out of the box. Tracing, auditing, and controls are automatic. Everything else is optional — turn it on when a real need shows up.
Most people start with just the proxy. That's the whole idea.
What you gain: one stable endpoint for 16 providers, automatic request tracing and cost tracking, response caching, rate limiting, model aliasing, failover routing, and a web console. All from one binary with no external dependencies.
What you give up: roughly 200ms of added latency per request (the middleware chain). Your requests route through one more hop. If you need sub-millisecond proxy overhead or horizontal scaling across many nodes, a lighter proxy or direct provider calls may be a better fit.
OpenAI, Anthropic, Google Gemini, Mistral, Cohere, Groq, Together AI, Fireworks, Perplexity, DeepSeek, Ollama, LM Studio, Azure OpenAI, AWS Bedrock, OpenRouter, and any OpenAI-compatible endpoint.
One binary. One SQLite file. Every request traced. Install in under a minute, no credit card needed.
Install Stockyard