Introduction
Carrot captures traces from your LLM usage, builds custom models, and serves them through an OpenAI-compatible API.
Carrot is an LLM platform that helps you go from off-the-shelf models to custom models optimized for your specific use case.
The workflow
- Trace — Install the Python SDK or add a header to your existing calls. Carrot automatically captures every LLM interaction — inputs, outputs, token usage, and latency.
- Build — Once enough data is collected, Carrot Labs builds a custom model optimized for your workload.
- Deploy — Your custom model is served through the same OpenAI-compatible API. Switch to it by changing the model name — no code changes needed.
What you get
- Observability — Full visibility into every LLM call across your application. Browse traces, filter by model or status, and debug issues in the dashboard.
- Custom models — Purpose-built models trained on your real production data, delivering better quality at lower cost.
- OpenAI-compatible API — Use the standard OpenAI SDK. Change
base_urlandmodel, and everything works. - Zero-overhead tracing — The SDK captures traces in the background with no impact on your application's latency.
- Streaming support — Full streaming support for both tracing and inference.