See everything. Miss nothing.
LLM monitoring built into the proxy. Every request traced automatically. Costs rolled up daily. Anomalies caught in real time. No extra tools, no extra config.
LLM monitoring built into the proxy. Every request traced automatically. Costs rolled up daily. Anomalies caught in real time. No extra tools, no extra config.
| Trace | Model | Tokens | Cost | Latency | Status |
|---|---|---|---|---|---|
| tr_8a4f2c | gpt-4o | 1,247 | $0.0187 | 842ms | 200 |
| tr_3b7e1a | claude-sonnet-4-20250514 | 3,891 | $0.0467 | 2.1s | 200 |
| tr_c91d5f | gpt-4o-mini | 412 | $0.0006 | 310ms | 200 |
| tr_f2a8b4 | gpt-4o | 0 | $0.00 | 28.3s | 429 |
| tr_71e3c9 | gpt-4o-mini | 876 | $0.0013 | 445ms | 200 |
Everything in Lookout is queryable via REST. Pull traces into your own dashboards, pipe alerts to Slack, export costs to your billing system.
Most LLM observability tools are SaaS products that require sending your prompts and completions to their servers. Helicone, Langfuse, and Braintrust all see your data in transit. Lookout runs inside the Stockyard proxy — traces never leave your infrastructure. Every request is automatically recorded with timing data for each middleware step, token counts for input and output, the model and provider used, and the full request and response bodies (configurable — you can redact or skip body logging for sensitive workloads).
The trace data powers three workflows. First, performance debugging: when a request takes longer than expected, the per-middleware timing breakdown shows exactly where the latency came from — was it a cache miss, a slow provider, or a guardrail evaluation? Second, cost analysis: aggregated traces show spend by model, by day, and by API key so you can identify which features cost the most. Third, quality monitoring: replay historical requests through updated middleware configurations to test changes before deploying them to production.
Lookout ships with every Stockyard instance. Self-hosted or Cloud.