model="gpt-4o" for model="claude-sonnet-4-5" and you’re A/B testing across providers. Continuous RL from production coming soon.
Quick Start
Point any OpenAI-compatible client atinference.hud.ai:
Supported Models
Full list at hud.ai/models.Anthropic
Anthropic
| Model | Routes |
|---|---|
claude-sonnet-4-5 | chat, messages |
claude-haiku-4-5 | chat, messages |
claude-opus-4-5 | chat, messages |
claude-opus-4-1 | chat, messages |
OpenAI
OpenAI
| Model | Routes |
|---|---|
gpt-5.1 | chat, responses |
gpt-5-mini | chat, responses |
gpt-4o | chat, responses |
gpt-4o-mini | chat, responses |
operator | responses |
Google Gemini
Google Gemini
| Model | Routes |
|---|---|
gemini-3-pro-preview | chat |
gemini-2.5-pro | chat |
gemini-2.5-computer-use-preview | gemini |
xAI & Others
xAI & Others
| Model | Routes |
|---|---|
grok-4-1-fast | chat |
z-ai/glm-4.5v | chat |
Telemetry
Wrap code in a plainhud.eval() to group inference calls. In the trace you’ll see the full conversation in sequence, not scattered API calls.
hud.eval(). Generate a UUID and pass it with each request in a task:
Routes
- chat —
/chat/completions(OpenAI-compatible) - messages —
/messages(Anthropic-compatible) - responses —
/responses(OpenAI Responses API) - gemini — Google Gemini native API