The platform for building
RL environments
Turn your production system into an environment.
Evaluate and train agents on your own software, web apps, or chat interfaces. Start from one of our templates and deploy in less than 30 minutes.
Start from a template
Deep Research
Advanced research environment with Exa search integration for deep web research capabilities.
Rubrics
Environment for creating and evaluating with rubrics, powered by TLDC methodology.
Browser
Local browser automation environment with HUD for web interaction and testing.
Coding
Full-featured coding environment with language servers, linters, and development tools.
claude-sonnet-4-5 successfully performing a financial analyst task. Read the SheetBench case study →
One API for testing any model.
Stop juggling API keys. Point any OpenAI-compatible client at inference.hud.ai and use Claude, GPT, Gemini, or Grok. Every call is traced on hud.ai.
from openai import AsyncOpenAI
import os
client = AsyncOpenAI(
base_url="https://inference.hud.ai",
api_key=os.environ["HUD_API_KEY"]
)
response = await client.chat.completions.create(
model="model-name",
messages=[{"role": "user", "content": "Hello!"}]
)
print(response.choices[0].message.content)Run evaluations and training at scale.
Our infrastructure handles 1000s of concurrent environments with sub-second latency. Run full benchmark suites in minutes, not hours.
Pricing
SDK
- ✓Turn any software into agent tools
- ✓Define scenarios for evaluation
- ✓Compatible with any agent framework
Cloud
- ✓100+ parallel environment instances
- ✓Live telemetry and debugging
- ✓Detailed trace analysis
Start with $10 in free credits!
Start evaluatingEnterprise
- ✓Train agents on your environments
- ✓SOC 2 compliant infrastructure
- ✓Volume pricing and dedicated support
Are you a student or researcher? Get $100 in free credits with a .edu email. Making an academic eval? Apply for a grant.