Prerequisites
- HUD API key: Remote training requires authentication. Set
HUD_API_KEYbefore running:
- Docker daemon: For local runs (using
--local) or when training against a local Docker image, ensure Docker Desktop is installed and the Docker daemon is running.
Quickstart
Install and download a taskset:1) Simple: Train (remote by default)
2) Run on your own machine/remote
Use any provider with at least 2 GPUs (one for inference, one for training). Run locally with the flag--local:
Recommended setups
- 2× A100: quick iteration, shorter runs
- 8× A100: higher throughput for larger tasksets
max_parallel_episodes).
3) Build your own environment (hud init)
Create a new MCP environment, develop with hot-reload, and train on a production image:Getting the best performance
Often training a good model requires many iterations over the parameters of the trainer. Take the config generated byhud rl and modify it to various values to do a hyperparameter sweep.
For easy launching, specify the tasks and config upfront, and add --yes to automatically launch vllm and training.
hud rl without training, and then:
- As many different tasks per gradient update as possible (runs with 4+ GPUs and batch size of 50+ are much more stable than single GPU runs)
- Batch size should be somewhere around 2/X where X is the accuracy of that given task on an untrained model.
Pricing
Below is the pricing by GPU type. Actual prices vary — see https://hud.ai/project/billing for current rates. vLLM GPU Pricing (2 Hosted GPUs)| GPU type | Memory | Est. price/hr |
|---|---|---|
| A100 80GB | 80 GB | $4.95 |
| H100 80GB | 80 GB | $7.95 |
| GPU type | Memory | Est. price/hr |
|---|---|---|
| A100 80GB | 80 GB | $3.95 |
| H100 80GB | 80 GB | $5.40 |