Overview
Reinforcement Fine-Tuning (RFT) allows you to fine-tune language models using reinforcement learning on HUD tasks. This creates specialized models optimized for your specific use cases.Access Required: RFT is currently available by invite only. Contact founders@hud.ai to request access.
RFT currently supports OpenAI models and requires a minimum of 10 tasks for effective training.
Prerequisites
- RFT access (contact founders@hud.ai to request)
- HUD API key set in environment (
HUD_API_KEY) - A HUD environment with remote tasks configured
- At least 10 tasks in your dataset
- No vision support required (RFT doesn’t support vision-based environments)
Step-by-Step Process
1. Prepare Your Tasks
First, ensure you have a tasks file with at least 10 tasks. You can download existing datasets or create your own:2. Launch RFT Training
Run the RFT command from within your environment directory:Use
--yes flag to skip all confirmation prompts for automation.3. Monitor Training Progress
Check the status of your training job:4. Use Your Fine-Tuned Model
Once training is complete (status shows “succeeded”), you’ll receive a fine-tuned model ID. Use it with the HUD inference API:Replace the model ID in the example with your actual fine-tuned model ID from the status command.
Command Reference
Launch Training
Check Status
Tips
- Start with smaller datasets (10-50 tasks) to test your approach
- Use
--verboseduring development to see detailed information - Monitor logs for any errors during training
- Fine-tuned models are optimized for tasks similar to your training data
Limitations
- Minimum 10 tasks required
- No vision support (text-based tasks only)
- Currently supports OpenAI models only
- Training time varies based on dataset size
Troubleshooting
If training fails:- Verify your tasks have valid remote configurations
- Ensure all environment variables are set
- Check that your environment doesn’t require vision support
- Use
--verboseflag for detailed error messages