Overview
Fine-tune text generation models for chatbots, Q&A systems, or content generation. This guide shows the minimal setup to get started.Your Data
Your training data should have two columns:- Input column - The user prompt or question
- Output column - The expected response or answer
train.parquet:
| text | sentiment |
|---|---|
| The product exceeded expectations | positive |
| Terrible customer service | negative |
| Average experience, nothing special | neutral |
Minimal Example
Key Parameters
Only these fields are required to start:| Parameter | Description | Example |
|---|---|---|
question_column | Name of your input/prompt column | "text", "question", "prompt" |
answer_column | Name of your output/response column | "sentiment", "answer", "response" |
epochs | Number of training passes (1-3 typical) | 1 |
Supported Models
Popular choices for text generation:meta-llama/Llama-3.2-1B-Instruct- Fast, good for Q&Ameta-llama/Llama-3.2-3B-Instruct- Balanced performancemeta-llama/Llama-3.1-8B-Instruct- Higher quality, slowerQwen/Qwen3-0.6B- Very fast, lightweight
See the full model list for all available options.
Monitor Progress
Check the status of your fine-tune:created, running, completed, errored
Next Steps
- Advanced parameters - Learning rate, batch size, LoRA configuration
- Deploy your model - Use your fine-tuned model for inference
- Full tutorial - End-to-end walkthrough with monitoring
Common Issues
Column not found error
Column not found error
Double-check your
question_column and answer_column names match your data exactly. Column names are case-sensitive.Out of memory error
Out of memory error
Reduce
batch_size to 1 or try a smaller model like Llama-3.2-1B-Instruct.Training taking too long
Training taking too long
Start with 1 epoch. If results aren’t good enough, try 2-3 epochs. More isn’t always better.