Documentation Index
Fetch the complete documentation index at: https://docs.monostate.ai/llms.txt
Use this file to discover all available pages before exploring further.
Choosing Your Training Approach
Not every AI problem needs the same solution. Sometimes you need full training, sometimes just fine-tuning, and sometimes no training at all.The Three Approaches
Prompt Engineering
No training neededUse existing models with clever prompts
- Instant results
- Zero training cost
- 0-10 examples needed
Fine-Tuning
Recommended approachAdapt pre-trained models to your needs
- Consistent behavior
- 100-10K examples
- Hours to days
Training from Scratch
Rarely neededBuild completely new models
- Full control
- Millions of examples
- Very expensive
1. Prompt Engineering (No Training)
Use existing models with carefully crafted prompts. What it is: Writing instructions that get the model to do what you want without any training. Example:- Testing ideas quickly
- Few examples available
- Task is within model’s capabilities
- Budget/time constraints
- Zero training time
- No training data needed
- Instant results
- Free to try
- Limited customization
- Inconsistent results
- Can’t add new knowledge
- Higher inference cost
2. Fine-Tuning (Recommended)
Adapt a pre-trained model to your specific needs. What it is: Taking a model that already understands language/images and teaching it your specific task. When to use:- Have hundreds to thousands of examples
- Need consistent behavior
- Want to add domain knowledge
- Building production systems
- Faster than training from scratch
- Needs less data
- Better performance than prompting
- Lower inference cost than prompting
- Requires training data
- Needs compute resources
- Takes time to train
- Can overfit on small datasets
3. Training from Scratch
Build a completely new model. What it is: Starting with random weights and training on massive datasets. When to use:- Creating foundational models
- Completely novel architectures
- Unlimited data and compute
- Research purposes
- Full control
- Can create novel capabilities
- No inherited biases
- Needs massive data (millions of examples)
- Extremely expensive
- Takes weeks to months
- Usually unnecessary
Decision Framework
Quick Decision Guide
Check your data
Do you have training examples?
- No → Go with Prompt Engineering
- Yes → Continue to Step 2
Count your examples
How many examples do you have?
- Less than 100 → Use Prompt Engineering
- 100-10,000 → Perfect for Fine-tuning
- Millions → Could train from scratch (but why?)
Detailed Comparison
| Aspect | Prompt Engineering | Fine-Tuning | Training from Scratch |
|---|---|---|---|
| Data Needed | 0-10 examples | 100-10,000 examples | Millions of examples |
| Time to Deploy | Minutes | Hours to days | Weeks to months |
| Cost | $0 upfront | $10-1,000 | $10,000+ |
| Customization | Limited | High | Complete |
| New Knowledge | No | Yes | Yes |
| Consistency | Variable | High | High |
| Maintenance | Update prompts | Retrain periodically | Continuous training |
Approach by Use Case
- Customer Service
- Content Generation
- Code Generation
- Document Analysis
When to use Prompt Engineering:
- General FAQs
- Simple routing
- Low volume
- Testing phase
- Company knowledge
- Brand voice
- High volume
- Complex products
Fine-Tuning Methods
Standard Fine-Tuning
Update all parametersPros:
- Maximum accuracy
- Needs more memory
LoRA
Low-Rank AdaptationPros:
- 90% less memory
- Swap adapters
- Faster training
QLoRA
Quantized LoRAPros:
- Works on consumer GPUs
- 4-bit quantization
- Slightly lower accuracy
Prompt/Prefix Tuning
Train only promptsPros:
- Minimal memory
- Very fast
- Limited capacity
Progressive Approach
Stage 1: Prompt Engineering
Start simple, test fast
- Test the concept
- Gather user feedback
- Identify limitations
- Collect training data
Stage 2: Few-Shot Fine-Tuning
Improve with examples
- Use collected examples
- Improve consistency
- Reduce prompt complexity
- Validate approach
Stage 3: Full Fine-Tuning
Scale for production
- Scale with more data
- Optimize performance
- Reduce inference costs
- Production deployment
Cost Considerations
Prompt Engineering
Prompt Engineering
Training Cost: $0Inference: $0.01-0.10 per 1K tokensBest for: Low volume, experimentationMonthly estimate (1M tokens): $10-100
Fine-Tuning
Fine-Tuning
Training Cost: $10-1,000 (one-time)Inference: $0.001-0.01 per 1K tokensBest for: High volume, productionMonthly estimate (1M tokens): $1-10 + hosting
Training from Scratch
Training from Scratch
Training Cost: $10,000-millionsInference: Variable based on sizeBest for: Foundation model creatorsNot recommended unless you’re OpenAI/Google
Data Requirements
Prompt Engineering
- Minimum: Zero-shot (no examples)
- Better: Few-shot (3-5 examples)
- Best: Many-shot (10+ examples in context)
Fine-Tuning
- Minimum: 50-100 examples
- Better: 500-1,000 examples
- Best: 5,000+ examples
Training from Scratch
- Minimum: 1M+ examples
- Better: 100M+ examples
- Best: Billions of examples
Quality vs Quantity Trade-offs
High Quality, Low Quantity
→ Fine-tune with careful data curation- Hand-picked examples
- Expert annotations
- Data augmentation
Low Quality, High Quantity
→ Use larger models with filtering- Automated cleaning
- Statistical filtering
- Ensemble methods
Mixed Quality
→ Progressive filtering approach- Start with all data
- Identify quality indicators
- Weight by quality
Common Mistakes to Avoid
Hybrid Approaches
RAG (Retrieval Augmented Generation)
Combine prompting with external knowledge. Use when:- Need updatable knowledge
- Can’t fine-tune frequently
- Have structured data
Ensemble Methods
Combine multiple approaches. Example:- Prompt for creativity
- Fine-tuned model for accuracy
- Vote/combine outputs
Chain of Thought + Fine-Tuning
Fine-tune on reasoning steps. Use when:- Need explainable outputs
- Complex reasoning tasks
- Educational applications
Making the Decision
Questions to Ask
-
What’s my budget?
- Low → Prompt engineering
- Medium → Fine-tuning
- High → Consider all options
-
How much data do I have?
- Less than 100 examples → Prompt engineering
- 100-10K → Fine-tuning
- More than 1M → Could train from scratch
-
How unique is my task?
- Common → Prompt engineering
- Specialized → Fine-tuning
- Novel → Training from scratch
-
What accuracy do I need?
- Acceptable → Prompt engineering
- High → Fine-tuning
- Perfect → Multiple iterations
-
How fast do I need results?
- Today → Prompt engineering
- This week → Fine-tuning
- This quarter → Any approach
Next Steps
Ready to start training?Quick Start
Try your first model
Choose Interface
Pick UI, CLI, or API
Model Types
Select architecture