4.5 KiB
4.5 KiB
Quick Guide: Pull Ollama Models for TradingAgents
⚠️ IMPORTANT: Pull Models Before Running
When you see a 404 error like:
ResponseError: 404 page not found (status code: 404)
It means the model isn't downloaded yet. You must pull it first!
📥 How to Pull Models
Open a terminal and run:
# RECOMMENDED - Start with this
ollama pull llama3.2
# OR choose from these tool-compatible models:
ollama pull llama3.2:1b # Fastest (1B)
ollama pull llama3.1 # Better quality (8B)
ollama pull mistral-nemo # Mistral (12B)
ollama pull qwen2.5:7b # Qwen (7B)
ollama pull qwen2.5-coder:7b # Coding-focused (7B)
✅ Verify Models Are Installed
ollama list
You should see your models listed:
NAME ID SIZE MODIFIED
llama3.2:latest abc123... 2.0 GB 2 minutes ago
mistral-nemo def456... 7.1 GB 1 hour ago
🎯 Recommended Setup for TradingAgents
For Quick Testing (Fastest)
ollama pull llama3.2:1b
- Size: ~1GB
- Speed: Very fast
- Quality: Good enough for testing
For Production Use (Balanced)
ollama pull llama3.2
- Size: ~2GB
- Speed: Fast
- Quality: Good
For Best Quality (Slower)
ollama pull llama3.1
- Size: ~5GB
- Speed: Medium
- Quality: Excellent
For Mistral Fans
ollama pull mistral-nemo
- Size: ~7GB
- Speed: Medium
- Quality: Excellent
For Qwen Models
# Standard Qwen
ollama pull qwen2.5:7b
# OR Coding-focused variant
ollama pull qwen2.5-coder:7b
- Size: ~4-5GB each
- Speed: Fast
- Quality: Very good
🚀 Complete Workflow
1. Pull a Model
ollama pull llama3.2
2. Verify It's Downloaded
ollama list
3. Run TradingAgents
python -m cli.main
4. Select Settings
- Provider: Ollama
- Quick-Thinking: llama3.2 (or your choice)
- Deep-Thinking: llama3.2 (or your choice)
📊 Model Comparison
| Model | Size | Download Time* | RAM Usage | Speed | Quality | Tools Support |
|---|---|---|---|---|---|---|
| llama3.2:1b | 1GB | ~1 min | 2GB | ⭐⭐⭐⭐⭐ | ⭐⭐⭐ | ✅ |
| llama3.2 | 2GB | ~2 min | 4GB | ⭐⭐⭐⭐ | ⭐⭐⭐⭐ | ✅ |
| llama3.1 | 5GB | ~5 min | 8GB | ⭐⭐⭐ | ⭐⭐⭐⭐⭐ | ✅ |
| mistral-nemo | 7GB | ~7 min | 12GB | ⭐⭐⭐ | ⭐⭐⭐⭐⭐ | ✅ |
| qwen2.5:7b | 4.7GB | ~5 min | 7GB | ⭐⭐⭐⭐ | ⭐⭐⭐⭐ | ✅ |
| qwen2.5-coder | 4.7GB | ~5 min | 7GB | ⭐⭐⭐⭐ | ⭐⭐⭐⭐ | ✅ |
*Approximate download time on typical broadband connection
⚡ Pro Tips
1. Pull Multiple Models
You can have multiple models installed and switch between them:
ollama pull llama3.2 # Fast for testing
ollama pull llama3.1 # High quality for production
2. Check Model Info
ollama show llama3.2
3. Remove Unwanted Models
ollama rm llama3 # Remove old llama3 (doesn't support tools)
4. Keep Models Updated
ollama pull llama3.2 # Updates to latest version
🐛 Troubleshooting
Error: "404 page not found"
Solution: Model not downloaded. Pull it first:
ollama pull llama3.2
Error: "model 'qwen2.5' not found"
Solution: Use full tag:
ollama pull qwen2.5:7b # Not just "qwen2.5"
Slow Performance
Solution: Use smaller model:
ollama pull llama3.2:1b
Out of Memory
Solution: Use smaller model or close other applications:
ollama pull llama3.2:1b # Only needs ~2GB RAM
Model Takes Forever to Download
Solution: Start with smallest model:
ollama pull llama3.2:1b # Only 1GB download
🎓 Learning Path
Beginner
- Start with:
ollama pull llama3.2:1b - Test with simple analysis
- Upgrade if needed
Intermediate
- Use:
ollama pull llama3.2 - Good balance of speed and quality
- Most popular choice
Advanced
- Try:
ollama pull llama3.1ormistral-nemo - Best quality for complex analysis
- Requires more resources
📝 Summary
TL;DR - Quick Start:
# 1. Pull the recommended model
ollama pull llama3.2
# 2. Verify it's there
ollama list
# 3. Run the app
python -m cli.main
That's it! 🚀
Need Help?
Check if Ollama is running:
ollama list
If you see an error, start Ollama:
ollama serve
Then pull your model and try again!