4.9 KiB
Ollama Models for TradingAgents
✅ Verified Tool-Compatible Models
These models support tool calling / function calling which is required for TradingAgents to work:
Recommended Models
| Model | Size | Speed | Quality | Command |
|---|---|---|---|---|
| llama3.2 ⭐ | 3B | Fast | Good | ollama pull llama3.2 |
| llama3.2:1b | 1B | Fastest | Moderate | ollama pull llama3.2:1b |
| llama3.1 | 8B | Medium | Better | ollama pull llama3.1 |
| mistral-nemo | 12B | Medium | Better | ollama pull mistral-nemo |
| qwen2.5 | 7B | Fast | Good | ollama pull qwen2.5 |
⭐ Best Choice for Most Users
ollama pull llama3.2
Why llama3.2?
- ✅ Supports tool calling
- ✅ Fast inference
- ✅ Good quality
- ✅ Reasonable memory usage (~4GB)
Model Details
llama3.2 (RECOMMENDED)
- Variants: 1B, 3B (default)
- Best For: General trading analysis
- Memory: ~2-4GB
- Speed: 2-3 minutes per analysis
- Tools: ✅ Full support
# Default (3B)
ollama pull llama3.2
# Smallest (1B) - fastest
ollama pull llama3.2:1b
llama3.1
- Variants: 8B, 70B, 405B
- Best For: Higher quality analysis
- Memory: ~8GB+ (for 8B)
- Speed: 3-5 minutes per analysis
- Tools: ✅ Full support
# Most common (8B)
ollama pull llama3.1
# High quality (70B) - requires powerful GPU
ollama pull llama3.1:70b
mistral-nemo
- Size: 12B
- Best For: Balanced quality/speed
- Memory: ~12GB
- Speed: 3-4 minutes per analysis
- Tools: ✅ Full support
ollama pull mistral-nemo
qwen2.5
- Variants: 0.5B, 1.5B, 3B, 7B, 14B, 32B, 72B
- Best For: Good multilingual support
- Memory: Varies (7B ~7GB)
- Speed: Fast
- Tools: ✅ Full support
# Default (7B)
ollama pull qwen2.5
# Smaller variants
ollama pull qwen2.5:3b
ollama pull qwen2.5:1.5b
❌ Models That DON'T Support Tools
These models will NOT work with TradingAgents:
- ❌
llama3(original) - ❌
llama2 - ❌
mistral(v0.1-0.2) - ❌
codellama(designed for code, not tools) - ❌ Most older models
Quick Start
1. Install Ollama
Download from: https://ollama.ai
2. Pull a Model
# RECOMMENDED
ollama pull llama3.2
# OR for better quality (slower)
ollama pull llama3.1
# OR for Mistral
ollama pull mistral-nemo
3. Verify Model Works
ollama list
You should see your model listed.
4. Use in TradingAgents
When running the CLI, select:
- Provider: Ollama
- Quick-Thinking LLM: llama3.2 (or your choice)
- Deep-Thinking LLM: llama3.2 (or your choice)
Performance Comparison
Speed Test (Single AAPL Analysis)
| Model | Time | Memory | Quality |
|---|---|---|---|
| llama3.2:1b | ~1-2 min | 2GB | ⭐⭐⭐ |
| llama3.2 (3B) | ~2-3 min | 4GB | ⭐⭐⭐⭐ |
| llama3.1 (8B) | ~3-5 min | 8GB | ⭐⭐⭐⭐⭐ |
| mistral-nemo | ~3-4 min | 12GB | ⭐⭐⭐⭐⭐ |
| qwen2.5 | ~2-3 min | 7GB | ⭐⭐⭐⭐ |
Times approximate on modern consumer hardware (RTX 3060+)
Advanced Options
Different Model Sizes
Many models have variants. List all available versions:
ollama list | grep llama3.2
Pull specific variants:
# Smallest llama3.2
ollama pull llama3.2:1b
# Default llama3.2
ollama pull llama3.2
# Latest llama3.2
ollama pull llama3.2:latest
Check Model Info
ollama show llama3.2
Remove Models
ollama rm llama3
ollama rm mistral
Troubleshooting
Error: "does not support tools"
Problem: You're using a model that doesn't support function calling.
Solution: Switch to a supported model:
ollama pull llama3.2
Slow Performance
Solution 1: Use a smaller model
ollama pull llama3.2:1b
Solution 2: Check GPU usage
# Make sure Ollama is using GPU
ollama show llama3.2 | grep gpu
Out of Memory
Solution: Use smaller model or reduce context
# Smallest option
ollama pull llama3.2:1b
Recommendations by Use Case
Development & Testing
Fastest: llama3.2:1b
ollama pull llama3.2:1b
Production (Free/Local)
Balanced: llama3.2 (3B default)
ollama pull llama3.2
High Quality (Local)
Best: llama3.1 (8B)
ollama pull llama3.1
Budget GPU
Efficient: qwen2.5:3b
ollama pull qwen2.5:3b
Future Models
New models are constantly being released. Check for tool support:
- Visit: https://ollama.ai/library
- Look for "Tools" or "Function Calling" in model description
- Test with:
python quick_test_ollama.py
Summary
✅ Best for most users: llama3.2
✅ Best quality (local): llama3.1
✅ Fastest: llama3.2:1b
✅ Balanced: mistral-nemo or qwen2.5
Command to get started:
ollama pull llama3.2
Then run:
python -m cli.main
And select Ollama as your provider! 🚀