TradingAgents/PULL_OLLAMA_MODELS.md

4.5 KiB

Quick Guide: Pull Ollama Models for TradingAgents

⚠️ IMPORTANT: Pull Models Before Running

When you see a 404 error like:

ResponseError: 404 page not found (status code: 404)

It means the model isn't downloaded yet. You must pull it first!

📥 How to Pull Models

Open a terminal and run:

# RECOMMENDED - Start with this
ollama pull llama3.2

# OR choose from these tool-compatible models:
ollama pull llama3.2:1b       # Fastest (1B)
ollama pull llama3.1          # Better quality (8B)
ollama pull mistral-nemo      # Mistral (12B)
ollama pull qwen2.5:7b        # Qwen (7B)
ollama pull qwen2.5-coder:7b  # Coding-focused (7B)

Verify Models Are Installed

ollama list

You should see your models listed:

NAME            ID              SIZE    MODIFIED
llama3.2:latest abc123...       2.0 GB  2 minutes ago
mistral-nemo    def456...       7.1 GB  1 hour ago

For Quick Testing (Fastest)

ollama pull llama3.2:1b
  • Size: ~1GB
  • Speed: Very fast
  • Quality: Good enough for testing

For Production Use (Balanced)

ollama pull llama3.2
  • Size: ~2GB
  • Speed: Fast
  • Quality: Good

For Best Quality (Slower)

ollama pull llama3.1
  • Size: ~5GB
  • Speed: Medium
  • Quality: Excellent

For Mistral Fans

ollama pull mistral-nemo
  • Size: ~7GB
  • Speed: Medium
  • Quality: Excellent

For Qwen Models

# Standard Qwen
ollama pull qwen2.5:7b

# OR Coding-focused variant
ollama pull qwen2.5-coder:7b
  • Size: ~4-5GB each
  • Speed: Fast
  • Quality: Very good

🚀 Complete Workflow

1. Pull a Model

ollama pull llama3.2

2. Verify It's Downloaded

ollama list

3. Run TradingAgents

python -m cli.main

4. Select Settings

  • Provider: Ollama
  • Quick-Thinking: llama3.2 (or your choice)
  • Deep-Thinking: llama3.2 (or your choice)

📊 Model Comparison

Model Size Download Time* RAM Usage Speed Quality Tools Support
llama3.2:1b 1GB ~1 min 2GB
llama3.2 2GB ~2 min 4GB
llama3.1 5GB ~5 min 8GB
mistral-nemo 7GB ~7 min 12GB
qwen2.5:7b 4.7GB ~5 min 7GB
qwen2.5-coder 4.7GB ~5 min 7GB

*Approximate download time on typical broadband connection

Pro Tips

1. Pull Multiple Models

You can have multiple models installed and switch between them:

ollama pull llama3.2     # Fast for testing
ollama pull llama3.1     # High quality for production

2. Check Model Info

ollama show llama3.2

3. Remove Unwanted Models

ollama rm llama3  # Remove old llama3 (doesn't support tools)

4. Keep Models Updated

ollama pull llama3.2  # Updates to latest version

🐛 Troubleshooting

Error: "404 page not found"

Solution: Model not downloaded. Pull it first:

ollama pull llama3.2

Error: "model 'qwen2.5' not found"

Solution: Use full tag:

ollama pull qwen2.5:7b  # Not just "qwen2.5"

Slow Performance

Solution: Use smaller model:

ollama pull llama3.2:1b

Out of Memory

Solution: Use smaller model or close other applications:

ollama pull llama3.2:1b  # Only needs ~2GB RAM

Model Takes Forever to Download

Solution: Start with smallest model:

ollama pull llama3.2:1b  # Only 1GB download

🎓 Learning Path

Beginner

  1. Start with: ollama pull llama3.2:1b
  2. Test with simple analysis
  3. Upgrade if needed

Intermediate

  1. Use: ollama pull llama3.2
  2. Good balance of speed and quality
  3. Most popular choice

Advanced

  1. Try: ollama pull llama3.1 or mistral-nemo
  2. Best quality for complex analysis
  3. Requires more resources

📝 Summary

TL;DR - Quick Start:

# 1. Pull the recommended model
ollama pull llama3.2

# 2. Verify it's there
ollama list

# 3. Run the app
python -m cli.main

That's it! 🚀


Need Help?

Check if Ollama is running:

ollama list

If you see an error, start Ollama:

ollama serve

Then pull your model and try again!