TradingAgents/docs/README_ADDITION.md

2.9 KiB

README Addition - Multi-Provider AI Support

Add this section to your README.md after the "Required APIs" section:


🚀 NEW: Multi-Provider AI Support

TradingAgents now supports multiple AI/LLM providers! You're no longer limited to OpenAI.

Supported Providers:

  • OpenAI (GPT-4, GPT-4o, GPT-3.5-turbo)
  • Ollama (Local models - FREE! Llama 3, Mistral, Mixtral, etc.)
  • Anthropic (Claude 3 Opus, Sonnet, Haiku)
  • Google (Gemini Pro, Gemini Flash)
  • Groq (Fast inference)
  • OpenRouter (Multi-provider access)
  • Azure OpenAI
  • Together AI
  • HuggingFace

📚 See Full Provider Guide | Quick Start Examples

Quick Examples

OpenAI (Default - No Changes Needed):

from tradingagents.graph.trading_graph import TradingAgentsGraph
from tradingagents.default_config import DEFAULT_CONFIG

ta = TradingAgentsGraph(config=DEFAULT_CONFIG)
_, decision = ta.propagate("NVDA", "2024-05-10")

Ollama (Local & Free):

config = DEFAULT_CONFIG.copy()
config.update({
    "llm_provider": "ollama",
    "deep_think_llm": "llama3:70b",
    "quick_think_llm": "llama3:8b",
    "backend_url": "http://localhost:11434"
})
ta = TradingAgentsGraph(config=config)

Anthropic Claude:

config = DEFAULT_CONFIG.copy()
config.update({
    "llm_provider": "anthropic",
    "deep_think_llm": "claude-3-opus-20240229",
    "quick_think_llm": "claude-3-haiku-20240307"
})
ta = TradingAgentsGraph(config=config)

Google Gemini:

config = DEFAULT_CONFIG.copy()
config.update({
    "llm_provider": "google",
    "deep_think_llm": "gemini-1.5-pro",
    "quick_think_llm": "gemini-1.5-flash"
})
ta = TradingAgentsGraph(config=config)

Groq (Fast & Affordable):

config = DEFAULT_CONFIG.copy()
config.update({
    "llm_provider": "groq",
    "deep_think_llm": "mixtral-8x7b-32768",
    "quick_think_llm": "llama3-8b-8192"
})
ta = TradingAgentsGraph(config=config)

See examples/llm_provider_configs.py for more pre-configured options!


Then update the "Implementation Details" section to say:

We built TradingAgents with LangGraph to ensure flexibility and modularity. The system now supports multiple LLM providers through a unified interface. You can use OpenAI (default), Ollama for local/free models, Anthropic Claude, Google Gemini, Groq, and others.

For OpenAI, we recommend using o4-mini and gpt-4o-mini for cost-effective testing, as our framework makes lots of API calls. For production, consider o1-preview and gpt-4o.

For free local inference, use Ollama with Llama 3 models. For the best quality, use Claude 3 Opus or GPT-4o. For the fastest inference, use Groq.

See the LLM Provider Guide for detailed recommendations and setup instructions.