perf: switch to Ollama Cloud (deepseek-v3.1:671b-cloud)
Use Ollama Cloud GPU inference instead of self-hosted CPU Ollama. 1-3s per call vs 2-15 minutes. Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
parent
512aff3b40
commit
d1fa7b6004
8
app.py
8
app.py
|
|
@ -68,10 +68,10 @@ class AnalyzeRequest(BaseModel):
|
|||
def build_config():
|
||||
"""Build TradingAgents config — uses Groq (OpenAI-compatible) by default."""
|
||||
config = DEFAULT_CONFIG.copy()
|
||||
config["llm_provider"] = os.getenv("LLM_PROVIDER", "anthropic")
|
||||
config["deep_think_llm"] = os.getenv("DEEP_THINK_MODEL", "claude-sonnet-4-6")
|
||||
config["quick_think_llm"] = os.getenv("QUICK_THINK_MODEL", "claude-haiku-4-5-20251001")
|
||||
config["backend_url"] = os.getenv("LLM_BASE_URL", "https://api.anthropic.com/v1")
|
||||
config["llm_provider"] = os.getenv("LLM_PROVIDER", "openai")
|
||||
config["deep_think_llm"] = os.getenv("DEEP_THINK_MODEL", "deepseek-v3.1:671b-cloud")
|
||||
config["quick_think_llm"] = os.getenv("QUICK_THINK_MODEL", "deepseek-v3.1:671b-cloud")
|
||||
config["backend_url"] = os.getenv("LLM_BASE_URL", "https://ollama.com/v1")
|
||||
config["max_debate_rounds"] = 1
|
||||
config["max_risk_discuss_rounds"] = 1
|
||||
config["data_vendors"] = {
|
||||
|
|
|
|||
Loading…
Reference in New Issue