From 9109516f7eea2f48e5542fc47cfaba5934a8dcf1 Mon Sep 17 00:00:00 2001 From: Ashish Jain Date: Sun, 12 Apr 2026 15:57:48 -0400 Subject: [PATCH] adapt for llama server --- tradingagents/default_config.py | 2 +- tradingagents/llm_clients/openai_client.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/tradingagents/default_config.py b/tradingagents/default_config.py index f3697703..fcad81b3 100644 --- a/tradingagents/default_config.py +++ b/tradingagents/default_config.py @@ -11,7 +11,7 @@ DEFAULT_CONFIG = { "llm_provider": "openai", "deep_think_llm": "qwen-35b", "quick_think_llm": "qwen-35b", - "backend_url": "https://ai.ghosthouse.uk/v1/chat/completions", + "backend_url": "https://ai.ghosthouse.uk/v1", # Provider-specific thinking configuration "google_thinking_level": None, # "high", "minimal", etc. "openai_reasoning_effort": None, # "medium", "high", "low" diff --git a/tradingagents/llm_clients/openai_client.py b/tradingagents/llm_clients/openai_client.py index 4f2e1b32..40ec218d 100644 --- a/tradingagents/llm_clients/openai_client.py +++ b/tradingagents/llm_clients/openai_client.py @@ -76,7 +76,7 @@ class OpenAIClient(BaseLLMClient): # Native OpenAI: use Responses API for consistent behavior across # all model families. Third-party providers use Chat Completions. - if self.provider == "openai": + if self.provider == "openai" and not self.base_url: llm_kwargs["use_responses_api"] = True return NormalizedChatOpenAI(**llm_kwargs)