diff --git a/cli/utils.py b/cli/utils.py index d3873360..217af9bf 100644 --- a/cli/utils.py +++ b/cli/utils.py @@ -151,6 +151,9 @@ def select_shallow_thinking_agent(provider) -> str: ], "ollama": [ ("llama3.2 local", "llama3.2"), + ], + "deepseek": [ + ("DeepSeek V3 - a 685B-parameter, mixture-of-experts model", "deepseek-chat") ] } @@ -212,6 +215,10 @@ def select_deep_thinking_agent(provider) -> str: ], "ollama": [ ("qwen3", "qwen3"), + ], + "deepseek": [ + ("DeepSeek V3 - a 685B-parameter, mixture-of-experts model", "deepseek-chat"), + ("DeepSeek-R1 - latest iteration of the flagship chat model family from the DeepSeek team.", "deepseek-reasoner"), ] } @@ -245,7 +252,8 @@ def select_llm_provider() -> tuple[str, str]: ("Anthropic", "https://api.anthropic.com/"), ("Google", "https://generativelanguage.googleapis.com/v1"), ("Openrouter", "https://openrouter.ai/api/v1"), - ("Ollama", "http://localhost:11434/v1"), + ("Ollama", "http://localhost:11434/v1"), + ("DeepSeek", "https://api.deepseek.com/v1") ] choice = questionary.select( diff --git a/requirements.txt b/requirements.txt index 1c7c2818..4d643427 100644 --- a/requirements.txt +++ b/requirements.txt @@ -22,3 +22,8 @@ redis chainlit rich questionary +langchain_anthropic +langchain_deepseek +langchain_google_genai +openai +sentence-transformers \ No newline at end of file diff --git a/tradingagents/agents/utils/memory.py b/tradingagents/agents/utils/memory.py index f3415765..e1428204 100644 --- a/tradingagents/agents/utils/memory.py +++ b/tradingagents/agents/utils/memory.py @@ -1,7 +1,9 @@ +import os + import chromadb from chromadb.config import Settings from openai import OpenAI - +from sentence_transformers import SentenceTransformer class FinancialSituationMemory: def __init__(self, name, config): @@ -9,17 +11,22 @@ class FinancialSituationMemory: self.embedding = "nomic-embed-text" else: self.embedding = "text-embedding-3-small" - self.client = OpenAI() self.chroma_client = chromadb.Client(Settings(allow_reset=True)) self.situation_collection = self.chroma_client.create_collection(name=name) def get_embedding(self, text): """Get OpenAI embedding for a text""" - - response = self.client.embeddings.create( - model=self.embedding, input=text - ) - return response.data[0].embedding + if os.environ.get("OPENAI_API_KEY"): + client = OpenAI() + response = client.embeddings.create( + model=self.embedding, input=text + ) + return response.data[0].embedding + else: + model = SentenceTransformer("all-MiniLM-L6-v2") + embeddings = model.encode(text, convert_to_numpy=True) + return embeddings + def add_situations(self, situations_and_advice): """Add financial situations and their corresponding advice. Parameter is a list of tuples (situation, rec)""" diff --git a/tradingagents/graph/trading_graph.py b/tradingagents/graph/trading_graph.py index eb06cf43..e4c5888a 100644 --- a/tradingagents/graph/trading_graph.py +++ b/tradingagents/graph/trading_graph.py @@ -6,6 +6,7 @@ import json from datetime import date from typing import Dict, Any, Tuple, List, Optional +from langchain_deepseek import ChatDeepSeek from langchain_openai import ChatOpenAI from langchain_anthropic import ChatAnthropic from langchain_google_genai import ChatGoogleGenerativeAI @@ -67,6 +68,9 @@ class TradingAgentsGraph: elif self.config["llm_provider"].lower() == "google": self.deep_thinking_llm = ChatGoogleGenerativeAI(model=self.config["deep_think_llm"]) self.quick_thinking_llm = ChatGoogleGenerativeAI(model=self.config["quick_think_llm"]) + elif self.config["llm_provider"].lower() == 'deepseek': + self.deep_thinking_llm = ChatDeepSeek(model=self.config["deep_think_llm"]) + self.quick_thinking_llm = ChatDeepSeek(model=self.config["quick_think_llm"]) else: raise ValueError(f"Unsupported LLM provider: {self.config['llm_provider']}")