save
This commit is contained in:
parent
817deb8247
commit
324ef8a98f
|
|
@ -5,11 +5,11 @@ from openai import OpenAI
|
||||||
|
|
||||||
class FinancialSituationMemory:
|
class FinancialSituationMemory:
|
||||||
def __init__(self, name, config):
|
def __init__(self, name, config):
|
||||||
if config["backend_url"] == "http://localhost:11434/v1":
|
if config["backend_embedding_url"] == "http://localhost:11434/v1":
|
||||||
self.embedding = "nomic-embed-text"
|
self.embedding = "nomic-embed-text"
|
||||||
else:
|
else:
|
||||||
self.embedding = "text-embedding-3-small"
|
self.embedding = "text-embedding-3-small"
|
||||||
self.client = OpenAI(base_url=config["backend_url"])
|
self.client = OpenAI(base_url=config["backend_embedding_url"])
|
||||||
self.chroma_client = chromadb.Client(Settings(allow_reset=True))
|
self.chroma_client = chromadb.Client(Settings(allow_reset=True))
|
||||||
self.situation_collection = self.chroma_client.create_collection(name=name)
|
self.situation_collection = self.chroma_client.create_collection(name=name)
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -4,7 +4,7 @@ from typing import Annotated
|
||||||
from .local import get_YFin_data, get_finnhub_news, get_finnhub_company_insider_sentiment, get_finnhub_company_insider_transactions, get_simfin_balance_sheet, get_simfin_cashflow, get_simfin_income_statements, get_reddit_global_news, get_reddit_company_news
|
from .local import get_YFin_data, get_finnhub_news, get_finnhub_company_insider_sentiment, get_finnhub_company_insider_transactions, get_simfin_balance_sheet, get_simfin_cashflow, get_simfin_income_statements, get_reddit_global_news, get_reddit_company_news
|
||||||
from .local_news import get_local_news
|
from .local_news import get_local_news
|
||||||
from .y_finance import get_YFin_data_online, get_stock_stats_indicators_window, get_balance_sheet as get_yfinance_balance_sheet, get_cashflow as get_yfinance_cashflow, get_income_statement as get_yfinance_income_statement, get_insider_transactions as get_yfinance_insider_transactions
|
from .y_finance import get_YFin_data_online, get_stock_stats_indicators_window, get_balance_sheet as get_yfinance_balance_sheet, get_cashflow as get_yfinance_cashflow, get_income_statement as get_yfinance_income_statement, get_insider_transactions as get_yfinance_insider_transactions
|
||||||
from .google import get_google_news
|
# Note: get_google_news not imported - incompatible signature with get_news tool interface
|
||||||
from .openai import get_stock_news_openai, get_global_news_openai, get_fundamentals_openai
|
from .openai import get_stock_news_openai, get_global_news_openai, get_fundamentals_openai
|
||||||
from .alpha_vantage import (
|
from .alpha_vantage import (
|
||||||
get_stock as get_alpha_vantage_stock,
|
get_stock as get_alpha_vantage_stock,
|
||||||
|
|
@ -98,11 +98,11 @@ VENDOR_METHODS = {
|
||||||
"local": get_simfin_income_statements,
|
"local": get_simfin_income_statements,
|
||||||
},
|
},
|
||||||
# news_data
|
# news_data
|
||||||
|
# Note: get_google_news removed - incompatible signature (expects look_back_days int, not end_date string)
|
||||||
"get_news": {
|
"get_news": {
|
||||||
"alpha_vantage": get_alpha_vantage_news,
|
"alpha_vantage": get_alpha_vantage_news,
|
||||||
"openai": get_stock_news_openai,
|
"openai": get_stock_news_openai,
|
||||||
"google": get_google_news,
|
"local": [get_finnhub_news, get_reddit_company_news],
|
||||||
"local": [get_finnhub_news, get_reddit_company_news, get_google_news],
|
|
||||||
"local_news": get_local_news
|
"local_news": get_local_news
|
||||||
},
|
},
|
||||||
"get_global_news": {
|
"get_global_news": {
|
||||||
|
|
|
||||||
|
|
@ -18,10 +18,14 @@ DEFAULT_CONFIG = {
|
||||||
# "quick_think_llm": "qwen3:30b",
|
# "quick_think_llm": "qwen3:30b",
|
||||||
# "deep_think_llm": "gpt-oss",
|
# "deep_think_llm": "gpt-oss",
|
||||||
# "quick_think_llm": "gpt-oss",
|
# "quick_think_llm": "gpt-oss",
|
||||||
|
# "temperature": 0.1,
|
||||||
"deep_think_llm": "glm-4.7-flash",
|
"deep_think_llm": "glm-4.7-flash",
|
||||||
"quick_think_llm": "glm-4.7-flash",
|
"quick_think_llm": "glm-4.7-flash",
|
||||||
# "backend_url": "http://localhost:8080/v1",
|
"temperature": 0.7,
|
||||||
"backend_url": "http://localhost:11434/v1",
|
"top_p": 0.95,
|
||||||
|
"backend_url": "http://localhost:8080/v1",
|
||||||
|
"backend_embedding_url": "http://localhost:8081/v1",
|
||||||
|
# "backend_url": "http://localhost:11434/v1",
|
||||||
# Debate and discussion settings
|
# Debate and discussion settings
|
||||||
"max_debate_rounds": 5,
|
"max_debate_rounds": 5,
|
||||||
"max_risk_discuss_rounds": 5,
|
"max_risk_discuss_rounds": 5,
|
||||||
|
|
@ -49,4 +53,6 @@ DEFAULT_CONFIG = {
|
||||||
# "get_insider_sentiment": "na",
|
# "get_insider_sentiment": "na",
|
||||||
"get_insider_transactions": "yfinance",
|
"get_insider_transactions": "yfinance",
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
# docker run --rm -it --gpus all -v ~/.cache/llama.cpp:/root/.cache/llama.cpp -p 8080:8080 ghcr.io/ggml-org/llama.cpp:server-cuda -hf unsloth/GLM-4.7-Flash-GGUF:Q4_K_M --port 8080 --host 0.0.0.0 -ngl 99 --jinja --ctx-size 32768 --min-p 0.01 --embeddings --pooling cls --ubatch-size 8192
|
||||||
|
|
@ -73,11 +73,23 @@ class TradingAgentsGraph:
|
||||||
|
|
||||||
# Initialize LLMs
|
# Initialize LLMs
|
||||||
if self.config["llm_provider"].lower() == "openai" or self.config["llm_provider"] == "ollama" or self.config["llm_provider"] == "openrouter":
|
if self.config["llm_provider"].lower() == "openai" or self.config["llm_provider"] == "ollama" or self.config["llm_provider"] == "openrouter":
|
||||||
self.deep_thinking_llm = ChatOpenAI(model=self.config["deep_think_llm"], base_url=self.config["backend_url"], temperature=0)
|
self.deep_thinking_llm = ChatOpenAI(model=self.config["deep_think_llm"],
|
||||||
self.quick_thinking_llm = ChatOpenAI(model=self.config["quick_think_llm"], base_url=self.config["backend_url"], temperature=0)
|
base_url=self.config["backend_url"],
|
||||||
|
temperature=self.config["temperature"],
|
||||||
|
top_p=self.config["top_p"])
|
||||||
|
self.quick_thinking_llm = ChatOpenAI(model=self.config["quick_think_llm"],
|
||||||
|
base_url=self.config["backend_url"],
|
||||||
|
temperature=self.config["temperature"],
|
||||||
|
top_p=self.config["top_p"])
|
||||||
elif self.config["llm_provider"].lower() == "anthropic":
|
elif self.config["llm_provider"].lower() == "anthropic":
|
||||||
self.deep_thinking_llm = ChatAnthropic(model=self.config["deep_think_llm"], base_url=self.config["backend_url"], temperature=0)
|
self.deep_thinking_llm = ChatAnthropic(model=self.config["deep_think_llm"],
|
||||||
self.quick_thinking_llm = ChatAnthropic(model=self.config["quick_think_llm"], base_url=self.config["backend_url"], temperature=0)
|
base_url=self.config["backend_url"],
|
||||||
|
temperature=self.config["temperature"],
|
||||||
|
top_p=self.config["top_p"])
|
||||||
|
self.quick_thinking_llm = ChatAnthropic(model=self.config["quick_think_llm"],
|
||||||
|
base_url=self.config["backend_url"],
|
||||||
|
temperature=self.config["temperature"],
|
||||||
|
top_p=self.config["top_p"])
|
||||||
elif self.config["llm_provider"].lower() == "google":
|
elif self.config["llm_provider"].lower() == "google":
|
||||||
self.deep_thinking_llm = ChatGoogleGenerativeAI(model=self.config["deep_think_llm"])
|
self.deep_thinking_llm = ChatGoogleGenerativeAI(model=self.config["deep_think_llm"])
|
||||||
self.quick_thinking_llm = ChatGoogleGenerativeAI(model=self.config["quick_think_llm"])
|
self.quick_thinking_llm = ChatGoogleGenerativeAI(model=self.config["quick_think_llm"])
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue