Update solution to use provider aware embedding logic if not using openai as the LLM provider

This commit is contained in:
sdk451 2025-07-23 11:49:09 +10:00
parent b600d59e31
commit c65d764908
2 changed files with 20 additions and 10 deletions

View File

@ -25,3 +25,4 @@ questionary
langchain_anthropic
langchain-google-genai
python-dotenv
sentence-transformers

View File

@ -1,25 +1,34 @@
import chromadb
from chromadb.config import Settings
from openai import OpenAI
import os
class FinancialSituationMemory:
def __init__(self, name, config):
if config["backend_url"] == "http://localhost:11434/v1":
self.embedding = "nomic-embed-text"
else:
self.config = config
self.provider = config.get("llm_provider", "openai").lower()
if self.provider == "openai":
self.embedding = "text-embedding-3-small"
self.client = OpenAI(base_url=config["backend_url"])
self.client = OpenAI(base_url=config["backend_url"])
self.embedding_model = None
else:
# Use a local embedding model for non-OpenAI providers
from sentence_transformers import SentenceTransformer
self.embedding_model = SentenceTransformer("all-MiniLM-L6-v2")
self.client = None
self.chroma_client = chromadb.Client(Settings(allow_reset=True))
self.situation_collection = self.chroma_client.create_collection(name=name)
def get_embedding(self, text):
"""Get OpenAI embedding for a text"""
response = self.client.embeddings.create(
model=self.embedding, input=text
)
return response.data[0].embedding
if self.provider == "openai":
response = self.client.embeddings.create(
model=self.embedding, input=text
)
return response.data[0].embedding
else:
# Use local embedding model
return self.embedding_model.encode(text).tolist()
def add_situations(self, situations_and_advice):
"""Add financial situations and their corresponding advice. Parameter is a list of tuples (situation, rec)"""