50 lines
1.4 KiB
Python
50 lines
1.4 KiB
Python
from typing import Optional
|
|
|
|
from .base_client import BaseLLMClient
|
|
from .openai_client import OpenAIClient
|
|
from .anthropic_client import AnthropicClient
|
|
from .google_client import GoogleClient
|
|
from .azure_client import AzureOpenAIClient
|
|
|
|
# Providers that use the OpenAI-compatible chat completions API
|
|
_OPENAI_COMPATIBLE = (
|
|
"openai", "xai", "deepseek", "qwen", "glm", "ollama", "openrouter", "minimax",
|
|
)
|
|
|
|
|
|
def create_llm_client(
|
|
provider: str,
|
|
model: str,
|
|
base_url: Optional[str] = None,
|
|
**kwargs,
|
|
) -> BaseLLMClient:
|
|
"""Create an LLM client for the specified provider.
|
|
|
|
Args:
|
|
provider: LLM provider name
|
|
model: Model name/identifier
|
|
base_url: Optional base URL for API endpoint
|
|
**kwargs: Additional provider-specific arguments
|
|
|
|
Returns:
|
|
Configured BaseLLMClient instance
|
|
|
|
Raises:
|
|
ValueError: If provider is not supported
|
|
"""
|
|
provider_lower = provider.lower()
|
|
|
|
if provider_lower in _OPENAI_COMPATIBLE:
|
|
return OpenAIClient(model, base_url, provider=provider_lower, **kwargs)
|
|
|
|
if provider_lower == "anthropic":
|
|
return AnthropicClient(model, base_url, **kwargs)
|
|
|
|
if provider_lower == "google":
|
|
return GoogleClient(model, base_url, **kwargs)
|
|
|
|
if provider_lower == "azure":
|
|
return AzureOpenAIClient(model, base_url, **kwargs)
|
|
|
|
raise ValueError(f"Unsupported LLM provider: {provider}")
|