From dc5232d769bc218ceb3cf9d2d656c6a2a2e7b085 Mon Sep 17 00:00:00 2001 From: Jeffrey Chu Date: Wed, 22 Oct 2025 19:16:46 +0800 Subject: [PATCH] align format --- cli/utils.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/cli/utils.py b/cli/utils.py index eda6fbfb..ec7b0f78 100644 --- a/cli/utils.py +++ b/cli/utils.py @@ -154,7 +154,7 @@ def select_shallow_thinking_agent(provider) -> str: ("llama3.2 local", "llama3.2"), ], "xai": [ - ("grok-4-fast-non-reasoning", "grok-4-fast-non-reasoning") + ("grok-4-fast-non-reasoning", "grok-4-fast-non-reasoning"), ] } @@ -212,14 +212,14 @@ def select_deep_thinking_agent(provider) -> str: ], "openrouter": [ ("DeepSeek V3 - a 685B-parameter, mixture-of-experts model", "deepseek/deepseek-chat-v3-0324:free"), - ("Deepseek - latest iteration of the flagship chat model family from the DeepSeek team.", "deepseek/deepseek-chat-v3-0324:free") + ("Deepseek - latest iteration of the flagship chat model family from the DeepSeek team.", "deepseek/deepseek-chat-v3-0324:free"), ], "ollama": [ ("llama3.1 local", "llama3.1"), ("qwen3", "qwen3"), ], "xai": [ - ("grok-4-fast-reasoning", "grok-4-fast-reasoning") + ("grok-4-fast-reasoning", "grok-4-fast-reasoning"), ] } @@ -254,7 +254,7 @@ def select_llm_provider() -> tuple[str, str]: ("Anthropic", "https://api.anthropic.com/"), ("Google", "https://generativelanguage.googleapis.com/v1"), ("Openrouter", "https://openrouter.ai/api/v1"), - ("Ollama", "http://localhost:11434/v1"), + ("Ollama", "http://localhost:11434/v1"), ] choice = questionary.select(