update
This commit is contained in:
parent
a84c69e42f
commit
fcca3236e9
12
cli/main.py
12
cli/main.py
|
|
@ -546,10 +546,8 @@ def get_user_selections():
|
||||||
)
|
)
|
||||||
selected_llm_provider, backend_url = select_llm_provider()
|
selected_llm_provider, backend_url = select_llm_provider()
|
||||||
|
|
||||||
provider_id = selected_llm_provider.lower()
|
|
||||||
|
|
||||||
# GitHub Copilot: run OAuth before proceeding
|
# GitHub Copilot: run OAuth before proceeding
|
||||||
if provider_id == "copilot":
|
if selected_llm_provider.lower() == "copilot":
|
||||||
console.print(
|
console.print(
|
||||||
create_question_box(
|
create_question_box(
|
||||||
"Step 5b: Copilot Auth",
|
"Step 5b: Copilot Auth",
|
||||||
|
|
@ -565,15 +563,15 @@ def get_user_selections():
|
||||||
"Step 6: Thinking Agents", "Select your thinking agents for analysis"
|
"Step 6: Thinking Agents", "Select your thinking agents for analysis"
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
selected_shallow_thinker = select_shallow_thinking_agent(provider_id)
|
selected_shallow_thinker = select_shallow_thinking_agent(selected_llm_provider)
|
||||||
selected_deep_thinker = select_deep_thinking_agent(provider_id)
|
selected_deep_thinker = select_deep_thinking_agent(selected_llm_provider)
|
||||||
|
|
||||||
# Step 7: Provider-specific thinking configuration
|
# Step 7: Provider-specific thinking configuration
|
||||||
thinking_level = None
|
thinking_level = None
|
||||||
reasoning_effort = None
|
reasoning_effort = None
|
||||||
anthropic_effort = None
|
anthropic_effort = None
|
||||||
|
|
||||||
provider_lower = provider_id
|
provider_lower = selected_llm_provider.lower()
|
||||||
if provider_lower == "google":
|
if provider_lower == "google":
|
||||||
console.print(
|
console.print(
|
||||||
create_question_box(
|
create_question_box(
|
||||||
|
|
@ -604,7 +602,7 @@ def get_user_selections():
|
||||||
"analysis_date": analysis_date,
|
"analysis_date": analysis_date,
|
||||||
"analysts": selected_analysts,
|
"analysts": selected_analysts,
|
||||||
"research_depth": selected_research_depth,
|
"research_depth": selected_research_depth,
|
||||||
"llm_provider": provider_id,
|
"llm_provider": selected_llm_provider.lower(),
|
||||||
"backend_url": backend_url,
|
"backend_url": backend_url,
|
||||||
"shallow_thinker": selected_shallow_thinker,
|
"shallow_thinker": selected_shallow_thinker,
|
||||||
"deep_thinker": selected_deep_thinker,
|
"deep_thinker": selected_deep_thinker,
|
||||||
|
|
|
||||||
|
|
@ -137,7 +137,9 @@ def select_research_depth() -> int:
|
||||||
def select_shallow_thinking_agent(provider) -> str:
|
def select_shallow_thinking_agent(provider) -> str:
|
||||||
"""Select shallow thinking llm engine using an interactive selection."""
|
"""Select shallow thinking llm engine using an interactive selection."""
|
||||||
|
|
||||||
|
# Define shallow thinking llm engine options with their corresponding model names
|
||||||
# Ordering: medium → light → heavy (balanced first for quick tasks)
|
# Ordering: medium → light → heavy (balanced first for quick tasks)
|
||||||
|
# Within same tier, newer models first
|
||||||
SHALLOW_AGENT_OPTIONS = {
|
SHALLOW_AGENT_OPTIONS = {
|
||||||
"openai": [
|
"openai": [
|
||||||
("GPT-5 Mini - Balanced speed, cost, and capability", "gpt-5-mini"),
|
("GPT-5 Mini - Balanced speed, cost, and capability", "gpt-5-mini"),
|
||||||
|
|
@ -170,7 +172,7 @@ def select_shallow_thinking_agent(provider) -> str:
|
||||||
("GPT-OSS:latest (20B, local)", "gpt-oss:latest"),
|
("GPT-OSS:latest (20B, local)", "gpt-oss:latest"),
|
||||||
("GLM-4.7-Flash:latest (30B, local)", "glm-4.7-flash:latest"),
|
("GLM-4.7-Flash:latest (30B, local)", "glm-4.7-flash:latest"),
|
||||||
],
|
],
|
||||||
"copilot": [], # populated dynamically by fetch_copilot_models()
|
"copilot": [],
|
||||||
}
|
}
|
||||||
|
|
||||||
if provider.lower() == "copilot":
|
if provider.lower() == "copilot":
|
||||||
|
|
@ -212,7 +214,9 @@ def select_shallow_thinking_agent(provider) -> str:
|
||||||
def select_deep_thinking_agent(provider) -> str:
|
def select_deep_thinking_agent(provider) -> str:
|
||||||
"""Select deep thinking llm engine using an interactive selection."""
|
"""Select deep thinking llm engine using an interactive selection."""
|
||||||
|
|
||||||
# Ordering: heavy → medium → light (most capable first for deep tasks)
|
# Define shallow thinking llm engine options with their corresponding model names
|
||||||
|
# Ordering: medium → light → heavy (balanced first for quick tasks)
|
||||||
|
# Within same tier, newer models first
|
||||||
DEEP_AGENT_OPTIONS = {
|
DEEP_AGENT_OPTIONS = {
|
||||||
"openai": [
|
"openai": [
|
||||||
("GPT-5.4 - Latest frontier, 1M context", "gpt-5.4"),
|
("GPT-5.4 - Latest frontier, 1M context", "gpt-5.4"),
|
||||||
|
|
|
||||||
|
|
@ -8,9 +8,6 @@ DEFAULT_CONFIG = {
|
||||||
"dataflows/data_cache",
|
"dataflows/data_cache",
|
||||||
),
|
),
|
||||||
# LLM settings
|
# LLM settings
|
||||||
# Set llm_provider to "copilot" to use GitHub Copilot (no explicit API key
|
|
||||||
# needed — authenticates via `gh auth token` from the GitHub CLI).
|
|
||||||
# Available models are fetched dynamically from the Copilot inference API.
|
|
||||||
"llm_provider": "openai",
|
"llm_provider": "openai",
|
||||||
"deep_think_llm": "gpt-5.2",
|
"deep_think_llm": "gpt-5.2",
|
||||||
"quick_think_llm": "gpt-5-mini",
|
"quick_think_llm": "gpt-5-mini",
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue