open-multi-agent/src/llm
JackChen bc31008f4e feat(llm): add fallback tool-call extraction for local models (#15)
Local models (Ollama, vLLM) sometimes return tool calls as text instead
of using the native tool_calls wire format. This adds a safety-net
extractor that parses tool calls from model text output when native
tool_calls is empty.

- Add text-tool-extractor with support for bare JSON, code fences,
  and Hermes <tool_call> tags
- Wire fallback into OpenAI adapter chat() and stream() paths
- Add onWarning callback when model ignores configured tools
- Add timeoutMs on AgentConfig for per-run abort (local models can
  be slow)
- Add 26 tests for extractor and fallback behavior
- Document local model compatibility in README
2026-04-05 11:49:10 +08:00
..
adapter.ts Merge branch 'main' into main 2026-04-04 13:43:40 +03:00
anthropic.ts feat: add baseURL and apiKey support for OpenAI-compatible APIs 2026-04-02 19:33:10 +08:00
copilot.ts feat(llm): add fallback tool-call extraction for local models (#15) 2026-04-05 11:49:10 +08:00
gemini.ts chore: support Node >=18, add optional @google/genai peer dependency and API key fallback 2026-04-04 22:12:28 +03:00
grok.ts feat(llm): add first-class Grok (xAI) support with dedicated GrokAdapter (#44) 2026-04-04 18:20:55 +08:00
openai-common.ts feat(llm): add fallback tool-call extraction for local models (#15) 2026-04-05 11:49:10 +08:00
openai.ts feat(llm): add fallback tool-call extraction for local models (#15) 2026-04-05 11:49:10 +08:00