Commit Graph

38 Commits

Author SHA1 Message Date
JackChen 3b6640b3a2 docs: refresh Star History and Contributors cache bust params 2026-04-05 13:21:52 +08:00
JackChen b4cdf741bd docs: add Translations section to README 2026-04-05 13:20:49 +08:00
JackChen 42fe51b3e1 docs: sync READMEs with recent PRs (#28, #32, #44, #45, #47, #49)
English README:
- Update source file count 27 → 33
- Add Human-in-the-Loop, Lifecycle Hooks, Loop Detection feature bullets
- Add Example 13 (Gemini)
- Add GrokAdapter to architecture diagram

Chinese README:
- Update source file count 27 → 33
- Add llama.cpp server to model-agnostic description
- Add Human-in-the-Loop, Lifecycle Hooks, Loop Detection feature bullets
- Update Quick Start API key list (add GEMINI_API_KEY, XAI_API_KEY)
- Add Example 12 (Grok) and Example 13 (Gemini)
- Add GrokAdapter to architecture diagram
- Add Grok and llama.cpp rows to Provider table
- Add Local Model Tool-Calling section (fallback, timeout, troubleshooting)
- Add LLM Configuration Examples section (Grok config)
- Update Contributing section (remove outdated Gemini reference)
2026-04-05 13:08:46 +08:00
JackChen bc31008f4e feat(llm): add fallback tool-call extraction for local models (#15)
Local models (Ollama, vLLM) sometimes return tool calls as text instead
of using the native tool_calls wire format. This adds a safety-net
extractor that parses tool calls from model text output when native
tool_calls is empty.

- Add text-tool-extractor with support for bare JSON, code fences,
  and Hermes <tool_call> tags
- Wire fallback into OpenAI adapter chat() and stream() paths
- Add onWarning callback when model ignores configured tools
- Add timeoutMs on AgentConfig for per-run abort (local models can
  be slow)
- Add 26 tests for extractor and fallback behavior
- Document local model compatibility in README
2026-04-05 11:49:10 +08:00
JackChen 9a81a13982
feat(llm): add Gemini adapter (#28)
feat: Add support for Gemini model
2026-04-05 11:48:32 +08:00
Trevin Chow 9e72c5029e docs: add llama.cpp server to local model providers
llama-server exposes an OpenAI-compatible API at /v1/chat/completions,
so it works with provider: 'openai' + baseURL like Ollama and vLLM.
Added to the supported providers table and the feature bullet.

Fixes #34
2026-04-04 10:49:21 -07:00
Ibrahim Kazimov 944918d5ae
Merge branch 'main' into main 2026-04-04 13:43:40 +03:00
JackChen 25b144acf3 chore: remove hardcoded cache bust param from Star History URLs 2026-04-04 18:29:11 +08:00
Marcelo Ceccon 10074c9b7d
feat(llm): add first-class Grok (xAI) support with dedicated GrokAdapter (#44)
feat(llm): add first-class Grok (xAI) support with dedicated GrokAdapter
2026-04-04 18:20:55 +08:00
JackChen 071d5dce61 docs: add Latent Space mention to README 2026-04-03 19:18:52 +08:00
JackChen 2253122515 docs: reposition README around TypeScript-native, goal-to-result, lightweight differentiators 2026-04-03 17:35:12 +08:00
JackChen 0111876264
feat: add onTrace observability callback (#18)
Add lightweight onTrace callback to OrchestratorConfig that emits
structured span events (llm_call, tool_call, task, agent) with timing,
token usage, and runId correlation. Zero overhead when not subscribed.

Closes #18
2026-04-03 15:28:59 +08:00
JackChen 17546fd93e docs: merge Gemma 4 examples, reorder README sections
- Merge examples 08 (runTasks) and 09 (runTeam) into a single Gemma 4 example
- Renumber: structured output → 09, task retry → 10
- Move Author and Contributors sections to bottom in both READMEs
- Add Author section to English README
2026-04-03 14:32:48 +08:00
JackChen 27c0103736 docs: add examples for structured output and task retry 2026-04-03 14:23:22 +08:00
JackChen 2187f3b310
docs: add Structured Output and Task Retry to feature list (#39) 2026-04-03 14:11:54 +08:00
JackChen 0db0a4d869 docs: adapt Star History chart for dark mode 2026-04-03 09:57:18 +08:00
JackChen 37bd56b193 docs: refresh Star History chart cache 2026-04-03 09:56:17 +08:00
JackChen 4fc7bb3f85 docs: add Gemma 4 to Model Agnostic feature description 2026-04-03 09:38:22 +08:00
JackChen d86ea766d3 feat: add Gemma 4 auto-orchestration example (runTeam)
Add examples/09-gemma4-auto-orchestration.ts demonstrating runTeam()
with Gemma 4 as the coordinator — the framework's key feature running
fully local. The coordinator successfully decomposes goals into JSON
task arrays, schedules dependencies, and synthesises results.

Verified on gemma4:e2b (5.1B params) with Ollama 0.20.0-rc1.
2026-04-03 09:28:45 +08:00
JackChen f9fcac0c38 feat: add Gemma 4 local model example with tool-calling
Add examples/08-gemma4-local.ts demonstrating a pure-local multi-agent
team using Gemma 4 via Ollama — zero API cost. Two agents (researcher +
summarizer) collaborate through a task pipeline with bash, file_write,
and file_read tools. Verified on gemma4:e2b with Ollama 0.20.0-rc1.

Update both READMEs: add example 08 to the examples table and note
Gemma 4 as a verified local model with tool-calling support.
2026-04-03 09:12:44 +08:00
MrAvalonApple 4e4226783d docs: update README to include Gemini model support 2026-04-02 22:06:55 +03:00
MrAvalonApple ec6be79dc4 Merge remote-tracking branch 'upstream/main' 2026-04-02 21:56:17 +03:00
JackChen 8d27c6a1fe docs: add supported providers section and clarify contributing guidelines
- Add Supported Providers table with 4 verified providers (Anthropic, OpenAI,
  Copilot, Ollama) and note that other OpenAI-compatible providers are unverified
- Update Contributing to distinguish baseURL verification (#25) from new adapters
- Note that local models via Ollama require no API key in Quick Start
2026-04-03 02:47:02 +08:00
MrAvalonApple 2bb220fc63 feat: add support for Gemini model 2026-04-02 21:41:56 +03:00
JackChen 94cccf24d7 docs: replace inline code examples with examples/ index table
Remove ~160 lines of duplicated code snippets from both READMEs.
Link to the runnable scripts in examples/ instead — single source of truth,
type-checked by npm run lint.
2026-04-03 02:18:58 +08:00
JackChen 54400580a2 docs: fix source file count in README (20 → 27) 2026-04-03 00:05:31 +08:00
JackChen 01b93d9897 docs: lead Quick Start with runTeam, add output example and Node version 2026-04-02 22:35:45 +08:00
JackChen 736121fe10 docs: add lightweight positioning tagline to README headers 2026-04-02 22:24:08 +08:00
JackChen 9f9f4e95d4 docs: highlight auto task decomposition, add run modes table, move contributors up
- Rewrite headline to emphasize automatic goal-to-task decomposition
- Add "Auto Task Decomposition" as first item in Why section
- Add "Three Ways to Run" table (runAgent / runTeam / runTasks)
- Move Contributors section higher for visibility
- Bust star-history cache to refresh chart
2026-04-02 21:55:04 +08:00
JackChen a32440728f docs: update READMEs for baseURL support and local model examples
- Add Ollama/local model agent example in multi-model teams section
- Update "Model Agnostic" description to mention local models and baseURL
- Update contributing section to reflect built-in OpenAI-compatible support
- Add author block with Xiaohongshu link in Chinese README
2026-04-02 19:42:59 +08:00
JackChen 7acd450707 Update READMEs: add Copilot adapter, contributors section 2026-04-02 12:20:21 +08:00
JackChen ad317610ff Bust GitHub cache for star history chart, ignore non-tech dirs 2026-04-02 00:56:15 +08:00
JackChen 376785c3fa Add Chinese README and GitHub topic tags for discoverability 2026-04-01 16:24:43 +08:00
JackChen ae6968a52c Publish to npm as @jackchen_me/open-multi-agent 2026-04-01 13:44:23 +08:00
JackChen 0c02327a56 Switch install method to GitHub, remove npm badges 2026-04-01 13:14:29 +08:00
JackChen 3b24b99056 Improve README: clearer intro, fewer code walls, contributing section 2026-04-01 13:04:40 +08:00
JackChen 6cbc5582fe Add star history chart to README and update .gitignore 2026-04-01 12:12:28 +08:00
JackChen a6244cfe64 Initial release: open-multi-agent v0.1.0
Production-grade multi-agent orchestration framework.
Model-agnostic, supports team collaboration, task scheduling
with dependency resolution, and inter-agent communication.
2026-04-01 04:33:15 +08:00