9 providers. 38 models. Quality-tuned prompting with chain-of-thought reasoning and auto-fallback. The same agentic architecture as Claude Code — bash, file ops, code search, streaming — running on Gemini, Llama, Mistral, Qwen, DeepSeek, and more.
Model quality is the bottleneck. These techniques compensate — disciplined prompting makes weaker models behave like stronger ones.
Every provider speaks the OpenAI wire format. One universal adapter with different base URLs. If one provider rate-limits, the system silently falls back to the next.
Based on the claurst reverse-engineering of Claude Code internals. 12 of 16 features shipped. 4 planned.
| Feature | Claude | Ours | Note |
|---|---|---|---|
| Agentic Tool Loop | ✓ | ✓ | LLM → tools → LLM, up to 15 turns |
| Bash Execution | ✓ | ✓ | Shell commands with timeout + cwd |
| File Read/Write/Edit | ✓ | ✓ | Find-and-replace, line numbers, create dirs |
| Code Search (grep/glob) | ✓ | ✓ | ripgrep with glob filters |
| Streaming Output | ✓ | ✓ | Real-time token streaming via SSE |
| Multi-Provider (9) | — | ✓ | Hot-switch with /provider command |
| 38 Models | — | ✓ | Gemini, Llama, Mistral, Qwen, DeepSeek, Gemma |
| Auto-Fallback | — | ✓ | 429? Silently tries next provider |
| Quality System Prompt | ✓ | ✓ | 2000+ token behavioral instructions |
| Chain-of-Thought | ✓ | ✓ | Internal reasoning, clean output |
| Parallel Tool Calls | ✓ | ✓ | Concurrent execution with semaphore |
| Conversation History | ✓ | ✓ | Multi-turn with full context |
| Coordinator Mode | ✓ | ◔ | Multi-agent orchestration (planned) |
| Memory Consolidation | ✓ | ◔ | autoDream pattern (planned) |
| Permission System | ✓ | ◔ | Risk classification (planned) |
| Browser Tool | ✓ | ◔ | Web browsing (planned) |
Claude Opus 4.6 is a frontier model. The architecture around it — the agentic loop, tool calling, streaming — is commodity code. ~1,550 lines of TypeScript replicates the entire pattern.
What makes it feel magical is the model quality — knowing when to read before editing, when to grep before writing, when to stop. Gemini 2.5 Pro gets you 85% of the way. Llama 3.1 405B on Together or SambaNova closes the gap further. Both are free.
Stack quality prompting + chain-of-thought + auto-fallback across 9 providers, and you have a system that produces frontier-competitive output from entirely open-source models. The remaining difference is taste, not capability.
Built on open-multi-agent (MIT). Internals informed by claurst. No proprietary code was used.