← back

Show HN: AgentForge – Multi-LLM Orchestrator in 15KB

chunktort | 2026-02-18 02:18 UTC | source

ChunkyTortoise/ai-orchestrator

1 points | 0 comments | original link
I built AgentForge, a minimal multi-LLM orchestrator. Total size: ~15KB of Python code.

Why? LangChain added 250ms overhead per request. I needed something simpler.

Performance vs LangChain (1,000 requests): - Avg latency: 420ms -> 65ms - Memory/request: 12MB -> 3MB - Cold start: 2.5s -> 0.3s - Test time: 45s -> 3s

Size: 15KB + 2 dependencies (httpx, pytest) vs LangChain's 15MB+ and 47 packages.

In production: 89% LLM cost reduction via 3-tier Redis caching (88% cache hit rate, verified benchmarks). 4.3M tool dispatches/sec in the core engine.

What it does: 1. Multi-agent orchestration -- route tasks to specialized agents with automatic fallbacks 2. Testing built-in -- MockLLMClient lets you assert agent behavior without API keys 3. Production patterns -- circuit breakers, rate limiting, caching included

Install: pip install agentforge

Live demo: https://ai-orchest-7mnwp9untg7gyyvchzevid.streamlit.app/

Packaged version with docs and deployment guide: https://chunkmaster1.gumroad.com

When to use it: production reliability matters, latency is a concern, you want full test coverage. When not to: prototyping, internal tools, team already on LangChain ecosystem.

Questions welcome.