$ swarm command --scale 250 "audit auth system" π Hive activated Β· 250 agents Β· 15 models Β· 3 families consensus: 94% β Cross-family validation passed β Shadow score: 96/100 β 3 critical findings synthesized β Final report delivered in 4m 12s
For small tasks, a single AI is fine. But for security audits, architecture reviews, and migration strategies β one model means one blind spot, one context window, one confident-sounding answer with no independent check. You need consensus from independent minds that verify each other's work.
One command. Tell the swarm what you need β security audit, code review, architecture analysis. Plain English.
Agents from Claude and GPT families compete and collaborate. Different models cross-pollinate and review each other's work.
Only findings validated across model families survive. Shadow scores gate quality. One synthesized answer emerges from the colony.
15 models, not one. Claude Opus & Sonnet. GPT-5.x series. Claude Haiku. Each brings different strengths β together they catch what any single model misses.
Different model families review each other's work. Claude checks GPT. GPT checks Claude. No echo chambers β only findings that survive independent scrutiny make the cut.
Hidden quality gates you can't game. Every agent is scored β failuresΒ Γ·Β totalΒ ΓΒ 100 β and they don't know they're being watched. Bad work gets caught automatically.
Every commander runs in its own context window.
Different model families ensure diverse perspectives.
5 Commanders Γ ~50 workers each = 250 agents, each with its own context window
Multiple independent minds converge on one synthesized truth.
Thorough analysis with full cross-family validation. Architecture reviews, security audits, migration planning.
$ swarm command --scale 100 "audit security posture"Every layer of the swarm is engineered to maximize signal while minimizing spend. Here's how.
Context shrinks at every layer β 128KΒ tokens at the Nexus compresses to just 128Β tokens at each worker. Parents strip rationale, narrow file scope, and tighten constraints so children only receive the bytes they need.
A three-state FSM (Closed β Open β Half-Open) monitors every layer. If 50-60% of agents fail, the breaker trips β no new agents spawn, costs stop climbing, and a recovery probe tests before the swarm resumes.
Timeout cascade (90β60β40β30s), token ceilings per layer, output size caps, retry budgets, a concurrent-agent cap of 50, and a hard cost ceiling ($5β$20 depending on scale) that kills all agents if breached.
Agents launch in three waves β CanaryΒ (1), ProbeΒ (3), Remainder β with health gates between each. If the canary fails, the full pod never deploys. One cheap test prevents many expensive failures.
Workers use Haiku and GPT-Mini β the lightest, cheapest models. Expensive Opus and Sonnet reasoning is reserved for Commanders and the Nexus where it matters most. 60% of agents cost 10Γ less.
SS-50 runs $1.50β$3.50. SS-100 runs $3.50β$8. SS-250 runs $8β$16. Hard ceilings at $5, $10, and $20 guarantee you never get a surprise bill β even if every agent retries at maximum.
| Scale | Agents | Typical Cost | Hard Cap | Wall-Clock |
|---|---|---|---|---|
| SS-50 | ~36-52 | $2.50 | $5 | ~30s |
| SS-100 | ~89 | $5.50 | $10 | ~45s |
| SS-250 | ~316 | $10 | $20 | ~65β90s |
Progressive refinement: discover β validate β confirm
consensus = confidenceΓ0.40 + evidenceΓ0.30 + scopeΓ0.15 + coverageΓ0.15 β conflict_penaltyOne command. Then type swarm command.
curl -fsSL https://raw.githubusercontent.com/DUBSOpenHub/swarm-command/main/quickstart.sh | bashRequires an active Copilot subscription