Every model gives you
a different answer.
Get the right one.
Stop copy-pasting prompts across OpenAI, Claude, and Codex. Stop comparing outputs in different tabs. Concilium runs them all in parallel, has them peer-review each other's work, and gives you one validated answer — in a single interface.

The Problem
You already know one model isn't enough. So you open multiple terminals, paste the same prompt into Claude, Opencode, and Codex, then spend 20 minutes reading and comparing their outputs. There has to be a better way.
How It Works
Parallel Execution
Send one prompt and three agents start working at the same time. No more opening multiple terminals or browser tabs.
» Claude, OpenAI, and OpenCode run in isolated subprocesses. You watch all three stream output simultaneously in the same window.
Blind Review
Instead of you reading and comparing outputs, multiple juror models do it for you. Anonymously. With no bias.
» Responses are labeled A, B, C. Jurors evaluate correctness, edge-case handling, and code quality. You skip the manual comparison entirely.
Synthesis
A Chairman model merges the strongest parts of each solution into one final answer. Better than any single model alone.
» The result combines the best architecture decisions, error handling, and implementation details from all three agents.
Why Concilium?
- ✕ Copy-pasting the same prompt into 3 different tools.
- ✕ Switching between browser tabs, terminals, and apps.
- ✕ Reading 3 long outputs and comparing them manually.
- ✕ No way to know which answer has the fewest bugs.
- ✓ One prompt, three agents, all running at the same time.
- ✓ A single desktop app with a unified interface.
- ✓ Automatic blind peer-review ranks the best response.
- ✓ Synthesized output validated by adversarial consensus.
Watch Concilium in Action
See how Concilium orchestrates multiple LLMs to reach consensus on complex coding tasks.
Full demo walkthrough • 2:15
Built-in Analytics
Every council run generates detailed telemetry. Concilium captures token usage, costs, timing, and rankings — then surfaces them in a full analytics dashboard so you can make informed decisions about your AI workflow.
Token Usage Breakdown
Track input and output tokens for every agent, juror, and chairman. Grouped bar charts show exactly where your tokens go, split by model across the entire pipeline.
Cost Analysis
Total spend, average cost per run, cost per 1k tokens, and the most expensive model — all at a glance. Cost efficiency rankings show which models give the best value for money.
Performance & Rankings
Win rates show which models get ranked #1 most often. Average ranking tables, quality-per-dollar efficiency scores, and total #1 counts help you pick the best agents.
Model Comparison & History
Full comparison table with runs, tokens, cost, average time, and average rank per model. Plus a sortable run history with status, prompt, duration, and cost for every council.
Why Analytics Matter
Know exactly which models give the best answers per dollar.
Stage timing reveals where the pipeline slows down.
Data-driven decisions on which agents to enable.
See how your usage, costs, and quality evolve run-over-run.
All analytics are computed locally from your run history. No data ever leaves your machine.
Open Source
matiasdaloia/concilium
MIT License
The entire codebase is open source. Browse the code, report issues, or contribute features.
Get Involved
- → Report bugs
Found something broken? Open an issue.
- → Submit PRs
Fix a bug or add a feature.
- → Request features
Have an idea? Start a discussion.
- → Join discussions
Shape the future of collective AI.
Terminal First
Same deliberation engine, zero GUI required. Run pipelines from your terminal, pipe results to other tools, or embed Concilium in scripts with the programmatic API.
The recommended approach combines a token bucket algorithm
with Redis-backed distributed state. Key improvements from
the top-ranked response (Response A):
run Run a full deliberation from the terminal --agents Choose agents (claude, codex, opencode) --json Machine-readable output for pipelines --output Save synthesis to a file history Browse and replay past deliberations --last Show the most recent run --synthesis Print only the final answer --json Export as JSON config Manage API keys, models, and preferences set api-key Store your OpenRouter key set jurors Default juror models set chairman Default chairman model models Discover available models from agents and OpenRouter --agent Filter by agent provider --council Browse OpenRouter catalog --json JSON output for scripting Programmatic API
Embed deliberations in scripts, CI pipelines, or agent skills with one function call.
Concilium Cloud
The same multi-model deliberation engine — without the setup. Concilium Cloud brings team collaboration, API access, and managed infrastructure so you can focus on shipping better code.
Zero Setup
No API keys to manage, no local agents to install. Sign in and start deliberating in seconds.
Team Collaboration
Share deliberation results, manage team API usage, and build shared prompt libraries across your org.
REST API
Trigger deliberations from CI/CD pipelines, scripts, or agent skills with a single API call.
Get Early Access
Be the first to try Concilium Cloud when it launches.
Get Started
Install the CLI globally, build the desktop app, or embed deliberations programmatically — pick the path that fits your workflow.
Install globally
Requires Node.js 18+ and at least one CLI agent installed:
claude,
codex, or
opencode.
Set your API key
Uses OpenRouter to access models for peer review and synthesis.
Run a deliberation
Agents run in parallel, review each other's work, and a chairman synthesizes the best answer.
Clone the repository
Requires Node.js 18+ and macOS 12+ or Linux.
Configure & build
Launch the app
At least one CLI agent must be installed:
claude,
codex, or
opencode.
Add to your project
Call deliberate()
Embed deliberations in scripts, CI pipelines, or agent skills with one function call.