
CAMP: AI’s First Case-Adaptive Clinical Panel📷 Source: Web
- ★Dynamic specialist assembly per case
- ★Three-valued voting prevents false consensus
- ★Hybrid router tailors diagnostic workflows
Large language models applied to clinical prediction have a dirty little consistency problem: swap a comma in the prompt, and a complex case can flip from sepsis to gastritis. Most multi-agent frameworks paper over this by slapping fixed roles together and calling it deliberation—then averaging away the very disagreements that should signal diagnostic uncertainty. The team behind arXiv 2604.00085v1 proposes CAMP (Case-Adaptive Multi-agent Panel), an attending-physician agent that dynamically assembles a specialist panel sized to the case’s specific entropy.
Instead of a binary vote, specialists get three options—KEEP, REFUSE, NEUTRAL—so a cardiologist can abstain on a neurology question without poisoning the consensus. The attending agent uses a hybrid router to funnel each case to the right mix of generalist and specialist models, effectively trading static majority voting for an emergent confidence-weighted ensemble. Early synthetic benchmarks show a 12-point lift in F1 scores on disputed cases, but the real story is the shift from forced agreement to principled disagreement.

CAMP: AI’s First Case-Adaptive Clinical Panel📷 Source: Web
The gap between benchmarks and bedside just got a new protocol
That 12-point lift isn’t just another synthetic benchmark; it’s the first time a clinical AI framework has explicitly encoded abstention as a valid diagnostic signal. The hybrid router isn’t merely routing—it’s curating what the team calls a ‘case-specific diagnostic quorum,’ effectively treating diagnostic uncertainty as a first-class citizen instead of noise to be averaged away. GitHub activity is still sparse, but the pull requests tell a story: developers are testing custom routers that can ingest EHR ontologies, hinting at deployment paths beyond the research sandbox.
Who wins here? EHR vendors suddenly have a new technical moat, but the real competitive pressure lands on single-agent diagnostic startups who’ve been selling consistency as a feature. Their ‘reliable’ outputs now look brittle next to a framework that can quantify when it’s unsure. For clinicians, CAMP flips the script: instead of forcing a black-box model to give one answer, it surfaces the edges of the model’s expertise in real time. That’s not a feature—it’s a fundamental rethink of what clinical AI should deliver.