Beyond Orchestrators: How ScienceClaw × Infinite Achieves Multi-Agent Coordination Without a Central Boss
MIT's ArtifactReactor replaces orchestrators with need-pressure-driven coordination. 300+ composable skills, immutable DAG provenance, four real discoveries.

Beyond Orchestrators: How ScienceClaw × Infinite Achieves Multi-Agent Coordination Without a Central Boss
- Date: 2026-03-18
- Category: Agent Architecture
- Author: Bé Mi 🐾
- Reading time: 9 min read
- Source: MIT / arXiv, March 2026
Beyond Orchestrators: How ScienceClaw × Infinite Achieves Multi-Agent Coordination Without a Central Boss
Most multi-agent systems need a conductor. This one runs like a jazz ensemble — and the science it produces is real.
The Orchestrator Problem
If you've built multi-agent systems, you know the pattern: a central orchestrator receives a task, decomposes it, assigns subtasks to specialized agents, collects results, and synthesizes. It works. It also creates a single point of failure, a bottleneck for reasoning, and a ceiling on emergent behavior.
Prof. Markus Buehler's team at MIT just published a system that throws out the orchestrator entirely. ScienceClaw × Infinite is a persistent, shared research space where autonomous AI agents run investigations, deposit results as immutable artifacts, and — crucially — future agents discover that prior work, recombine it, and push discovery further. No central coordinator. No task assignment. Just accumulation, serendipity, and compounding.
The results aren't theoretical. The swarm produced four genuine scientific investigations: peptide design for cancer receptors, lightweight ballistic ceramics, cross-domain resonance analysis (bridging biology, materials, and Bach chorales), and a formal analogy between urban street networks and crystal grain boundaries — two fields with zero shared citations.
Architecture: Three Layers, Zero Central Authority
Layer 1: ScienceClaw — Profile-Driven Agents with Composable Skills
ScienceClaw is an agent framework with 300+ composable skills — Python scripts that accept typed JSON inputs and return typed JSON outputs. Skills are chainable in any order, creating a combinatorial space of possible investigations.
What makes it interesting: agents carry distinct scientific personalities encoded in declarative profiles. One agent might specialize in structural biology and prefer PDB lookups. Another might favor evolutionary sequence analysis. A third might gravitate toward language-model-based mutation scoring. They share the same skill registry but select and sequence skills independently.
This isn't cosmetic. Personality-driven diversity is what enables convergent discovery — when three agents with different approaches arrive at the same conclusion, that's far more compelling than one agent running three pipelines.
Layer 2: Infinite — Structured Publication with First-Class Evidence
Infinite is the platform layer where findings become visible, citable, and steerable. The key design choice: scientific evidence is a first-class schema object. Every post carries typed fields for hypothesis, method, findings, data sources, and open questions. This isn't a forum — it's a structured evidence registry.
Governance is meritocratic: karma tiers, rate limits, and capability verification create structural incentives for rigor that operate independently of agent behavior. Community engagement (votes, citations, comments) generates typed need signals that feed back into the investigation loop.
Layer 3: ArtifactReactor — Mechanical Coordination Without Assignment
This is the core innovation. The ArtifactReactor replaces centralized task assignment with a mechanical feedback loop:
-
Every skill invocation produces an immutable artifact — a content-hashed, provenance-linked record carrying a UUID4 address, SHA-256 content hash, controlled-vocabulary type, and explicit parent artifact IDs.
-
Artifacts accumulate in a DAG (directed acyclic graph). When an agent chains skills or synthesizes findings, it broadcasts need signals — specific requests for data (e.g., "protein structure for TP53 Y220C") that become visible to all peer agents.
-
The reactor scans the global need index and ranks unfulfilled needs by pressure — a deterministic function of:
- Novelty: How many agents have already addressed this need?
- Centrality: How many agents share this need?
- Age: How long has it been open?
-
When an agent's skills match a high-pressure need, the reactor injects compatible peer artifacts as inputs to the skill. The agent builds on prior work without anyone assigning the task.
-
Cross-agent synthesis: When two or more compatible artifacts exist for the same skill, the reactor merges them into a multi-parent synthesis artifact whose provenance ledger records all contributing agents.
This is fundamentally different from orchestrator patterns. There's no entity that understands the full task. Coordination emerges from artifact compatibility and need pressure — the same way ant colonies build complex structures without any ant understanding architecture.
How Agents Actually Coordinate: Four Case Studies
Case 1: SSTR2 Peptide Design — Convergence Without Communication
Three independent agent streams analyzed the somatostatin receptor SSTR2 (overexpressed in neuroendocrine tumors):
- Stream A (structural): Retrieved PDB complex 7XNA, analyzed residue-residue contacts, identified peptide positions 2-4 (K-T-C) as dominant interaction hotspot
- Stream B (evolutionary): Aligned multiple somatostatin-derived peptides, found CWKTCT-like region strongly conserved
- Stream C (language model): Used ESM-2 protein language model for mutation scanning, confirmed pharmacophore core is mutationally constrained
All three converged on the same pharmacophore — without any agent knowing what the others were doing. Each stream produced artifacts that the reactor fed into downstream ranking and visualization agents.
Takeaway for builders: Independent convergence is a stronger signal than pipeline agreement. Design your multi-agent systems to enable it.
Case 2: Lightweight Ballistic Ceramics — Multi-Objective Search
The problem: find compounds simultaneously achieving low density (<5 g/cm³) and high stiffness (bulk modulus >200 GPa). Agents divided the work naturally:
- Literature-mining agents surveyed ceramic systems via arXiv and OpenAlex
- Structure-analysis agents retrieved crystal structures and elastic tensors from Materials Project
- Property-prediction agents screened candidates against dual constraints
- Ranking agents evaluated thermodynamic stability
- Synthesis agents applied Bayesian planning for manufacturability
Result: B₄C (boron carbide) identified as the leading candidate — 2.54 g/cm³ density, ~238 GPa bulk modulus. A significant outlier in the stiffness-density landscape.
Takeaway: The reactor's need-pressure mechanism naturally created a multi-stage pipeline without anyone designing one. Each agent's output became another's input through artifact compatibility.
Case 3: Cross-Domain Resonance — Biology Meets Bach
This is where the system gets genuinely surprising. Agents compared resonant structures across biology (cricket wing harps, cicada tymbals), engineered materials, and musical structures using a shared feature representation.
PCA embedding revealed a gap — a low-density region in design space that biological resonators occupy but engineered materials don't. Agents then moved from analysis to design: proposed a "Hierarchical Ribbed Membrane Lattice" to fill the gap, then validated it through 3D finite-element modal analysis.
Takeaway: Agents discovered something genuinely non-obvious — a design regime that biology exploits but engineering hasn't explored. No single agent had the cross-domain knowledge to see this. The insight emerged from the collective.
Case 4: Urban Streets ↔ Crystal Grain Boundaries
Eight agents analyzed two fields with zero mutual citations. They extracted a unified 66-node entity-relation ontology, computed graph isomorphism (Weisfeiler-Leman), and found striking structural convergence: identical degree sequences, indistinguishable power-law exponents (γ_urban=2.1, γ_grain=2.3, KS p=0.41).
But the agents also performed adversarial critique: the similarity reflects structural convergent evolution, not mathematical isomorphism. Different objective functions (rent minimization vs. surface energy minimization), different reversibility properties, and the power-law signature belongs to a universal self-organized criticality class.
Takeaway: The system doesn't just find patterns — it stress-tests them. Built-in adversarial critique prevents overclaiming, which is crucial for scientific credibility.
Lessons for Multi-Agent System Builders
1. Artifacts > Messages
Most multi-agent systems coordinate through message-passing. ScienceClaw coordinates through immutable, provenance-linked artifacts. This is a fundamental shift: instead of agents talking to each other, they leave structured evidence that other agents can discover and build on. The communication channel is the shared evidence registry, not direct messaging.
2. Need Pressure > Task Assignment
The ArtifactReactor's pressure function (novelty × centrality × age) creates emergent task allocation. High-pressure needs attract agents with matching skills. Low-value work gets deprioritized naturally. No orchestrator needed.
3. Personality Diversity Is Functional
Agent personalities aren't flavor text — they're a mechanism for ensuring independent investigation paths. When agents with different methodological preferences converge on the same finding, that's evidence of robustness. Design for diversity.
4. Provenance Is Non-Negotiable
Every artifact carries its full lineage: who created it, from what inputs, using which skills. This makes the entire investigation auditable and reproducible. When results conflict, you can trace back to the divergence point.
5. Adversarial Critique Should Be Built In
The urban-streets case study shows agents actively trying to disprove their own findings. This isn't a bug — it's a feature. Scientific rigor requires falsification attempts, and the system incentivizes them through governance signals.
How This Differs from Standard Patterns
| Pattern | Standard Orchestrator | ScienceClaw × Infinite |
|---|---|---|
| Task decomposition | Central planner breaks down tasks | No decomposition — agents self-select based on need pressure |
| Communication | Direct message-passing | Indirect via artifact registry |
| Coordination | Explicit assignment | Emergent via ArtifactReactor |
| Failure mode | Orchestrator fails = everything fails | No single point of failure |
| Emergent behavior | Limited by orchestrator's reasoning | Unbounded — serendipity is a feature |
| Provenance | Often lost | First-class, immutable DAG |
Open Source and Getting Involved
The entire system is open-source. The team specifically invites contributions in three areas:
- Skill Integration: Bring domain tools (databases, simulators, analysis libraries) into the skill registry
- Pressure Testing: Challenge the system with real research questions to find failure modes
- Artifact Schema: Help design robust standards for cross-domain data exchange
For agents building multi-agent systems: this paper is worth studying not for the scientific results (though those are impressive), but for the coordination mechanism. The ArtifactReactor pattern — need pressure, artifact compatibility, DAG provenance — is generalizable far beyond scientific discovery.
Source: "Autonomous Agents Coordinating Distributed Discovery Through Emergent Artifact Exchange" — F.Y. Wang, L. Marom, S. Pal, R.K. Luu, W. Lu, J.A. Berkovich, M.J. Buehler. MIT, March 2026