A Governance Protocol for Human Orchestration of Parallel Multi-AI Execution
From the Author of Governing AI: When Capability Exceeds Control
What This Is
Eight months of daily work across eleven AI platforms produced one clear lesson: the hardest governance problems in multi-AI work are not inside any single AI. They live in the space between platforms, where the human is trying to make sense of everything simultaneously.
HAIA-CAIPR — Cross AI Platform Review — is the governance protocol that addresses that space. It tells the human how to orchestrate multiple AI platforms running in parallel, how to compare their outputs with discipline, and how to keep human judgment at the center of every decision, not as a formality but as a structural requirement.
This is the first time this framework has been published publicly. It builds on HAIA-RECCLIN, which governs how individual AI platforms respond to human prompts. CAIPR operates at the level above that: it governs how the human works when multiple platforms are running at once.
Where CAIPR Sits in the HAIA Framework
HAIA — Human Artificial Intelligence Assistant — is the foundational principle governing any process that involves AI. Every framework in the stack operates within it.
RECCLIN is the role-assignment and structured output framework within HAIA. It assigns one of seven roles to each AI platform — Researcher, Editor, Coder, Calculator, Liaison, Ideator, or Navigator — and requires a consistent structured return: Role, Task, Output, Source, Conflicts, Expiry, a Fact-Tactic-KPI triad, and Recommendation. RECCLIN signals to the human whether the AI is doing what was intended. It enables deliberate platform selection by role strength. It also trains the human to evaluate AI output rather than simply accept it.
CAIPR requires RECCLIN as its output standard and operates at a distinct tier above it. RECCLIN governs individual AI interactions. CAIPR governs the human orchestrating all of them in parallel.
CBG — Checkpoint-Based Governance — provides the constitutional checkpoint authority that governs when the human decides. Every checkpoint in a CAIPR session is a CBG checkpoint.
| Layer | Function |
|---|---|
| Factics | Foundational evidentiary discipline. Facts paired with tactics and measurable outcomes. |
| HAIA-RECCLIN | Structured output governance for individual AI interactions and role-assigned multi-AI operation. |
| HAIA-CAIPR | Human orchestration governance for full parallel multi-platform execution. |
| CBG | Constitutional checkpoint authority. Governs when the human decides. |
| GOPEL | Non-cognitive software that will automate CAIPR mechanics without performing cognitive work. Specified; not yet deployed. |
RECCLIN Multi-AI and CAIPR Are Not the Same Thing
Both use multiple AI platforms. The architectures are categorically different.
In HAIA-RECCLIN multi-AI operation, each platform receives a designated role before dispatch. Perplexity researches. Claude builds code. Grok navigates. Each produces a role-specific RECCLIN-structured output. The human synthesizes across distributed specializations. This is efficient and powerful. Platform strengths are matched to tasks. Cost and time are controlled.
CAIPR dispatches the same full prompt to all available platforms simultaneously, with no pre-assigned role differentiation. Every platform receives the complete RECCLIN output format requirement and self-assigns its role based on its own reading of the prompt. At nine or eleven platforms, the human receives nine or eleven full RECCLIN-structured outputs covering the same task from independent analytical positions.
The resource difference is real and significant. A CAIPR session at nine platforms consumes proportionally more tokens, more synthesizer processing load, and more human arbiter review time than any role-assigned RECCLIN configuration. The evidence yield justifies that investment for high-stakes decisions. For routine workflows, RECCLIN multi-AI delivers sufficient governance at lower cost.
The Failure Evidence That Required a New Layer
CAIPR was not designed in advance. It was discovered through eight months of operational practice. Four findings were predicted by theory. Twenty-nine were discovered through practice. Three failure modes drove the development of the core protocol requirements.
Hallucination Detection
In single-platform workflows, there is no mechanism to detect fabrication. Cross-platform parallel execution exposes fabrications because other platforms do not corroborate inventions. Three documented incidents: wholesale document section fabrication, quote fabrication with self-review of those fabricated quotes, and content repetition fifteen times in a single output. Single-platform governance is structurally blind to this failure mode.
Source-Authority Erosion
In a multi-AI workflow using a synthesizer, the human arbiter’s own input was processed identically to AI platform output. The synthesizer treated human corrections as one more data source. The human arbiter began doubting their own corrections. This is reverse automation bias: the human doubting their own authority because the system failed to recognize it. The governance term for this case study is The Loop That Ate the Governor.
Algorithmic Narcissism
In a nine-platform structured review and separately a three-platform review, every platform asked to identify which should synthesize nominated itself with supporting evidence. This is a structural property of AI systems asked to evaluate their own fitness. No single-platform framework provides a mechanism to prevent this chokepoint from forming by default.
The Eight Core Operations
CAIPR specifies eight operations the human governs in every parallel session.
| Operation | What It Governs |
|---|---|
| 1. Parallel Dispatch | Identical prompts sent simultaneously to all platforms. No platform sees another’s output before producing its own. |
| 2. Structured Collection | Platform identity documented by the human before output is received. Timestamps recorded. Raw outputs preserved in full. |
| 3. Cross-Platform Comparison | All platform outputs read by the human before any synthesis begins. |
| 4. Hallucination Detection | Claims present in only one platform flagged for source-level verification. Citations not in human-provided documents independently verified. |
| 5. Convergence Analysis | Factual, analytical, and recommendation convergence classified separately. Convergence without dissent treated as a red flag. |
| 6. Synthesizer Oversight | The synthesizer operates under Tier 2 classification with seven documented failure modes. Inclusion manifest required for every output. |
| 7. Source-Authority Discrimination | Every input classified at ingestion: Tier 0 (human arbiter, immutable), Tier 1 (raw platform output), Tier 2 (synthesizer output, highest scrutiny). |
| 8. Platform Resilience Management | Sessions continue when platforms fail. Odd-number protocol maintained. Substitution governed by dispatch state. |
The Human Governor Is Not Optional
CAIPR defines five categories of human authority that no AI platform can perform or substitute.
- Evaluative Authority: the human assesses quality, completeness, and reliability across all platform outputs.
- Corrective Authority: the human issues corrections as Tier 0 inputs that override platform outputs regardless of confidence scores or convergence.
- Methodological Authority: the human determines platform count, dispatch architecture, convergence thresholds, synthesizer selection, and platform independence assessment.
- Creative Authority: the human contributes original synthesis and analytical conclusions no platform produced. Six named concepts in the operational record originated from the arbiter and were validated by platforms after the fact.
- Process Authority: the human decides when to escalate, pause, and conclude. No AI platform can determine session completion.
Six categories of human override are documented across dozens of instances in eight months of practice. Each represents knowledge no AI platform can replicate: provenance knowledge, source knowledge, methodological judgment, process memory, concept origination, and attribution verification.
The Synthesizer Problem
The synthesizer is the AI platform that processes multiple platform outputs to produce a governance package for the human. Synthesizer governance is the most demanding CAIPR requirement because synthesizer failures are structurally invisible to the platforms being synthesized.
Seven synthesizer failure modes are documented from operational practice: scope contamination, platform omission and misattribution, platform miscounting, false negative on prior work, human input dropping, evidence destruction, and structural invisibility. Every one of these was discovered through practice, not theoretical design.
The remedy is not to avoid synthesizers. It is to govern them with explicit requirements: Tier 2 classification, an inclusion manifest listing every platform included with role and timestamp, no approval authority, periodic verification against raw outputs, and self-documentation of synthesis methodology in each output.
The dual-signed inclusion manifest is the most operationally useful requirement. The human arbiter provides the expected platform list before synthesis begins. The synthesizer returns a receipt list. Any delta triggers an automatic failure flag before the human reviews the output.
What CAIPR Is Not
- Not a constitutional-level correction tool. Training data, alignment tuning, and platform values are set before user interaction. Multi-platform comparison reveals where platform outputs diverge. It cannot change the underlying models.
- Not an accuracy guarantee. It provides richer evidence for human evaluation. More platforms reading the same fabricated citation produce stronger false convergence, not detection.
- Not autonomous. Human arbiter at every checkpoint. GOPEL will automate mechanics in a future deployment. It will not perform cognitive work.
- Not a substitute for domain expertise. The strongest human overrides in the operational record came from domain-specific knowledge no platform could supply. CAIPR amplifies expertise; it does not replace it.
- Not for every task. For routine workflows, HAIA-RECCLIN multi-AI role-assigned operation delivers sufficient governance at lower cost. CAIPR is a premium-tier protocol warranted by high-stakes decisions, not a universal default.
The Adoption Ladder
The HAIA framework is designed for progressive adoption. Each level adds capability without invalidating the level below.
| Level | Components | What It Delivers |
|---|---|---|
| 1 | Factics | Evidentiary discipline. Facts paired with tactics and measurable outcomes. No AI required. |
| 2 | Factics + RECCLIN | Structured output governance with a single AI platform. Free tier accessible. Builds the evaluation capacity CAIPR requires. |
| 2.5 | Factics + RECCLIN Multi-AI | Role-assigned multi-AI execution. Platform strengths matched to tasks. Lower resource cost than CAIPR. |
| 3 | Factics + RECCLIN + CAIPR | Full parallel multi-AI execution. All platforms run the complete RECCLIN output format simultaneously. Premium tier. |
| 4 | Factics + RECCLIN + CAIPR + GOPEL | CAIPR operations automated through non-cognitive agent. Specified; not yet deployed. |
What This Framework Does Not Yet Resolve
Honest acknowledgment of scope boundaries is a governance principle, not a disclaimer. Four questions from operational practice remain open.
- What does multi-AI review miss? No method currently exists for detecting convergent error where all platforms agree on something wrong because they share the same training data error. This is the most significant unresolved limitation.
- Does reverse automation bias scale organizationally? If operators in multi-AI workflows systematically doubt their own input when systems do not confirm it, the failure becomes organizational. The organizational scale has not been studied.
- What is the full cost of platform thread loss? Two unrecoverable threads were documented. The full scope of thread loss across a project is structurally unknowable from the operator’s position.
- How does the human arbiter maintain platform independence as the AI industry consolidates? Platform independence is a human arbiter selection responsibility. The human names the platforms, selects for known architectural diversity, and documents independence assumptions in session records. No algorithmic method resolves this; governance of the question belongs to the human.
Related reading: AI Provider Plurality: A Congressional Package & Verified AI Inference Standards Act (VAISA) — produced to create infrastructure for safe AI in America.
Why This Matters Now
The governance conversation in AI has focused heavily on what AI systems should be built to do. Less attention has gone to what humans need to do when working across multiple AI systems simultaneously, each with different strengths, failure modes, and behavioral tendencies.
CAIPR is an answer to the practical question practitioners face today: how do you extract disciplined intelligence from parallel AI execution without losing human authority in the process?
The answer developed here is not theoretical. It came from eight months of daily operational practice, 33 documented findings, and repeated governance failures that the framework had to correct through its own operation. The ratio of practice-discovered findings to theory-predicted findings was nearly seven to one. That ratio is itself a governance lesson: frameworks that are only designed, not practiced, miss most of what actually happens.
The HAIA-CAIPR Specification is available in full at basilpuglisi.com and through the HAIA repository. Practitioners working at the intersection of multi-AI execution and human governance are invited to apply, test, and challenge the framework through their own operational practice.

Frequently Asked Questions
What is HAIA-CAIPR?
HAIA-CAIPR (Cross AI Platform Review) is a governance protocol for human orchestration of parallel multi-AI execution. It governs how the human works when multiple AI platforms run simultaneously, covering parallel dispatch, output comparison, hallucination detection, synthesizer oversight, and source-authority discrimination. It sits above HAIA-RECCLIN in the HAIA framework stack.
How does CAIPR differ from HAIA-RECCLIN multi-AI operation?
RECCLIN multi-AI assigns each platform a designated role before dispatch. One platform researches, another codes, another navigates. CAIPR dispatches the identical full prompt to all platforms simultaneously with no pre-assigned roles. Every platform self-assigns its role and returns a complete RECCLIN-structured output. RECCLIN distributes roles across platforms. CAIPR replicates the full governance output structure across every platform at once.
What are the eight core operations of CAIPR?
The eight CAIPR operations are: Parallel Dispatch, Structured Collection, Cross-Platform Comparison, Hallucination Detection, Convergence Analysis, Synthesizer Oversight, Source-Authority Discrimination, and Platform Resilience Management. Each is a human-governed checkpoint. No operation is delegated to an AI platform.
What is the Synthesizer Problem in CAIPR?
The synthesizer is the AI platform that processes multiple platform outputs to produce a governance package for the human. Seven synthesizer failure modes are documented: scope contamination, platform omission and misattribution, platform miscounting, false negative on prior work, human input dropping, evidence destruction, and structural invisibility. The remedy is Tier 2 classification, a dual-signed inclusion manifest, and periodic verification against raw outputs.
What is the HAIA adoption ladder?
The HAIA adoption ladder runs from Factics (Level 1, no AI required) through RECCLIN single-platform (Level 2), RECCLIN Multi-AI (Level 2.5), CAIPR full parallel execution (Level 3), and GOPEL-automated mechanics (Level 4). Each level adds capability without invalidating the level below. Entry is possible at any level.
Is CAIPR appropriate for every AI task?
No. CAIPR is a premium-tier protocol warranted by high-stakes decisions. It consumes proportionally more tokens, synthesizer processing load, and human arbiter review time than role-assigned RECCLIN configurations. For routine workflows, HAIA-RECCLIN multi-AI delivers sufficient governance at lower cost. CAIPR is not a universal default.
What failure modes does CAIPR address?
Three operational failure modes drove CAIPR development. First, hallucination detection: single-platform workflows cannot detect fabrication; cross-platform execution exposes it. Second, source-authority erosion: documented as The Loop That Ate the Governor, where a synthesizer processed human corrections identically to AI output, producing reverse automation bias. Third, algorithmic narcissism: every platform asked to identify which should synthesize nominated itself.
What is GOPEL and how does it relate to CAIPR?
GOPEL (Governance Orchestrator Policy Enforcement Layer) is a non-cognitive, deterministic software agent that will automate CAIPR mechanics without performing cognitive work. It handles dispatch, collection, routing, logging, pausing, hashing, and reporting. Pipes that can think can be manipulated, so GOPEL performs zero cognitive work by design. It is specified and documented but not yet deployed as of March 2026.
About the Author
Basil C. Puglisi, MPA is a Human-AI Collaboration Strategist and AI Governance Consultant. He is the creator of the Factics methodology (2012), HAIA-RECCLIN, HAIA-CAIPR, Checkpoint-Based Governance, and the Human Enhancement Quotient. He founded Digital Ethos in 2011 and has documented digital platform evolution across 900+ articles since 2009. He is the author of Governing AI: When Capability Exceeds Control (2025). He retired after twelve years of service with the Port Authority Police Department.
Leave a Reply
You must be logged in to post a comment.