Upholding Epistemic Agency: A Brouwerian Assertibility Constraint for Responsible AI

This academic paper proposes a formal 'assertibility constraint' requiring AI systems to provide publicly contestable proof for claims in high-stakes domains. The framework mandates a three-status interface (Asserted, Denied, Undetermined) where 'Undetermined' is required when proof is unavailable. The approach aims to preserve human epistemic agency by making AI outputs answerable to challengeable warrants rather than confidence scores alone.

Upholding Epistemic Agency: A Brouwerian Assertibility Constraint for Responsible AI

This academic paper proposes a fundamental shift in how high-stakes AI systems should communicate uncertainty, arguing that current systems' tendency to convert probabilistic uncertainty into definitive outputs undermines democratic discourse and public trust. The author introduces a formal "assertibility constraint" requiring AI to provide publicly contestable proof for its claims, a framework with profound implications for deploying generative AI in legal, medical, and governmental domains where justification matters as much as the answer itself.

Key Takeaways

  • The paper critiques how generative AI can displace "epistemic agency" by presenting uncertain outputs as authoritative verdicts, eroding the public's role in evaluating claims.
  • It proposes a formal "assertibility constraint": in high-stakes domains, AI may only assert or deny a claim if it can produce a publicly inspectable and contestable certificate of entitlement; otherwise, it must return "Undetermined."
  • This creates a mandatory three-status interface (Asserted, Denied, Undetermined), where "Undetermined" is not a tunable reject option but a required status when forcing proof is unavailable.
  • The framework is operationalized through "decision-layer gating," using internal mathematical witnesses (like sound bounds) and an output contract with reason-coded abstentions.
  • The core goal is to make AI outputs "answerable to challengeable warrants rather than confidence alone," preserving human oversight in public justification processes.

A Framework for Accountable AI Outputs

The paper, arXiv:2603.03971v1, presents a philosophical and technical framework to address a critical flaw in modern AI: its epistemic overreach. The author argues that by generating fluent, authoritative-sounding text from inherently uncertain internal processes, generative AI systems "displace the justificatory work on which democratic epistemic agency depends." In essence, they shortcut the public process of evaluating evidence and reasoning, potentially eroding trust and accountability.

As a corrective, the author proposes a Brouwer-inspired assertibility constraint. For any high-stakes claim, an AI system is only permitted to output a definitive "Asserted" or "Denied" if it can simultaneously provide a "publicly inspectable and contestable certificate of entitlement." This certificate acts as a "boundary object" connecting the system's internal reasoning to public scrutiny. If such a certificate cannot be generated, the system is constrained to return a third status: "Undetermined."

The technical implementation involves decision-layer gating. Instead of merely applying a confidence threshold or an argmax function to a model's logits to get a final answer, the system must check for the existence of an "internal witness"—a concrete, verifiable piece of evidence like a sound mathematical bound or a separation margin in a feature space. The output is then governed by a formal contract that includes "reason-coded abstentions" explaining why an "Undetermined" status was returned. A key design lemma proves that any system with a total, certificate-sound binary interface has, in fact, already decided the predicate on its declared scope. Therefore, the "Undetermined" status is not a flexible reject option to be tuned for accuracy but a mandatory epistemic duty when justification is absent.

Industry Context & Analysis

This paper directly challenges the prevailing paradigm in commercial AI, where systems are optimized to always provide an answer, often camouflaging uncertainty with high but meaningless confidence scores. Unlike OpenAI's ChatGPT or Google's Gemini, which are designed to generate coherent and helpful responses across a vast range of queries with minimal abstention, this framework mandates systematic and justifiable silence. This mirrors a growing but niche focus on uncertainty quantification in research models like DeepMind's Chinchilla or Meta's Llama 2, but goes far beyond statistical confidence to demand producible, verifiable proof.

The proposed "assertibility constraint" finds a philosophical cousin in the concept of interpretability or explainable AI (XAI). However, most XAI research, such as feature attribution methods (SHAP, LIME) or concept-based explanations, provides post-hoc rationalizations for a model's decision. The author's framework is more rigorous, requiring the proof (the certificate) to be a necessary precondition for the decision itself, not a subsequent justification. This aligns more closely with formal methods in safety-critical software, where correctness must be provably verifiable.

The practical implications are starkest in regulated industries. In medical diagnostics, an AI reading an X-ray might achieve 94% accuracy on a benchmark like CheXpert, but under this framework, it could only output "Positive for pneumonia" if it could point to a specific, contestable region and a validated margin of separation from healthy scans. In legal document review, an AI couldn't simply assert "this clause contains a liability risk" with 99% confidence; it would need to produce a certificate tracing that judgment to specific legal precedents or logical rules. This moves the industry benchmark from raw performance on tasks like MMLU (Massive Multitask Language Understanding) or HumanEval for code to a new metric: the rate of justifiably determined outputs versus necessary "Undetermined" returns within a high-stakes domain.

What This Means Going Forward

The immediate beneficiaries of this research are regulators, ethicists, and developers working on AI for governance, justice, and healthcare—domains where automated decisions enter the public record and require justification. It provides a formal vocabulary and technical blueprint for building epistemically humble AI that knows the limits of its justificatory power. This could lead to new regulatory standards, akin to the EU AI Act's requirements for high-risk AI systems, mandating not just accuracy but the ability to demonstrate and communicate evidential support.

For AI companies, adopting such a framework would represent a significant shift in product design and marketing. A model that frequently returns "Undetermined" in complex scenarios may be more trustworthy but could be perceived as less capable than a competitor that always gives a confident, if ungrounded, answer. This creates a tension between perceived utility and epistemic integrity. The market may begin to segment, with "certificate-driven AI" becoming a premium, trusted category for critical applications, while always-answering models remain in lower-stakes, conversational domains.

The critical trend to watch is whether this theoretical framework gains traction in real-world system design. Key indicators will be its adoption in academic challenges, its influence on DARPA or NSF-funded projects in assured AI, and whether any major lab (e.g., Anthropic with its constitutional AI focus) begins to prototype "certificate-gated" interfaces. If successful, it could redefine the contract between AI and society, transforming systems from oracles that speak to participants that must show their work, fundamentally preserving human epistemic agency in an automated age.

常见问题