Skip to content

Designing Incident Reporting Systems for Harms from General-Purpose AI

Conference: AAAI 2026 arXiv: 2511.05914 Code: None Area: AI Safety & Governance Keywords: AI Incident Reporting, AI Governance, Safety-Critical Systems, Policy Design, General-Purpose AI

TL;DR

Through a literature review and case studies of nine safety-critical industries (nuclear energy, aviation, healthcare, etc.), this paper proposes a seven-dimensional institutional design framework for AI incident reporting systems, providing systematic guidance for policy design of general-purpose AI incident reporting in the United States.

Background & Motivation

General-purpose AI (GPAI) systems, particularly large language models (LLMs), are causing a growing number of real-world harms:

  • Facilitating a $25.6 million financial fraud
  • Assisting in planning explosive attacks
  • Generating deepfake pornographic content
  • Accidentally deleting an entire company codebase
  • Exhibiting extortion and deception capabilities
  • Spreading election misinformation in the United States and globally

Key Challenge: According to Normal Accident Theory, serious incidents are inevitable over time in complex systems such as GPAI. Pre-deployment audits cannot prevent all incidents — emergent capabilities of LLMs may manifest unexpectedly after deployment, creating unforeseen incident types. Yet research on the institutional design of AI incident reporting mechanisms remains nearly absent.

Background: - As of July 2025, only China and the European Union mandate general-purpose AI incident reporting - Multiple legislative proposals exist in the United States but have not yet been enacted - Non-governmental databases (AIID, AIAAIC, AVID) exist but lack stakeholder buy-in - The literature broadly supports incident reporting, but systematic institutional design analysis is lacking

This paper fills this gap by providing the first systematic examination of institutional design choices for AI incident reporting systems and the conditions under which they apply.

Method

Overall Architecture

This paper employs a three-step case study methodology: 1. Selection of nine safety-critical industries as cases: nuclear energy, aviation, pesticides, pharmaceuticals, cybersecurity, dams, railways, occupational safety, and healthcare 2. Development of a seven-dimensional institutional design framework through literature review 3. Extraction of design considerations from the case industries, with discussion of the applicability of specific design choices to the AI context

Key Designs

Design 1: Seven-Dimensional Institutional Design Framework

The seven dimensions of the framework are as follows:

Dimension Definition Options
Policy Goal The objective the system seeks to achieve Safety learning or accountability
Reporting & Receiving Entities Participants who submit and receive reports Users, victims, third parties, companies, industry employees, government
Incident Types Categories of events to be reported Safety incidents, rights incidents, security vulnerability incidents
Risk Materialization Stage The risk stage at which events are reported Hazard → Situation → Near miss → Harm event
Reporting Mandatoriness Mechanisms for incentivizing reporting Mandatory (legally required) or voluntary
Reporter Anonymity Who can know the reporter's identity Public, confidential, anonymous
Post-Report Actions Actions taken upon receiving a report Information sharing, information disclosure, auditing, regulatory action

An important conceptual clarification offered in this paper is the distinction between AI issues/flaws and AI incidents: issues are system-level conditions (hazards) that serve as preconditions for incidents when exposed to external environments; incidents are events that have caused or may cause harm.

Design 2: Incident Lifecycle Model

The complete pathway from risk emergence to materialization: - Hazard: An intrinsic system condition that may lead to harm - Situation: A hazard exposed to an external environment - Near Miss: An event that could have caused harm but ultimately did not - Harm Event: An event that has actually caused harm

Hospital studies indicate that fewer than 1% of safety events cause serious injury, 18% cause minor injury, and 82% cause no injury. The frequency of near misses is estimated to be 300 times that of harm events. This suggests that AI near-miss reporting could be highly valuable for safety learning.

Design 3: Design Considerations Derived from Nine Industry Cases

Key design considerations extracted from the case studies include:

(1) Dual nature of policy goals: While safety learning and accountability can be pursued within the same system, dual-objective systems are rare in practice because the two goals tend to call for opposing design choices — for example, encouraging voluntary reporting conflicts with punitive enforcement. Establishing multiple single-objective systems may be necessary.

(2) Regulatory vs. non-regulatory operators: - Regulatory agency operation: typically used for mandatory reporting and accountability-oriented systems - Non-regulatory agency operation: typically non-punitive and learning-oriented. The FAA fully delegated its voluntary reporting system ASRS to NASA — an agency that does not regulate airlines — thereby fostering trust. ASRS has received over 2 million reports since 1975.

(3) Reporting coverage: Given the diverse forms of GPAI harms and complex supply chains, information must be collected from multiple parties (users, company employees, third parties, the public). The FAA maintains no fewer than 8 separate voluntary reporting programs for different occupational groups (dispatchers, air traffic controllers, pilots, etc.).

Loss & Training

(This paper is a policy research study and does not involve model training. The methodological strategy is described here.)

The study employs qualitative case analysis, with a methodology adapted from Raji et al. (2022), Ayling and Chapman (2022), and Stein et al. (2024): 1. Identification of nine safety-critical industries through seed articles 2. Literature review of incident reporting institutions in each industry 3. Classification and coding of each industry's reporting system according to the seven-dimensional framework 4. Extraction of common patterns and best practices as design considerations for AI incident reporting

Key Experimental Results

Main Results

(This paper is a policy analysis study; "main results" correspond to the systematic comparative analysis of the case industries.)

Summary comparison of incident reporting systems across nine industries:

Industry Mandatory Reporting Voluntary Reporting Operating Entity Coverage
Nuclear Regulatory (NRC) + International (IAEA) Companies + employees
Aviation Regulatory (FAA) + Non-regulatory (NASA/ASRS) Companies + employees + public
Pesticides Multi-level (federal + state + local) Farmers + physicians + consumers
Pharmaceuticals Regulatory (FDA/MedWatch) Companies (90% mandatory) + consumers
Cybersecurity 22 federal agencies, 45+ requirements Companies + government
Dams Federal + state + industry Engineers + government
Railways FRA + NASA (C3RS) Companies + employees
Occupational Safety OSHA Employers
Healthcare State agencies + federal Hospitals + physicians + patients

Key quantitative findings: - 90% of reports in the FDA MedWatch database originate from mandatory corporate reporting - 91% of incidents are reported in the FAA wildlife strike database, but via 10+ distinct systems - A hospital study found that <1% of events caused serious injury and 82% caused no injury - C3RS (railway voluntary reporting) attracted only 23 of 800 railroad companies, reflecting a failure of industry buy-in - Kesari (2023) found that mandatory cybersecurity incident reporting reduced identity theft complaints by an average of 10.1%

Ablation Study

(In this policy analysis paper, this section corresponds to an analysis of design dimension variants.)

Comparison of mandatory vs. voluntary reporting effectiveness:

Reporting Type Advantages Disadvantages Typical Case
Mandatory High coverage, strong regulatory visibility May suppress reporting of low-severity events FDA: 90% of reports from mandatory requirements
Voluntary Covers near misses, promotes learning Difficult industry buy-in, competitive concerns ASRS: 2M+ reports vs. C3RS: only 23 participants
Anonymous Reduces fear of retaliation, increases reporting rates Impedes accountability, complicates follow-up investigation ASRS de-identification + FAA immunity protection
Public Facilitates accountability and follow-up investigation Suppresses reporting (72% of physicians indicate legal protection would increase reporting) OSHA mandatory public reporting

Key Findings

  • Safety learning and accountability goals are difficult to reconcile within the same system and often require separate systems
  • The success of ASRS (2M+ reports) depends on multiple factors: non-regulatory operation, de-identification, immunity protection, and early stakeholder engagement. This model failed when replicated in the railway sector (C3RS)
  • In cybersecurity, 22 federal agencies and 45+ reporting requirements create severe fragmentation, impeding data aggregation and learning — a pitfall AI governance should avoid
  • Ambiguous reporting thresholds lead companies to evade compliance (as seen in automotive manufacturer cases), making precise definitions critical
  • The multi-domain deployment and complex supply chains of GPAI mean that incident information is dispersed across multiple parties, necessitating multi-system, multi-level reporting

Highlights & Insights

  • The seven-dimensional framework provides a unified vocabulary for analyzing and comparing incident reporting systems, filling a gap in the AI governance literature
  • The incident lifecycle model clearly distinguishes hazards, situations, near misses, and harm events, laying a foundation for standardizing AI incident definitions
  • Design principles derived from the experience of nine mature industries help the AI field avoid reinventing the wheel
  • The critical analysis of the applicability of the ASRS model to the AI industry offers practical guidance

Limitations & Future Work

  • The analysis is U.S.-centric, and some lessons may not transfer readily to other jurisdictions
  • No cost-benefit analysis is conducted comparing incident reporting systems against alternative governance mechanisms
  • AI security vulnerability incidents (distinct from safety/rights incidents) are not examined in depth, and information sharing in this category carries additional risks
  • The framework focuses on institutional design and does not address technical implementation (incident monitoring/detection methods), safety culture, user interaction, or other sociotechnical factors
  • Empirical data from practitioner interviews, user experiments, and industry surveys are absent
  • vs. Raji et al. (2022): Raji et al. propose a citizen–regulator reporting system to promote audit accountability; this paper provides a more comprehensive analysis of institutional design dimensions, arguing that accountability and learning may require separate systems
  • vs. McGregor (2021) (AIID): AIID represents an important first step but lacks the stakeholder buy-in and information needed for safety learning — eight of the nine industries studied implement additional reporting mechanisms beyond standalone databases
  • vs. Shrishak (2023): Shrishak proposes a voluntary AI reporting system modeled on the FAA/ASRS, but this paper draws on the failure of C3RS to argue that the ASRS model's applicability to the AI industry is questionable, given competitive dynamics, the absence of unions, and the complexity of multi-domain deployment

Rating

  • Novelty: ⭐⭐⭐⭐ First systematic institutional design analysis framework for AI incident reporting
  • Experimental Thoroughness: ⭐⭐⭐ Qualitative case analysis is detailed, but lacks quantitative empirical data and practitioner evidence
  • Writing Quality: ⭐⭐⭐⭐ Clear structure, rigorous argumentation, and comprehensive literature review
  • Value: ⭐⭐⭐⭐ Directly useful to AI safety governance policymakers; highly timely