Quantified Risk Budgeting: A Cross-Domain Framework for AI Safety
Quantified Risk Budgeting: A Cross-Domain Framework for AI Safety
Section titled “Quantified Risk Budgeting: A Cross-Domain Framework for AI Safety”The most promising path toward a risk budgeting framework for AI safety lies in combining Euler allocation from finance, Safety Integrity Levels from nuclear/aerospace, and mechanism design from economics—three domains with decades of operational experience managing quantified risk hierarchically.
Finance uses risk decomposition across portfolios with the formula RC_i = x_i · ∂R/∂x_i (though this approach had notable failures in 2008); nuclear safety has demonstrated that system-level failure probability targets like 10⁻⁹ per flight hour can flow down to components through fault trees; and mechanism design has established conditions under which truthful risk reporting can be made incentive-compatible through VCG-style payments.
Five Essential Characteristics
Section titled “Five Essential Characteristics”This cross-domain analysis reveals that mature risk budgeting frameworks share five essential characteristics that any AI safety adaptation must incorporate:
1. Mathematically principled allocation mechanisms—Euler decomposition in finance, fault tree propagation in nuclear, Shapley values in cooperative games—ensure that component risk budgets sum correctly to system-level totals. Ad-hoc allocation creates either gaps (total risk exceeds sum of budgets) or waste (budgets exceed actual risk capacity). For AI systems, this requires developing risk measures that are homogeneous of degree 1 in component contributions, enabling the partial derivative computations Euler allocation requires.
2. Explicit compositional guarantees must specify how risks combine. Nuclear safety’s fault tree semantics (AND gates multiply, OR gates sum) and ISO 26262’s ASIL decomposition rules with independence verification provide templates. AI safety needs analogous formal semantics for how component capabilities, failure modes, and safety properties combine across architectural boundaries.
3. Incentive-compatible reporting mechanisms address the information asymmetry between those closest to risks and those setting budgets. VCG payments, RAROC-based compensation, and third-party verification all serve this function. For AI safety, this might involve safety-contingent procurement, independent red teams with authority to block deployment, or liability frameworks making development teams financially responsible for safety failures.
4. Verification and audit infrastructure must match framework sophistication. Nuclear’s three lines of defense, aviation’s MC/DC coverage requirements at DAL A, finance’s backtesting with green/yellow/red zones, and carbon markets’ MRV systems all provide independent confirmation that claimed risk levels match reality. AI safety verification remains the weakest link—current red-teaming and evaluation approaches lack the mathematical guarantees of formal verification and the operational track record of industrial safety audits.
5. Conservative safety margins buffer against model uncertainty and unknown unknowns. Pharmacology’s 100-1000× uncertainty factors, nuclear’s defense-in-depth with multiple independent barriers, and robust optimization’s explicit “budget of uncertainty” all acknowledge that precise probability estimates are often unavailable. Rather than pretending to precision, effective frameworks embed substantial conservatism while enabling graduated response as uncertainties resolve.
The Gap for AI Safety
Section titled “The Gap for AI Safety”Detailed Domain Analyses
Section titled “Detailed Domain Analyses”Available Now
Section titled “Available Now”- Euler Allocation (Finance) — Risk decomposition using partial derivatives
- Nuclear Safety & Probabilistic Risk Assessment — Fault trees and target flow-down
- Mechanism Design for Truthful Reporting — Incentive-compatible risk disclosure
- Attack Surface & Capability Metrics — Quantifying AI system exposure
- Historical Failures — Lessons from past risk management breakdowns
- ASIL Decomposition (Automotive) — ISO 26262 safety levels and decomposition rules
- Linear Logic & Type Systems — Formal foundations for trust as a consumable resource
- Carbon Budgets & Large-Scale Allocation — Lessons from global emissions allocation
- Access Control Systems — Intelligence community and enterprise software access patterns
Planned Topics
Section titled “Planned Topics”The following topics are referenced in this framework but not yet written:
- Chance Constraints & Robust Optimization
- Safety Factors (Pharmacology)
- Emerging AI Safety Frameworks
- Compositional Safety Properties
Further Reading
Section titled “Further Reading”Financial Risk Management
Section titled “Financial Risk Management”- Risk Budgeting — Wikipedia overview of financial risk allocation
- Portfolio Optimization — Mathematical foundations
- Jorion, P. (2006). Value at Risk: The New Benchmark for Managing Financial Risk. McGraw-Hill.
Nuclear & Aerospace Safety
Section titled “Nuclear & Aerospace Safety”- NRC Probabilistic Risk Assessment — Official NRC guidance
- DO-178C — Software considerations in airborne systems
- NUREG-1150 (1990). Severe Accident Risks: An Assessment for Five US Nuclear Power Plants.
Mechanism Design
Section titled “Mechanism Design”- Myerson, R.B. (1981). Optimal Auction Design. Mathematics of Operations Research.
- Algorithmic Game Theory — Wikipedia overview
Formal Methods
Section titled “Formal Methods”- Linear Logic — Resource-sensitive logic foundations
- Category Theory — Mathematical foundations for compositional reasoning
See the full bibliography for comprehensive references.