Estimates Registry
Estimates Registry
Section titled “Estimates Registry”This section provides a centralized registry of all quantitative estimates used in the Delegation Risk Framework. Each estimate includes:
- Point estimate: A single representative value
- Distribution: Uncertainty quantification in Squiggle notation
- Confidence level: How certain we are about this estimate
- Source: Where the estimate comes from
- Last updated: When the estimate was last reviewed
Why Formalize Estimates?
Section titled “Why Formalize Estimates?”Most risk frameworks use point estimates like “probability = 2%” or “damage = $5,000”. This understates uncertainty and can lead to:
- Overconfidence in risk calculations
- Poor decisions when parameters are near critical thresholds
- Inability to identify which uncertainties matter most
- No path to improvement through better calibration
By expressing estimates as probability distributions, we can:
- Calculate confidence intervals on total risk
- Perform sensitivity analysis to find critical parameters
- Update estimates as we gather more data
- Compare architectures under parameter uncertainty
Estimate Categories
Section titled “Estimate Categories”Default failure probabilities by component type (deterministic code, narrow ML, general LLM, RL agents). These are the priors before observing any track record.
Damage magnitude distributions by category (infrastructure, data, reputation, regulatory, catastrophic). Heavy-tailed distributions for rare-but-severe events.
How much risk reduction do various interventions provide? Verifiers, sandboxing, rate limiting, human review, formal verification.
How to update trust based on track record. Bayesian updating formulas and calibrated priors.
Reference points from nuclear safety, aviation, finance, and other domains with mature risk quantification.
How to gather calibrated probability and damage estimates from domain experts when historical data is unavailable.
How to ground your risk estimates in real-world AI incident data from public databases.
Using These Estimates
Section titled “Using These Estimates”For Practitioners
Section titled “For Practitioners”Use these as starting points for your own risk assessment. Adjust based on:
- Your specific domain (healthcare vs. internal tools)
- Your organization’s risk appetite
- Your observed track record
- Expert judgment from your team
For Researchers
Section titled “For Researchers”These estimates are hypotheses to be tested. Help improve them by:
- Analyzing AI incident databases
- Conducting expert elicitation studies
- Publishing calibration results
- Proposing better distribution families
Notation Guide
Section titled “Notation Guide”We use Squiggle notation for distributions:
| Notation | Meaning | Example |
|---|---|---|
beta(a, b) | Beta distribution with shape parameters a, b | beta(10, 100) → ~9% mean |
lognormal(mu, sigma) | Lognormal with log-mean μ, log-std σ | lognormal(log(1000), 0.5) |
normal(mean, std) | Normal/Gaussian distribution | normal(0.1, 0.02) |
uniform(low, high) | Uniform between bounds | uniform(0.01, 0.1) |
mixture([d1, d2], [w1, w2]) | Weighted mixture of distributions | mixture([normal(0,1), normal(5,1)], [0.8, 0.2]) |
Interpreting Confidence Levels
Section titled “Interpreting Confidence Levels”| Level | Meaning |
|---|---|
| High | Based on substantial data or expert consensus; unlikely to change significantly |
| Medium | Reasonable estimate with some uncertainty; may change with new data |
| Low | Educated guess or extrapolation; significant revision possible |
| Speculative | Placeholder for discussion; not yet suitable for decisions |
Contributing
Section titled “Contributing”We welcome contributions to improve these estimates:
- Data-driven updates: If you have incident data or calibration studies
- Domain expertise: If you can provide better priors for specific domains
- Methodological improvements: Better distribution families, aggregation methods
- Error corrections: If you find mistakes or inconsistencies
See our contribution guidelines for how to submit improvements.