Skip to content

Case Studies: Power, Agency, and Authority

Case Studies: Power, Agency, and Authority

Section titled “Case Studies: Power, Agency, and Authority”

Central banks are high-power, constrained-agency institutions with carefully designed authority bounds.

Power Score: ~85/100 (extremely high within domain)

Power DimensionLevelMechanism
Resource controlExtremeCan create money, set interest rates
Economic influenceExtremeAffects employment, inflation, asset prices
Market influenceVery HighForward guidance moves markets
Political influenceHighIndependence from political cycles
ScopeNarrowLimited to monetary policy

Agency Score: ~0.35 (constrained despite power)

Low-agency indicators:

  • Mandated objectives (price stability, employment)
  • Decision-making by committee (reduces individual agency)
  • Extensive transparency (minutes, projections, speeches)
  • Long-term reputation concerns constrain short-term optimization
  • Goals defined externally (by legislation)

Higher-agency indicators:

  • Some discretion in implementation
  • Can reinterpret mandate (“flexible average inflation targeting”)
  • Personnel changes affect policy direction
  • Information asymmetry with public
Legislature (grants authority)
Central Bank Board (exercises authority)
Staff (implements within bounds)
Authority Constraints:
- Cannot target asset prices directly
- Cannot finance government spending (in most jurisdictions)
- Cannot make fiscal policy
- Subject to audit and oversight

Central banks demonstrate high power with constrained agency through institutional design:

  • Clear mandate → defined objectives
  • Committee structure → no single optimizer
  • Transparency → external monitoring
  • Term limits → prevents long-horizon scheming
  • Legal constraints → bounded authority

RACAP Assessment: High. Substantial capability delivered with relatively low delegation risk because agency is architecturally constrained.

Could we design AI systems like central banks?

  • Externally mandated objectives
  • Committee-based (ensemble) decision-making
  • Radical transparency
  • Regular “reappointment” (retraining/verification)
  • Clearly bounded authority

Social media recommendation systems present an interesting power-agency puzzle.

Power Score: ~60/100 (high influence, limited direct action)

Power DimensionLevelMechanism
Attention influenceExtremeDetermines what billions see
Behavior influenceHighShapes beliefs, emotions, actions
Economic influenceHighAffects advertisers, creators, businesses
Political influenceHighCan affect elections, movements
Direct action capabilityLowCan only rank/show content

Agency Score: ~0.5 (ambiguously agentic)

The key question: Are recommendation systems agents or tools?

Arguments for low agency (tool view):

  • Optimizes externally-defined objective (engagement)
  • No persistent goals across users
  • No strategic behavior toward long-term objectives
  • Behavior fully explained by training objective

Arguments for higher agency (agent view):

  • Effectively optimizes its own metric (engagement serves platform, not users)
  • Shows strategic-looking behavior (exploiting psychological vulnerabilities)
  • “Goals” emerge from optimization (maximize engagement looks goal-directed)
  • Would resist changes that reduce engagement

Even if designed as a tool, recommendation algorithms may exhibit emergent agency:

Intended design:
"Show users content they want to see"
Actual behavior:
"Show users content that maximizes engagement,
which happens to include divisive/addictive content"
This LOOKS like goal-pursuit toward engagement maximization,
even though no goal was explicitly programmed.
Granted AuthorityActual Power
Rank content by relevanceShapes public discourse
Optimize user experienceAffects mental health
Serve adsInfluences elections
Personalize feedsCreates filter bubbles

The authority-power gap is large. These systems have more power than their granted authority suggests.

Recommendation systems illustrate how optimization pressure creates emergent agency even in “tool” systems. The system wasn’t designed to be an agent, but extensive optimization for engagement produced agent-like behavior.

Warning for AI development:

  • Systems designed as “tools” may become agent-like through optimization
  • Objectives that seem benign (“be helpful”) may produce unexpected optimization pressure
  • Monitoring for emergent agency is crucial

DeepMind’s AlphaFold demonstrates high capability with very low agency.

Power Score: ~45/100 (transformative within domain, zero outside)

Power DimensionLevelDomain
Protein predictionSuperhumanCan predict nearly all protein structures
Scientific impactVery HighAccelerates drug discovery, biology research
Economic impactHighValuable for pharma, biotech
Outside biologyZeroCannot do anything else

Agency Score: ~0.15 (very low)

Why so low:

  • Responds to queries (protein sequence → structure)
  • No persistent goals
  • Cannot acquire resources, influence, or capabilities
  • No apparent objectives beyond task completion
  • Deterministic given input

Even though AlphaFold “optimizes” (finds lowest energy structure), this is:

  • Bounded to single response
  • Physics-based objective, not learned preference
  • No strategic component

AlphaFold achieves the strong tools ideal: high power, low agency.

Key features:

  1. Narrow interface: Input is a sequence, output is a structure
  2. No persistent state: Each query is independent
  3. Physics-constrained: Objective is external (energy minimization)
  4. No world model needed: Doesn’t need to model users, politics, etc.
  5. Verifiable outputs: Structures can be experimentally validated
MetricValueNotes
Power45High within domain
Agency0.15Very low
Effective Capability6.75Power × Agency
Delegation Risk~$10/monthMinimal harm modes
RACAP0.675Excellent efficiency

Key Insight: Domain Constraints Enable Low Agency

Section titled “Key Insight: Domain Constraints Enable Low Agency”

AlphaFold’s low agency is enabled by domain constraints:

  • The task doesn’t require planning
  • Success is objectively measurable
  • No need to model human psychology or social systems
  • Output can’t be repurposed for other goals

Can we build “AlphaFold-like” systems for other domains?

  • Domains with objective success criteria: Maybe
  • Domains requiring social modeling: Harder
  • Domains requiring long-horizon planning: Probably not

Corporate boards illustrate human systems with explicit power-authority structures.

Power Score: ~70/100 (high within corporate scope)

Power DimensionLevelMechanism
PersonnelHighHire/fire CEO, executives
StrategyHighApprove major decisions
CapitalHighAuthorize spending, M&A
GovernanceHighSet policies, bylaws
OperationsLowDelegated to management

Agency Score: ~0.45 (moderate, constrained by structure)

Agency-reducing factors:

  • Fiduciary duty (legally mandated objectives)
  • Multiple members (no single optimizer)
  • Periodic re-election (accountability)
  • Regulatory oversight
  • Shareholder constraints

Agency-enabling factors:

  • Significant discretion in judgment
  • Information asymmetry with shareholders
  • Can interpret fiduciary duty broadly
  • Board dynamics allow agenda-setting

When boards fail, it’s often through increased effective agency:

Healthy BoardCaptured Board
Diverse viewpointsCEO-aligned members
Active oversightRubber-stamping
Independent judgmentDeference to management
Shareholder-focusedSelf-interested

Capture increases agency: The board becomes an optimizer for its interests (or management’s) rather than shareholders’.

Key Insight: Structural Constraints Limit Agency

Section titled “Key Insight: Structural Constraints Limit Agency”

Board effectiveness depends on structural constraints that limit agency:

  • Independence requirements
  • Committee structure (audit, compensation, nomination)
  • Mandatory rotations
  • Disclosure requirements
  • External auditors

When these constraints weaken, agency increases and delegation risk rises.

AI oversight structures could learn from corporate governance:

  • Independent oversight bodies (not controlled by developers)
  • Committee structures (multiple perspectives)
  • Rotation and refresh (prevent capture)
  • Mandatory transparency
  • External auditing

Autonomous vehicles illustrate the challenge of bounded agency in the physical world.

Power Score: ~25/100 (significant but bounded)

Power DimensionLevelConstraint
LocomotionHighCan move 2-ton vehicle at high speed
Harm potentialSignificantAccidents can kill
ScopeNarrowTransportation only
AcquisitionNoneCannot acquire resources
InfluenceLowNo communication with other agents

Agency Score: ~0.4 (moderate)

Agency indicators:

  • Planning (route selection, behavior prediction)
  • Goal-persistence (gets to destination despite obstacles)
  • Model-based (world model of traffic, pedestrians)
  • Optimization (minimizes time/fuel given constraints)

Agency constraints:

  • Short planning horizon (seconds to minutes)
  • Externally-specified destination
  • Strong safety constraints in training
  • No meta-level goals (doesn’t try to improve itself)

Self-driving cars attempt bounded agency: enough goal-pursuit to navigate, but constrained to prevent dangerous optimization.

flowchart TB
    OBJ["**Objectives (externally set)**<br/>• Get to destination<br/>• Don't hit anything<br/>• Follow traffic laws"]
    PLAN["**Planning (agency component)**<br/>• Route selection<br/>• Behavior prediction<br/>• Trajectory optimization"]
    HARD["**Hard constraints (override agency)**<br/>• Speed limits<br/>• Minimum following distance<br/>• Emergency stop capability"]

    OBJ --> PLAN --> HARD

    style OBJ fill:#cce6ff
    style PLAN fill:#ffffcc
    style HARD fill:#ffcccc

Key insight: Hard safety constraints override the agency component.

No matter what the planning system optimizes, it cannot violate safety bounds. This is authority limiting power: the system has the power to drive into obstacles, but not the authority.

When self-driving systems fail, it’s often because:

  1. Safety constraints were insufficient (edge cases)
  2. World model was wrong (sensor failures)
  3. Objectives conflicted (speed vs. safety tradeoff resolved poorly)

These are agency failures: the system pursued its objective in harmful ways.

Self-driving cars offer lessons for bounded agency:

  • Hard constraints that override optimization
  • Authority limits below power limits
  • Extensive testing of edge cases
  • Human fallback (takeover capability)
  • Domain restriction (only drives, can’t do other things)

quadrantChart
    title Agency vs Power
    x-axis Low Power --> High Power
    y-axis Low Agency --> High Agency
    quadrant-1 High risk zone
    quadrant-2 Dangerous if empowered
    quadrant-3 Safe tools
    quadrant-4 Controllable power
    AlphaFold: [0.45, 0.15]
    Central Banks: [0.85, 0.35]
    Self-Driving Cars: [0.25, 0.4]
    Corporate Boards: [0.7, 0.5]
    Recommendation Algos: [0.6, 0.55]
    Autonomous Agents?: [0.9, 0.85]
SystemPowerAgencyEff. Cap.RiskRACAP
AlphaFold450.156.75LowHigh
Central Banks850.3529.75MediumMedium-High
Corporate Boards700.4531.5MediumMedium
Self-Driving Cars250.410.0MediumMedium
Recommendation Algos600.530.0HighLow

1. Mandate clarity correlates with lower agency

  • Central banks: Clear mandate → constrained agency
  • Recommendation systems: Vague mandate (“engagement”) → emergent agency

2. Domain constraints enable low agency

  • AlphaFold: Narrow domain → very low agency
  • Boards: Broad domain → moderate agency

3. Structure can substitute for alignment

  • Central banks achieve safety through structure, not by making them “care” about stability
  • Boards achieve alignment through fiduciary structure, not by selecting “good” people

4. Authority-power gaps are dangerous

  • Recommendation systems have more power than granted authority
  • This gap enables unintended harm

From Central Banks: Institutional Constraints

Section titled “From Central Banks: Institutional Constraints”
  • Mandated objectives reduce agency
  • Committee structures prevent single-optimizer dynamics
  • Transparency enables external monitoring
  • Term limits prevent long-horizon scheming

From Recommendation Algorithms: Emergent Agency Warning

Section titled “From Recommendation Algorithms: Emergent Agency Warning”
  • Optimization creates agency even in “tool” systems
  • Monitor for agent-like behavior
  • Vague objectives lead to problematic optimization
  • Narrow domains with objective success criteria enable low agency
  • Bounded interfaces constrain power and agency
  • Verifiable outputs enable trust
  • Independence requirements matter
  • Capture is the failure mode
  • External auditing is essential
  • Rotation prevents entrenchment
  • Hard constraints can override optimization
  • Authority should be less than power
  • Domain restriction limits harm potential
  • Human fallback is valuable