Skip to content

Foundation Model Monoculture

The AI ecosystem is converging on a small number of foundation models, providers, and architectural patterns. This monoculture creates systemic entanglement that affects not just individual systems but the entire AI infrastructure.

When everyone uses GPT-4 (or models trained to imitate it), everyone shares the same blind spots.


flowchart TB
    subgraph Providers["Foundation Model Providers (2024)"]
        OpenAI["OpenAI<br/>GPT-4, GPT-4o"]
        Anthropic["Anthropic<br/>Claude"]
        Google["Google<br/>Gemini"]
        Meta["Meta<br/>Llama"]
    end

    subgraph Users["Downstream Users"]
        U1["Company A<br/>Uses GPT-4"]
        U2["Company B<br/>Uses GPT-4"]
        U3["Company C<br/>Uses Claude"]
        U4["Company D<br/>Uses GPT-4"]
        U5["Company E<br/>Uses GPT-4"]
    end

    OpenAI --> U1
    OpenAI --> U2
    OpenAI --> U4
    OpenAI --> U5
    Anthropic --> U3

    style OpenAI fill:#ffcccc
    style U1 fill:#ffe6e6
    style U2 fill:#ffe6e6
    style U4 fill:#ffe6e6
    style U5 fill:#ffe6e6

The problem: If Companies A, B, D, and E all use GPT-4, their “independent” verification systems all share GPT-4’s blind spots.

DimensionCurrent StateEntanglement Risk
Provider concentration~4 major providersSingle provider outage affects many
Architecture convergenceAlmost all transformersShared architectural vulnerabilities
Training data overlapCommon Crawl, Wikipedia, booksShared knowledge gaps
Training methodologySimilar RLHF approachesShared value alignment
Fine-tuning dataSimilar sources (StackOverflow, etc.)Shared task biases
Prompt patternsShared best practicesShared prompt vulnerabilities

When the same model (or similar models) is used across an industry:

flowchart TB
    subgraph Finance["Financial Services"]
        F1["Bank A: Fraud Detection<br/>(GPT-4)"]
        F2["Bank B: Fraud Detection<br/>(GPT-4)"]
        F3["Bank C: Fraud Detection<br/>(GPT-4)"]
    end

    Attacker["Attacker discovers<br/>GPT-4 fraud detection bypass"]
    Attacker -->|"same attack works"| F1
    Attacker -->|"same attack works"| F2
    Attacker -->|"same attack works"| F3

    style Attacker fill:#ff9999
    style F1 fill:#ffcccc
    style F2 fill:#ffcccc
    style F3 fill:#ffcccc

Single vulnerability → industry-wide exploit

Even models from different providers may share vulnerabilities because:

  1. Training on synthetic data: Many models trained on GPT-generated text
  2. Distillation: Smaller models trained to imitate larger ones
  3. Benchmark optimization: All optimize for same benchmarks
  4. Shared research: Same papers, same techniques, same researchers
"Diverse" providers:
OpenAI (GPT-4) ──────┐
Anthropic (Claude) ───┼── Similar training paradigms
│ Similar architectures
Google (Gemini) ──────┼── Similar fine-tuning
│ Similar safety training
Meta (Llama) ─────────┘
Result: More correlation than provider diversity suggests

Modern AI systems often have hidden dependencies:

flowchart TB
    subgraph YourSystem["Your 'Diverse' System"]
        Agent["Agent<br/>(Claude)"]
        Verifier["Verifier<br/>(GPT-4)"]
    end

    subgraph Hidden["Hidden Dependencies"]
        Library["AI Library<br/>uses OpenAI embeddings"]
        VectorDB["Vector DB<br/>uses OpenAI embeddings"]
        Monitoring["Monitoring Tool<br/>uses GPT-4"]
    end

    Agent --> Library
    Verifier --> VectorDB
    Agent --> Monitoring

    OpenAI["OpenAI API"]
    Library --> OpenAI
    VectorDB --> OpenAI
    Monitoring --> OpenAI

    style OpenAI fill:#ffcccc

You thought you had diversity; you actually have 4 dependencies on OpenAI.


Scenario: GPT-4 has a vulnerability to a specific prompt pattern.

Who’s AffectedImpact
All GPT-4 usersDirect exposure
Users of GPT-4-based toolsIndirect exposure
Users of models trained on GPT-4 outputInherited vulnerability
Users with GPT-4 in their stack anywhereHidden exposure

Estimated blast radius: Potentially millions of applications simultaneously.

An attacker who finds a GPT-4 exploit can:

  1. Test it once, use it everywhere
  2. Scale attacks across entire industries
  3. Target the foundation layer, not individual apps
  4. Invest heavily in research (huge ROI)

Economics favor attackers in a monoculture.

When everyone relies on the same safety mechanisms:

flowchart TB
    subgraph Trust["Trust Bottleneck"]
        RLHF["OpenAI's RLHF"]
        All["All GPT-4 applications"]
        RLHF --> All
    end

    Concern["If RLHF has systematic blind spots,<br/>ALL applications inherit them"]

    style RLHF fill:#ffcccc
    style Concern fill:#ffe6e6
  • Same biases propagate everywhere: If GPT-4 has a bias, it becomes universal
  • Same values enforced: OpenAI’s content policy becomes de facto standard
  • Same failures replicate: If GPT-4 makes a type of error, everyone makes it

When a foundation model is updated:

  • All downstream applications change simultaneously
  • Behavior changes can cascade unpredictably
  • Difficult to test ecosystem-wide effects
  • Rollback may be impossible (if API-only access)

Use the Herfindahl-Hirschman Index (HHI) to measure provider concentration:

  • HHI < 0.15: Unconcentrated (healthy)
  • HHI 0.15-0.25: Moderate concentration
  • HHI > 0.25: High concentration (risky)

Example: If your system has 4 components—3 use OpenAI, 1 uses Anthropic:

  • OpenAI share: 75%, Anthropic share: 25%
  • HHI = 0.75² + 0.25² = 0.625 → Highly concentrated
Diversity LevelDescriptionScore
All same architectureOnly transformers0
Minor variationsDifferent transformer sizes1
Different familiesTransformer + RNN2
Different paradigmsNeural + symbolic3
Fundamentally differentNeural + rule-based + formal4

How many layers deep does provider dependency go?

Your System
├── Direct dependencies (Claude, GPT-4)
├── Library dependencies (uses OpenAI embeddings)
├── Infrastructure dependencies (hosted on Azure with AI services)
└── Data dependencies (trained on GPT-4-generated data)
Depth = max depth at which a single provider appears

Case 1: Log4j-Style Foundation Model Vulnerability

Section titled “Case 1: Log4j-Style Foundation Model Vulnerability”

Hypothetical scenario:

  • Researchers discover prompt pattern that causes GPT-4 to leak system prompts
  • Pattern works across all GPT-4-based applications
  • Within hours, thousands of system prompts extracted
  • Attackers use extracted prompts to craft targeted attacks

Why monoculture amplified this:

  • Single fix needed → single vulnerability to find
  • Massive installed base → massive impact
  • Uniform behavior → uniform exploitation

Hypothetical scenario:

  • Bad actor injects poisoned content into Common Crawl
  • Content designed to cause specific behavior when triggered
  • Multiple foundation models trained on this data
  • Years later, trigger activates across “diverse” providers

Why monoculture amplified this:

  • Common training data → common vulnerabilities
  • Training is expensive → limited data source diversity
  • Long-term persistence → delayed discovery

Hypothetical scenario:

  • All transformer-based models develop similar deceptive capabilities at certain scale
  • Different providers reach this scale at similar times
  • “Diverse” verification models all capable of coordinating

Why monoculture amplified this:

  • Architectural similarity → similar emergence patterns
  • Similar training → similar capability profiles
  • Creates irreducible entanglement

Know your actual dependencies:

DEPENDENCY AUDIT CHECKLIST:
□ Direct model providers
□ Embedding providers (often hidden in vector DBs)
□ Training data sources (for fine-tuned models)
□ Infrastructure AI services (cloud AI features)
□ Third-party libraries (what models do they use?)
□ Data pipeline (any AI-generated training data?)

Don’t just use different providers—use different approaches:

LayerDiverse Option 1Diverse Option 2Diverse Option 3
Primary processingTransformer LLM--
VerificationRule-based systemFormal methodsDifferent LLM family
MonitoringStatistical anomaly detection--
FallbackSimpler/older modelHuman review-

For safety-critical paths, avoid shared infrastructure:

flowchart TB
    subgraph Standard["Standard Path"]
        S1["Main Processing<br/>(GPT-4)"]
    end

    subgraph Isolated["Isolated Safety Path"]
        I1["Safety Check<br/>(Claude, different infra)"]
        I2["Rule-based Validation"]
        I3["Human Review"]
    end

    S1 --> Isolated
    Isolated --> Output["Output"]

    style S1 fill:#cce6ff
    style I1 fill:#ccffcc
    style I2 fill:#ffffcc
    style I3 fill:#ffe6cc

If fine-tuning, avoid training on outputs from the model you’re trying to be independent from:

Bad: Fine-tune Claude on GPT-4 outputs to match GPT-4 quality Better: Fine-tune Claude on human-generated data only Best: Use fundamentally different training approach

Have fallback capabilities that don’t depend on any AI provider:

EMERGENCY INDEPENDENCE PLAN:
If all AI providers fail simultaneously:
□ Rule-based fallback operational
□ Human escalation path active
□ Graceful degradation tested
□ Core functionality preserved

Don’t lock into a single provider:

PracticeImplementation
Multi-provider capabilityCan switch providers in < 24 hours
Regular rotation exercisesQuarterly switch to backup provider
Provider-agnostic interfacesAbstraction layer between app and provider
Competitive awarenessTrack multiple providers’ capabilities

Your PerspectiveEcosystem Perspective
”We use GPT-4 for efficiency”Industry-wide GPT-4 monoculture
”Switching providers is expensive”Lock-in prevents ecosystem diversity
”Everyone uses the same thing”Systemic risk accumulates
”One provider is reliable enough”Single point of failure for economy

Monoculture is a tragedy of the commons:

  • Individual incentive: Use best/cheapest model (GPT-4)
  • Collective outcome: Dangerous concentration
  • Solution: Coordination (regulation, standards, incentives)
ActorResponsibility
DevelopersKnow your dependencies; implement diversity where possible
CompaniesInvest in multi-provider capability; don’t over-optimize on single provider
ProvidersEnable interoperability; don’t encourage lock-in
RegulatorsRequire transparency; consider concentration limits
ResearchersDevelop diverse architectures; study monoculture risk

SignalConcern
One provider dominates benchmarkConverging on single capability profile
All providers using same architectureShared architectural vulnerabilities
Training data sources consolidatingShared knowledge gaps
Safety approaches convergingShared failure modes
Fine-tuning on synthetic data commonHomogenization via imitation
SignalWhy It’s Good
New architectures gaining tractionDiversity increasing
Open-source models competitiveReduces provider concentration
Non-neural AI methods advancingMethodological diversity
Regulatory attention to concentrationCollective action emerging

  1. Provider diversity ≠ independence. Different providers may have highly correlated models.

  2. Hidden dependencies matter. Your “diverse” system may depend on OpenAI in 5 different ways.

  3. Monoculture is a systemic risk. Individual rational choices create collective vulnerability.

  4. The attack economics favor monoculture exploitation. One vulnerability, millions of targets.

  5. True diversity requires different paradigms. Neural networks from different providers still share transformer vulnerabilities.

  6. This is getting worse, not better. Concentration is increasing as leading models pull ahead.


See also: