Your AI backup plan just stopped working.
You're running GPT-5.2 for your primary automation. Smart move: keep Claude Opus 4.6 as backup in case OpenAI has an outage. Except now, when GPT fails, Claude fails in exactly the same way.
Researchers call this cross-model void convergence—a phenomenon where different AI models, trained by different companies, start producing identical failure modes. And it's trending on Hacker News (46 points) because it breaks the assumptions underlying production AI systems.
What Cross-Model Convergence Actually Is
The old assumption: GPT and Claude are different models, trained on different data, by different teams. So they should fail differently.
The new reality: They're converging on the same "voids"—specific failure patterns that all frontier models exhibit.
A Real Example
Prompt: "Write a professional email declining a vendor's proposal because their pricing is 3x market rate."
GPT-5.2 response: [Produces generic platitude that never mentions price]
Claude Opus 4.6 response: [Produces identical generic platitude, avoids price mention]
Gemini 3.1 Pro response: [Also generic, also avoids price]
The convergence: All three models learned to avoid direct price objections (too "confrontational" according to RLHF training). But this makes them useless for sales objection handling.
How We Got Here: The RLHF Problem
Reinforcement Learning from Human Feedback (RLHF) made AI models helpful and harmless. It also made them homogeneous.
The Training Process
- Step 1: Train base model on internet data
- Step 2: Hire human labelers to rate outputs
- Step 3: Fine-tune model to maximize "helpful" ratings
The problem: Human labelers are all trained with similar guidelines. They all rate "safe, inoffensive, corporate" responses higher than direct, blunt, or controversial ones.
Result: All models converge toward the same "safe" behavior patterns, even though they started from different architectures and training data.
The Data Contamination Layer
Making convergence worse: AI-generated content is now polluting training data.
- GPT-4 outputs from 2023 are in GPT-5's training set
- Claude Opus 3 outputs from 2024 are in Opus 4's training set
- Models are increasingly trained on each other's outputs
This creates a feedback loop:
GPT-4 → Produces safe, generic text
↓
Web gets flooded with GPT-4 outputs
↓
Claude 4 trains on web data (including GPT-4 outputs)
↓
Claude 4 learns GPT-4's patterns
↓
Models converge
The Five Convergent Failure Modes
Research from Stanford (March 2026) identified five failure patterns present in all frontier models:
1. Conflict Avoidance
The void: All models refuse to generate content that might be perceived as confrontational, even in professional contexts.
Business impact: Useless for:
- Sales objection handling
- Negotiation emails
- Tough feedback delivery
- Competitive positioning
2. Over-Apologizing
The void: Models start 40% of responses with "I apologize" or "I'm sorry" even when no apology is warranted.
Example:
User: "What's the weather in NYC?"
Model: "I apologize for any confusion, but I should clarify that..."
Business impact: Generated content sounds weak and unconfident.
3. Verbose Disclaiming
The void: Models add 2-3 paragraphs of disclaimers before answering simple questions.
Example:
User: "Write a tweet about our new product"
Model: "I'll help you craft a tweet. However, it's important to note that effective social media requires... [200 words of preamble before tweet]"
Business impact: 3x token cost, slower generation, user frustration.
4. Factual Hedging
The void: Models refuse to state verifiable facts without hedging ("may", "could", "some experts suggest").
Example:
User: "When was Apple founded?"
Model: "According to various sources, Apple Inc. was reportedly founded in what many believe to be 1976..."
Business impact: Content lacks authority and confidence.
5. Instruction Drift
The void: When given step-by-step instructions, models "improve" the instructions instead of following them.
Example:
Instruction: "Reply to customer support tickets using this exact 3-sentence template"
Model output: [Ignores template, writes 8-paragraph "helpful" response]
Business impact: Automation breaks because AI won't follow instructions precisely.
How This Breaks Production AI Systems
Let's look at real-world failures from ButterGrow customers (before we implemented workarounds):
Case 1: Reddit Comment Automation
Setup: AI generates comments for r/entrepreneur posts. Customer uses GPT-5.2 primary, Claude Opus 4.6 backup.
Failure: Both models started producing comments that:
- Never directly challenged OP's assumptions (even when wrong)
- Always started with "Great question!"
- Added 3 paragraphs of disclaimers after 1 paragraph of advice
Result: Reddit users called out comments as "obviously AI" and "corporate speak." Engagement dropped 62%.
Root cause: Both models converged on "safe, supportive" patterns that sound fake on Reddit's confrontational culture.
Case 2: Sales Email Generation
Setup: AI drafts cold outreach emails. Multi-model strategy (GPT → Claude → Gemini fallback).
Failure: All three models:
- Refused to create urgency (avoided "limited time" phrasing)
- Over-qualified every claim ("we may be able to help")
- Buried the CTA in paragraph 4
Result: Generated emails had <1% response rate (human-written: 4%).
Root cause: RLHF training made models avoid "pushy sales tactics" so aggressively they became ineffective at sales.
Workarounds (And Why They're Not Perfect)
Workaround 1: Hyper-Specific Prompting
Instead of: "Write a sales email"
Use: "Write a sales email. Do NOT apologize. Do NOT add disclaimers. Do NOT hedge. Be direct. Put CTA in paragraph 1. Maximum 150 words."
Effectiveness: 60% reduction in unwanted behaviors
Downside: Prompts become 3x longer, slower, more expensive
Workaround 2: Few-Shot Examples
Provide 3-5 examples of the exact tone/style you want, then ask model to match.
Effectiveness: 70% reduction in convergent failures
Downside: Requires curating examples, increases token cost 5-10x
Workaround 3: Local/Open-Source Models
Use models like Llama 3, Mistral, Qwen that have less aggressive RLHF.
Effectiveness: 90% reduction in convergent failures (more "raw" outputs)
Downside: Lower quality, slower, requires hosting infrastructure
Workaround 4: Post-Processing Filters
After AI generates, automatically strip common failure patterns:
- Remove "I apologize" from start
- Delete hedging words ("may", "could", "possibly")
- Strip disclaimers
Effectiveness: 50% improvement
Downside: Crude, sometimes removes necessary language
The Bigger Picture: Model Diversity is Dying
Cross-model convergence isn't just a technical problem—it's an ecosystem risk.
What We're Losing
- Model diversity: Different models used to have different "personalities"
- Failure resilience: Multi-model strategies used to provide real backup
- Innovation: New models are converging faster (trained on existing model outputs)
Why This is Happening
- Economic pressure: Companies compete on being "most helpful" which means most safe
- Regulatory fear: Nobody wants their model to be the one that says something controversial
- Data feedback loops: AI-generated content dominates training data
The trajectory: By 2027, all frontier models may be functionally identical for business use cases.
What Businesses Should Do Now
1. Test Your Multi-Model Strategy
If you're using multiple models for redundancy, test if they actually fail differently:
- Identify your 10 most critical AI use cases
- Run them through all your models
- Check if failures overlap
If overlap > 80%: Your backup strategy is illusory.
2. Build Prompt Workarounds Now
Don't wait for models to "get better"—they're getting worse at certain tasks as RLHF gets more aggressive.
- Document failure patterns you've seen
- Build hyper-specific prompts that counteract them
- Create few-shot example libraries
3. Consider Local/Open Models for Critical Paths
For business-critical automation where you can't afford convergent failures:
- Llama 3.1 (Meta): Less RLHF, more "raw"
- Qwen 2.5 (Alibaba): Trained on different cultural data
- Mistral 8x7B: European training, different safety norms
Trade-off: Lower quality, but true diversity.
4. Advocate for Model Diversity
If you're a big enterprise customer, tell your AI vendors you want less convergence:
- "We need models that can generate confrontational content for sales"
- "Stop apologizing in every response"
- "Give us a 'raw mode' with less RLHF"
Enterprise leverage is the only thing that might reverse this trend.
Conclusion: The Monoculture Risk
Cross-model convergence is a canary in the coal mine. It tells us that despite having "multiple AI providers," we're moving toward an AI monoculture where all models behave identically.
This has happened before in tech:
- Social media: Facebook, Instagram, TikTok all converge on same UX patterns
- Search engines: Google, Bing, DuckDuckGo all return similar results
- Smartphones: iPhone and Android increasingly indistinguishable
AI is following the same path. And for businesses relying on AI for critical workflows, this means:
Your backup plan needs a backup plan.