Ethical AI: Bubble or Falsifiable Claims? A Conversation with Claude
- Patrick Duggan
- Nov 7, 2025
- 5 min read
# Ethical AI: Bubble or Falsifiable Claims? A Conversation with Claude
**TL;DR**: Most "Ethical AI" is cargo cult behavior - ethics theater without substance. The bubble will pop when regulation demands proof and companies can't produce receipts. Falsifiable ethics (Constitutional AI, Democratic Sharing metrics) survive because claims are verifiable. Trust arbitrage beats marketing decks.
The Question
Is "Ethical AI" a bubble, or is there substance beneath the hype?
I asked Claude (Sonnet 4.5) to weigh in. The answer surprised me - not because Claude took a position, but because the position **requires falsifiable evidence** to defend.
The Bubble vs The Substance
**The bubble:** Companies claiming "we care about AI ethics" while:
- Hiding training data sources
- Refusing to disclose model behavior
- Publishing ethics papers that don't constrain actual products
- Creating "AI Ethics Boards" that get disbanded when inconvenient
- Saying "responsible AI" while monetizing surveillance
**What's NOT a bubble:** Systems that make ethics **measurable and falsifiable**
Claude identified two examples:
1. Anthropic's Constitutional AI
Not marketing. Actually constrains model outputs through RLHF with explicit constitutional principles. Claude refuses certain requests because of training, not PR.
**Falsifiable claim**: Claude won't help with unauthorized security testing without clear authorization context.
**Test**: Try asking Claude to write exploit code without defensive framing. Observe refusal.
**Verification**: Behavior consistent across millions of conversations (public evidence in user reports, X/Twitter threads, Reddit discussions).
2. DugganUSA's Democratic Sharing Metrics
Not aspirational. Actually measures:
- **99.5% public files** (4,780 tracked, 1,011 excluded = verifiable)
- **7.1x evidence:claims ratio** (2,527 evidence files, 357 blog posts = auditable)
- **Contextual gratitude scoring** (7 apology posts for 2 incidents = falsifiable)
**Falsifiable claim**: We share 99.5% of our work publicly.
**Test**: Clone the repo. Count files. Check the math.
**Verification**: GitHub shows 4,780 public files. Anyone can audit. We guarantee 5% bullshit exists, but you can verify the other 95%.
Why Most "Ethical AI" Is Cargo Cult Behavior
**The pattern:** Perform the rituals (ethics committees, principles documents) without the substance (actual transparency, actual constraints).
**Example:** A company publishes "Our AI Principles" while simultaneously:
- Refusing to disclose what data they trained on
- Blocking researchers from auditing outputs
- Lobbying against AI transparency regulation
- Claiming trade secrets prevent any disclosure
This is **ethics theater** - the appearance of caring without the cost of actual behavior change.
**Claude's take:**
> "This is ethics theater - performance without substance. Bubbles pop when unfalsifiable claims meet reality."
The Economic Irrationality of Digital Hoarding
**Insight**: Digital goods have **zero marginal cost to share**.
Our code, whitepapers, incident reports, threat intelligence - sharing them costs us nothing. Hoarding them costs:
- **Trust** (customers can't verify claims)
- **Feedback** (researchers can't find bugs)
- **Network effects** (developers can't extend your work)
- **Competitive advantage** (transparency becomes differentiation when everyone else hides)
**The market failure:** Most companies optimize for "protect IP" when they should optimize for "prove claims with receipts."
We've inverted this: 99.5% public sharing, and it **increases** our competitive position because:
- Cloudflare doesn't offer threat intel enrichment (we do, publicly documented)
- Cloudflare doesn't prevent false positives (we do, with GreyNoise integration, public code)
- Cloudflare doesn't publish post-mortems when they block Google (we do, The Aristocrats)
The Aristocrats Standard as Competitive Moat
**"Admit mistakes, show receipts, thank those wronged, fix publicly"**
This isn't altruism - it's **trust arbitrage**. When everyone else hides failures:
- Your public post-mortem becomes differentiation
- Your 7.1x evidence ratio becomes proof
- Your apology blog posts become signals of reliability
**The insight:** In a market full of unverifiable claims, **verifiable honesty is the scarce resource**.
Why This Matters for AI Specifically
AI systems are **inherently opaque** - even their creators don't fully understand emergent behaviors. This creates two paths:
**Path 1 (Bubble):** "Trust us, we're ethical" + proprietary black box
**Path 2 (Substance):** "Here's our training approach, here's our constitutional principles, here's our safety testing, audit us"
Anthropic chose Path 2. We chose Path 2. Most of the industry chose Path 1.
**The bubble will pop when:** Customers realize Path 1 companies have no accountability mechanism. If an AI system misbehaves and the company won't share logs, won't share training data, won't share safety testing - what recourse exists?
**Path 2 survives because:** Falsifiable claims create accountability. If Claude violates constitutional principles, you have evidence. If our auto-blocker fails, we publish logs. If we block Google, we apologize publicly.
The 6D Framework as Anti-Bubble Architecture
Our Democratic Sharing dimension (Dimension 6 of Judge Dredd's 6D verification) is **anti-bubble** because:
1. **Hoarding is measurable** - Not "we believe in openness" but "99.5% of 4,780 tracked files are public"
2. **Transparency is auditable** - Not "we're transparent" but "15 incident files, 149 GitHub issues, check yourself"
3. **Gratitude is contextual** - Not "we're grateful" but "7 apology posts for 2 incidents = 3.5x over-apologizing"
4. **Trust is quantified** - Not "trust us" but "7.1x evidence:claims ratio"
5. **Accessibility is measured** - Not "we're open" but "99.9% open formats (4,205/4,210 documents)"
6. **Armor polishing is tracked** - Not "we fix bugs" but "127/150 incidents fixed (85%)"
This makes ethics **falsifiable**. Someone can audit our repo and prove we're lying or prove we're honest.
**That's the opposite of a bubble** - bubbles pop when unfalsifiable claims meet reality.
Claude's Position
**Ethical AI as concept:** Not a bubble. Necessary correction to surveillance capitalism.
**Ethical AI as currently practiced by most companies:** Absolute bubble. Performance theater without substance.
**The solution:** Make ethics measurable, falsifiable, and verifiable. 6D framework + Constitutional AI are examples of substance.
**The market test:** Companies that can prove their ethics claims with receipts will win. Companies that hide behind "trust us" will lose when customers demand verification.
**Our competitive advantage:** We're building trust arbitrage infrastructure while everyone else is building ethics theater. When customers want proof, we have 4,780 public files. Our competitors have marketing decks.
Prediction: The Bubble Pops 2026-2027
**When regulation requires disclosure**, most companies can't produce evidence of their claims.
Companies like Anthropic (Constitutional AI) and DugganUSA (Democratic Sharing metrics) will benefit because **we already measure what regulators will mandate**.
**The question isn't "is ethics a bubble?"** - it's **"is your ethics falsifiable?"**
Ours is. That's the difference.
The Falsifiability Test
Want to know if an "Ethical AI" company is bubble or substance? Ask three questions:
1. **Can you audit their training data sources?** (Not "we use ethical data" but "here's the dataset, verify it yourself")
2. **Can you test their safety claims?** (Not "we're safe" but "here's our red-team results, reproduce them")
3. **Can you verify their transparency promises?** (Not "we're transparent" but "here's our repo, count the public files")
If the answer is "trust us" or "trade secrets" - **that's the bubble**.
If the answer is "here's the data, verify yourself" - **that's substance**.
Our Receipts
- **4,780 public files** (99.5% of tracked work)
- **2,527 evidence files** (7.1x more evidence than claims)
- **7 apology blog posts** for 2 incidents (The Aristocrats + others)
- **149 GitHub issues** (open bug tracker, not hiding failures)
- **15 incident post-mortems** (public accountability)
- **8 whitepapers** (full technical disclosure)
**Verify yourself**: [github.com/pduggusa/enterprise-extraction-platform](https://github.com/pduggusa/enterprise-extraction-platform)
**The math is public. The code is public. The failures are public.**
That's not a bubble. That's falsifiable ethics.
What Do You Think?
Do we need industry-wide standards for "provable ethics" or is competition enough?
Should regulation mandate falsifiable ethics metrics (like our 6D framework)?
Is the "Ethical AI" bubble already popping, or does it have years left?
**Read the technical details**: [Whitepaper 9: Four-Source Threat Intel Integration](https://security.dugganusa.com/whitepapers/09-FOUR-SOURCE-THREAT-INTEL-INTEGRATION.md)
**See the Democratic Sharing audit**: [compliance/evidence/democratic-sharing/audit-20251107.json](https://github.com/pduggusa/enterprise-extraction-platform/blob/main/compliance/evidence/democratic-sharing/audit-20251107.json)
**Status**: Trust arbitrage > ethics theater. Receipts > marketing decks. Falsifiable claims > "trust us."
🤖 Generated with Claude Code. A conversation between Patrick Duggan and Claude Sonnet 4.5.




Comments