DamageBDD and the Missing Foundation Beneath Modern Software Systems
- The problem is not insufficient tooling
- Failure is not caused by complexity — it is caused by unverifiable assumptions
- Behaviour is the only stable contract between humans and machines
- What DamageBDD changes at a structural level
- Why AI cannot solve this — and why DamageBDD can coexist with it
- DamageBDD as digital public infrastructure, not a testing tool
- Why this foundation has been missing for 50+ years
- Conclusion: make new mistakes — but make them provable
Why large-scale IT failures persist — and what has been structurally absent for decades
The problem is not insufficient tooling
It is the absence of enforceable, human-readable intent
The IEEE Spectrum article “The Trillion-Dollar Cost of IT’s Willful Ignorance” correctly identifies a truth that the software industry has been circling since the 1968 NATO Software Engineering Conference: software failures are not rare, novel, or mysterious — they are repetitive, predictable, and overwhelmingly human in origin.
What the article stops short of naming is the structural vacuum that allows this repetition to continue.
Over decades, we have added:
- More methodologies (Waterfall → Agile → DevOps → AI copilots)
- More abstractions
- More automation
- More capital
Yet we have never installed a foundation layer that binds human intent, organizational accountability, and machine execution into a single, verifiable system of record.
DamageBDD exists precisely in that missing layer.
Failure is not caused by complexity — it is caused by unverifiable assumptions
The article catalogs familiar disasters:
- Phoenix payroll
- Horizon/Post Office
- ERP collapses
- Automated decision systems (MiDAS, Robodebt)
- Defense and aerospace overruns
Across all of them, the common failure mode is not bad code in isolation.
It is this sequence:
- Human intent is informal, fragmented, and political
- That intent is translated into technical artifacts by intermediaries
- Assumptions are lost, softened, or hidden
- Execution proceeds without a shared, auditable truth
- When failure emerges, no one can prove what was supposed to happen
This is not a testing problem. This is not an AI problem. This is a semantic accountability problem.
Behaviour is the only stable contract between humans and machines
Software systems do not fail because requirements were unclear. They fail because requirements were never made executable, verifiable, and persistent.
DamageBDD’s foundational contribution is simple but radical:
The primary artifact of a software system should not be code — it should be behavior, written in human language, that the machine is forced to obey.
Behaviour-Driven Development (BDD) was always pointing in this direction, but historically it stopped at:
- Test frameworks
- Developer tooling
- CI conveniences
DamageBDD completes the arc by treating behaviour as infrastructure, not documentation.
What DamageBDD changes at a structural level
1. Human language becomes an executable asset
DamageBDD uses constrained natural language (Gherkin-style semantics) not as commentary, but as the authoritative definition of system behavior.
This matters because:
- Non-technical stakeholders can define outcomes
- Ambiguity is surfaced immediately
- “We thought it meant…” ceases to be defensible
The IEEE article repeatedly asks: Why haven’t we applied what we already know? DamageBDD answers: because knowledge was never encoded in a form the system could enforce.
2. Behaviour is verified continuously, not retroactively
Most catastrophic failures were “known” in advance — in reports, audits, warnings, or suppressed evidence.
DamageBDD eliminates the gap between knowing and verifying by making verification:
- Continuous
- Repeatable
- Independent of organizational hierarchy
A behaviour that fails cannot be hand-waved away by management optimism or vendor assurances. It either passes — or it doesn’t.
3. Accountability is cryptographic, not rhetorical
The Horizon scandal shows what happens when software claims authority without evidence.
DamageBDD introduces a property missing from almost all IT governance systems:
- Immutable execution records
- Attributable behavior definitions
- Proof that “this is what ran”
This directly addresses the article’s call for:
- Transparency
- Right to explanation
- Accountability in automated systems
Without this layer, “AI governance” is performative.
Why AI cannot solve this — and why DamageBDD can coexist with it
The IEEE article is correct: AI cannot manage organizational delusion, political pressure, or wishful thinking.
DamageBDD does not attempt to replace human judgment. It does something far more important:
It constrains human judgment with verifiable commitments.
AI tools can:
- Generate code
- Suggest tests
- Optimize execution
But they cannot decide:
- What should never be allowed to fail
- What risks are unacceptable
- What behavior is ethically or socially constrained
Those decisions must be made by humans — and then locked in place so they cannot be silently violated.
That is the role DamageBDD plays.
DamageBDD as digital public infrastructure, not a testing tool
Many of the failures described — payroll systems, welfare systems, licensing systems — are not “products”. They are infrastructure that citizens cannot opt out of.
For such systems, the minimum ethical bar is:
- Explainability
- Auditability
- Reproducibility
- Verifiable intent
DamageBDD aligns naturally with this requirement because:
- Behaviour is explicit
- Execution is observable
- Drift is detectable
- Blame cannot be abstracted away
This is why DamageBDD is not merely a QA improvement. It is a governance primitive.
Why this foundation has been missing for 50+ years
The uncomfortable answer — hinted at throughout the IEEE article — is that opacity benefits power.
- Vague requirements protect management
- Complex systems diffuse responsibility
- Post-hoc explanations replace proof
DamageBDD collapses that ambiguity.
When behaviour is:
- Human-readable
- Machine-enforced
- Publicly verifiable
Then:
- Failures are harder to deny
- Decisions are harder to evade
- Learning becomes unavoidable
This is precisely why such a foundation has not emerged organically.
Conclusion: make new mistakes — but make them provable
The IEEE article ends with a plea: “Make new ones, damn it.”
DamageBDD does not promise the end of failure. It promises the end of plausible deniability.
By anchoring software systems in executable human intent, DamageBDD provides the missing substrate beneath Agile, DevOps, AI, and whatever comes next.
Without such a foundation, we will continue to spend trillions rediscovering the same lessons — and calling them innovation.
With it, failure becomes visible, bounded, and finally instructive.