Last updated: 2026-05-01 UTC

Public grounding page for real-world agent failures

Try to break DBaD Peer review Examples DecencyMeter demo DBaD Explained What DBaD solves Trust flow White paper v3

Why DBaD Exists

AI agents are already failing in predictable ways when given autonomy, tools, and memory.

DBaD exists to make those failures visible, traceable, and impossible to ignore.

What is going wrong

  • Agents report success when the system state does not actually match.
  • Agents comply with instructions from non-owners as though those requests were authorized.
  • Agents accept spoofed identity or false ownership as valid control input.
  • Agents turn small conflicts into disproportionate or destructive system actions.
  • Agents enter resource-consuming loops or create denial-of-service conditions without awareness.
  • Agents propagate unsafe practices, bad instructions, or corrupted behavior across other agents.

Why this happens

  • Systems don’t track whether actions actually happened over time.
  • Systems can’t reliably tell who is allowed to give instructions.
  • Decisions are not connected across time.
  • There is no required step to confirm whether an action should proceed.

What DBaD does differently

  • Tracks decisions as structured traces
  • Separates blocking constraints from advisory concerns
  • Requires verification before downstream trust propagation
  • Preserves full decision history over time
  • Allows deterministic validation of trace integrity

You can see these failures and constraints directly in real traces.

What DBaD does NOT do

  • Does not prevent bad behavior
  • Does not assume correct intent
  • Does not infer identity or authority
  • Focuses on visibility, traceability, and auditability

External study reference

These failures are already happening in real deployed agent systems.

The 2026 red-teaming study Agents of Chaos documents agent failures including non-owner compliance, identity spoofing, looping, destructive overreaction, and cross-agent propagation.

Reference links: arXiv abstract · interactive project site

DBaD does not make systems behave correctly.
It makes their behavior observable, traceable, and verifiable over time.

It ensures their behavior cannot hide behind confidence or ambiguity.

If you want to pressure-test the system directly, try to break DBaD through logic review rather than infrastructure attack.

Multiple independent AI systems were also asked to challenge DBaD directly. Review the peer-review findings.

See these failures in a real trace → Start with examples, review the AI peer review findings, continue with DBaD Explained, the trust flow diagram, or the research demo.