Skip to Content
HeadGym PABLO
Skip to Content
PostsIndustry and Societal ImpactCargo-Cult Agent Design: Avoiding Illusion in AI Prompts
Tags:#ai_and_agents#software_engineering

Cargo‑Cult Agent Design: Why Most “Advanced” Prompts Don’t Work

Modern AI systems are increasingly governed by text.

Not code.
Not constraints.
Not enforcement.

Text.

Long system prompts now claim to define autonomy, quality control, expertise, and decision‑making. They read like operating manuals. They look serious. They feel engineered.

And yet, most of them do nothing.

This is not a failure of large language models. It is a failure of engineering judgment. What we are witnessing—especially in agent and prompt design—is a classic case of cargo‑cult engineering.


Cargo‑Cult Engineering, Applied to AI

Cargo‑cult engineering is the practice of copying the visible form of a successful system without implementing the causal mechanisms that make it work.

The term comes from Richard Feynman’s famous description of post‑war Pacific islanders who, after seeing military aircraft deliver supplies, rebuilt what they remembered:

  • runways
  • control towers
  • radio headsets

All carved from wood and bamboo.

They reproduced the appearance of the system perfectly.
The planes never came back.

In AI, prompts are our bamboo control towers.


The Cargo‑Cult Turn in Prompt Writing

As AI systems have become more capable, prompt writing has become more theatrical.

Instead of asking what the system can enforce, prompts increasingly describe what the system should aspire to be.

You see instructions like:

“Include quality control mechanisms and self‑verification steps.”

Or:

“You are an elite AI agent architect specializing in crafting high‑performance agent configurations.”

These statements sound responsible. They signal rigor. They resemble engineering.

But they collapse under a single question:

Where is the mechanism?

There is no verifier.
No external check.
No failure mode if “self‑verification” is skipped, ignored, or hallucinated.

The prompt names a concept without instantiating it. That is the defining move of cargo‑cult engineering.


Persona Inflation Is Not Control

One of the most common rituals in cargo‑cult agent design is persona inflation.

“Elite.”
“Expert.”
“Autonomous.”
“High‑performance.”

These words do not constrain behaviour.
They do not grant authority.
They do not enforce correctness.

They are narrative devices.

Personas can influence tone. They cannot substitute for permission models, budget limits, escalation paths, or validation loops. Calling a model “elite” does not make it reliable—just as wearing a bamboo headset does not connect you to air traffic control.


Checklists Without Causality

Cargo‑cult prompts love checklists:

  • Extract core intent
  • Design expert persona
  • Architect comprehensive instructions
  • Optimize for performance

Each item gestures at something real. None of them guarantees it happens.

Optimization without metrics is fiction.
Escalation without an escalation path is theater.
Quality control without an external check is wishful thinking.

These lists simulate engineering seriousness without doing engineering work.


Example Theater and Narrative Compliance

Another warning sign is example theater: elaborate fictional dialogues showing how an agent ought to behave.

Examples bias models, but they do not enforce behavior. If the runtime environment does not require tool use, penalize deviation, or validate outputs, examples are just stories.

Real systems rely on constraints.
Cargo‑cult systems rely on narrative compliance.


A One‑Question Test for Cargo‑Cult Agents

There is a simple diagnostic for cargo‑cult agent design:

What breaks if this instruction is ignored, misunderstood, or hallucinated?

If the answer is “nothing obvious,” the instruction is decorative.

More bluntly:

If removing it would not change system behavior, it is not doing real work.


Why Cargo‑Cult Prompts Persist

Cargo‑cult agent design survives because it satisfies organizational incentives:

  • It looks sophisticated
  • It produces impressive artifacts
  • It deflects critique (“we follow best practices”)
  • It avoids hard architectural decisions

But it is brittle. It accumulates complexity without leverage. And when it fails, it fails silently—until it fails catastrophically.


What Non‑Cargo‑Cult Agent Design Looks Like

Real agent design starts from a different assumption:

The model will misunderstand you. Design accordingly.

Non‑cargo‑cult agents are built around mechanisms, not prose.

Principles of non‑cargo‑cult agent design:

  • Constraints beat descriptions
    Define what the agent cannot do before describing what it should do.

  • Verification is external
    Quality control lives outside the model, not inside its instructions.

  • Failure is explicit
    Agents have clear abort paths, escalation rules, and fallback behaviors.

  • Authority is scoped
    Every action is permissioned. Autonomy is narrow, revocable, and observable.

  • Behavior is measurable
    Outputs are structured, testable, and auditable—not merely plausible.

In short:

Real agents are governed. Cargo‑cult agents are narrated.


The Runway Lesson, Relearned

The islanders were not foolish. They were rational, given what they could see.

They rebuilt the visible parts of the system. They could not see the invisible ones.

Prompt writers today make the same mistake. The prompt is visible. It feels powerful. It feels like control.

But planes do not land because runways look right.

They land because of everything you don’t see.

Until AI systems are designed with that humility—assuming failure, misunderstanding, and non‑compliance—we will keep building beautiful runways and wondering why nothing arrives.

Last updated on