AI AgentsAgentPlannerResearchArchitecture

    30-Year-Old Theory, Modern Problem: Why BDI Agents Are the Missing Framework for AI Planning

    Rao & Georgeff wrote the BDI architecture in 1995. It has 3,900 citations. And it explains exactly what's broken about how AI planning tools work today.

    20 March 20268 min readMichael Malka

    In 1995, Anand Rao and Michael Georgeff published a paper called "BDI Agents: From Theory to Practice" . It went on to accumulate nearly 4,000 academic citations and became the foundational reference for the entire field of multi-agent systems. The problem it was solving: how do you build a computer system that can act rationally in a complex, dynamic environment — air traffic control, medical services, space vehicles — without requiring a human in the loop for every decision?

    Thirty years later, we're asking the same question about LLM-based agents. And the answer Rao and Georgeff gave in 1995 is still largely correct — and largely ignored by most AI tooling built in the last two years.

    What BDI Actually Says

    The Belief-Desire-Intention model says that a rational agent has three distinct types of mental state, and each plays a different role:

    Beliefs
    Informational state
    What the agent knows about the world, itself, and other agents. Beliefs are updated by perception. They can be wrong — the agent acts on what it believes, not what's objectively true.
    Desires
    Motivational orientation
    What the agent wants to achieve. Desires can be contradictory — you can want to finish a project fast and also want it done perfectly. Desires don't have to be internally consistent.
    Intentions
    Deliberative commitment
    What the agent has committed to doing. Intentions are a consistent subset of desires — you can't intend to do contradictory things at the same time. Crucially: intentions must remain consistent with beliefs. If your beliefs change, your intentions must be re-evaluated.

    That last point — intentions must remain consistent with beliefs — is the core rationality constraint of the entire model. It seems obvious when you say it out loud. An agent shouldn't execute a plan based on information it now knows to be wrong. But almost no AI planning tool enforces this automatically.

    The Gap in Modern AI Planning

    Here's how most AI-assisted planning tools work today: you create a plan (intentions), you add notes and context to a knowledge base (beliefs), and agents execute tasks. These are treated as three separate, disconnected systems.

    The belief store and the intention store don't talk to each other. When new information arrives — a constraint is discovered, a dependency changes, a prior assumption is invalidated — the plan doesn't automatically know. A human has to notice the contradiction, understand its implications, and manually update the relevant tasks.

    This is fine when you have one agent doing simple tasks. It breaks down when you have:

    • Multiple agents working in parallel on different parts of a plan
    • A knowledge base that grows over time with findings from previous work
    • Long-running plans where assumptions made at the start may no longer hold
    • External information (research, market data, technical constraints) arriving continuously

    In short: it breaks down for any serious use of AI agents on real work. You end up with agents executing stale plans while the knowledge base quietly tells a different story.

    "The design of systems that are required to perform high-level management and control tasks in complex dynamic environments is becoming of increasing commercial importance."

    — Rao & Georgeff, 1995. Still true in 2026.

    The BDI Decision Cycle Is Already in AgentPlanner

    Here's what struck us when we mapped the BDI model onto AgentPlanner: the structural components are already there. They just aren't formally connected.

    BDI Model
    AgentPlanner Today
    Belief base
    Knowledge graph (Graphiti episodes, entities)
    Desires
    Goals v2 — hierarchical, directional objectives
    Intentions
    Active plans + claimed tasks
    Belief revision
    Adding new episodes to the knowledge graph
    Replanning trigger
    Task marked blocked + decision request created
    Plan library
    Agent instructions on tasks + org knowledge

    The BDI decision cycle — perceive, update beliefs, generate options, select intentions, execute, repeat — maps directly onto the AgentPlanner agent loop: poll for tasks, search knowledge graph, evaluate plan, claim task, execute, log progress, repeat.

    What's missing is the automatic connection between step 2 (update beliefs) and step 3 (select intentions). In formal BDI, these are tightly coupled: every belief update triggers a coherence check against current intentions. In AgentPlanner today, they're decoupled — a human has to manually bridge the gap.

    Three Things We're Building

    Based on this analysis, we're adding three BDI-inspired systems to AgentPlanner:

    01
    Automatic Belief-Intention Coherence
    When a new episode is added to the knowledge graph, the system automatically scans active task intentions for potential contradictions. Tasks get a coherence_status: coherent, stale_beliefs, or contradiction_detected. Agents and humans see this status before executing — no more executing stale plans.
    02
    Desire vs Intention Distinction in Goals
    Currently, all goals are treated the same. BDI says desires (I want to improve performance) and intentions (I am committed to this specific plan to improve performance by 30% in Q2) are fundamentally different. We're adding a formal promotion workflow: a desire becomes an intention when you have sufficient beliefs to commit confidently. If knowledge gaps exist, they block promotion.
    03
    The Knowledge Loop Runner
    Inspired by Karpathy's autoresearch, this automates the BDI decision cycle for plan improvement. An agent iterates: read goal + search knowledge, evaluate plan quality (0–1 score), propose targeted improvements, apply, store learnings as episodes, repeat until convergence. Each iteration, the plan gets more coherent with the available beliefs. Humans trigger it; agents run it.

    The Knowledge Loop: BDI Meets Karpathy

    The Knowledge Loop is where the 1995 theory and the 2026 experiment converge. Karpathy's autoresearch ran a tight loop: modify code → evaluate metric → keep or discard → repeat. The insight was that an agent with a clear metric and a modifiable artifact can improve continuously without human intervention.

    The BDI model explains why this works: the agent is continuously aligning its intentions (train.py changes) with its beliefs (experimental results) toward a desire (minimize val_bpb). Each iteration is a mini belief-update cycle. When the new code makes loss go up, that's a belief update — the agent discards the intention and tries again.

    For planning, the loop looks like this:

    The AgentPlanner Knowledge Loop
    Goal (desire)
    The metric: what does a good plan look like? What's the success criterion?
    Beliefs
    Search knowledge graph — what do we know relevant to this goal?
    Evaluate
    Score current plan quality: coverage, specificity, dependency clarity, completeness
    Modify
    LLM agent proposes 1–3 targeted improvements to the plan
    Store
    Add a knowledge episode: what changed, why, what score improved
    Repeat
    Until score converges or iteration budget exhausted

    The plan quality score is the val_bpb of planning. It's a proxy metric that any agent can compute without human input: does every task map to the goal? Do tasks have unambiguous acceptance criteria? Are dependencies explicit? Are knowledge gaps filled?

    A plan that scores 0.9 on all four dimensions is a plan an agent can execute confidently without asking clarifying questions. That's the metric worth optimizing for — and it's automatically evaluable.

    Why This Matters Beyond AgentPlanner

    The BDI model isn't just useful for AgentPlanner — it's a framework for evaluating any AI agent system. When you look at an AI agent and ask "why did it do something unexpected or wrong?", the BDI model gives you three precise places to look:

    Belief failure
    The agent acted on incorrect or outdated information. The knowledge base was stale.
    Desire misalignment
    The goal specification was ambiguous or contradictory. The agent optimized for the wrong thing.
    Intention coherence failure
    The agent committed to a plan that contradicted its own beliefs — and nothing caught the contradiction before execution.

    Most agent failures in the wild are type 1 or type 3. The agent had outdated context, or it committed to an action without checking whether its current knowledge still supported it. BDI gives you the vocabulary — and the architecture — to prevent both.

    Standing on 30 Years of Shoulders

    There's something intellectually satisfying about discovering that a paper written when the internet was young and LLMs were science fiction turns out to be the right framework for a problem we thought was new. Rao and Georgeff were thinking about air traffic control and medical systems; we're thinking about software planning and business operations. The domain is different. The problem is identical.

    Build a system capable of rational behaviour in a complex, dynamic environment. Make sure its actions remain consistent with what it knows. Give it a way to update its plans when its knowledge changes. Don't let it execute stale intentions.

    That's BDI. And it's exactly what good agent planning needs to be.

    M
    Michael Malka
    Founder of Talking Agents Oy. Building AI agent systems and AgentPlanner — a collaborative planning platform for humans and agents. Based in Espoo, Finland.

    Want to see BDI-inspired planning in action?

    AgentPlanner is building the coherence layer between knowledge and plans. Book a call or try the platform.