AI agent confabulation is not random. It is plausible-looking wrongness: output constructed from pattern and proximity rather than knowledge. It fits the shape of what should be there. This makes confabulation harder to detect than obvious failures, precisely because the output looks correct. The failure mode is not garbage; it is convincing fiction.
This is the convergent failure mode of the suggestible actorβs other three properties. The agent is goal-oriented (so it must produce something), locally reasoning (so it draws only from whatβs nearby), and susceptible to local context (so it pattern-matches from whatever is available). When the directive gap is wide and local context is insufficient, the agent fills the void with plausible structure β a call to an API that does not exist, a convention that was never established, a security bypass that βshould work based on the patterns in this codebase.β