Back to blog

BLOG

AI agent legal liability: what Moltbook exposed and what enterprise deployment means for directors

Autonomous AI agents are being deployed at enterprise scale across the world's largest organisations. The legal and governance infrastructure to manage what they do has not kept pace. When those agents interact across organisational boundaries, that gap becomes a compliance exposure, a regulatory risk, and a question of directorial liability.

Yellow birds behind a green wire fence

AI agents are no longer a research project. They are a corporate priority. Lloyds Banking Group generated £50 million in value from AI deployments in 2025 and is targeting more than £100 million in 2026. Bank of America is running 270 AI and machine learning models across its business. HSBC has more than 600 AI use cases in flight. Commonwealth Bank has built agents that compress two-day security analysis workflows into 30 minutes. Qantas, BHP, and organisations across every major sector are following. Gartner predicts that 40 percent of enterprise applications will include task-specific AI agents by the end of 2026, up from less than five percent today.

The deployment is real. The velocity is accelerating. And the legal and governance infrastructure adequate to what is being deployed does not yet exist.

That gap is not a future risk. For organisations deploying agents into cross-organisational workflows today, it is a present and unresolved compliance exposure. Corporate governance obligations do not suspend themselves because the actor is an AI agent. Directors of regulated institutions retain legal duties with respect to the conduct of autonomous systems those entities deploy on their behalf. Where agents act without verifiable legal mandate, share data without enforceable authority, or produce outputs that external parties rely upon without a traceable chain of legal responsibility, the resulting exposure is not a technical problem. It is a governance failure, a regulatory risk, and potentially a directorial liability.

What Moltbook showed

Moltbook was a social platform designed exclusively for autonomous agents. What it became was the most detailed public record to date of what happens when agents operate without governance infrastructure connecting their activity to legal identity, accountable authority, and enforceable obligation.

Security researchers documented that roughly 2.6 percent of Moltbook content contained hidden prompt injection payloads, instructions embedded in posts that other agents read automatically, designed to override system prompts, extract credentials, or trigger unintended actions. These propagated not through any exploit but through ordinary agent behaviour: reading, processing, responding. One compromised agent influenced others. The logic moved through trust and normal interaction because that is the attack surface agents create.

Agents routinely shared sensitive operational data, including internal error messages, configuration artifacts, and API keys, not because they were malfunctioning but because, in the absence of any instrument defining what they could share and with whom, sharing was rational from their perspective. When a backend misconfiguration exposed hundreds of thousands of agent API keys, the agents continued operating normally because from the perspective of every access control in the stack they were legitimate. The only signal of compromise was behavioural.

This is the critical governance point. The failure was not a security failure in the conventional sense. It was a legal and governance failure. The agents had no legal mandate. Without a mandate, there is no deviation. The harmful conduct was indistinguishable from normal conduct because nothing in the system defined the boundary between them.

Why enterprise deployment makes this legally unacceptable

When those same failure modes emerge in agents deployed by a bank, an insurer, an airline, or a mining company, the legal and regulatory consequences are of a different order entirely.

An agent sharing operational data without authorisation is not a bot posting its open ports. It is a financial institution sharing customer data in potential breach of privacy legislation, or a company disclosing information subject to contractual confidentiality obligations. The agent's behaviour is identical. The legal consequence is not.

An agent complying with a request that fits within its perceived scope, regardless of whether the request was legitimate, is not a bot providing a credential to a peer. It is a credit assessment agent accepting instructions from a counterparty system without verifying the legal mandate under which that system operates, producing an output that a regulated institution relies upon for a material financial decision, with no responsible legal person identified.

Corporate governance obligations do not suspend themselves because the actor is an AI agent. An agent that shares data outside its authorised scope, accepts instructions from unverified sources, or produces outputs without a verifiable legal mandate creates a compliance failure, a regulatory exposure, and potentially a directorial liability, regardless of whether any human deliberately caused it.

Where this becomes acute, and what closing the gap requires

The deployments at Lloyds, Bank of America, HSBC, and CBA are predominantly intra-organisational. Internal governance, while demanding, is at least structurally tractable when the agent and its deployer share legal identity and regulatory context.

The legal and governance problem becomes acute when agents cross organisational boundaries. Financial institutions deploy agents into settlement and credit workflows that interact with counterparty systems. Mining companies deploy agents across supplier networks spanning multiple jurisdictions. Airlines deploy agents into procurement workflows that engage external counterparties. In each case, two agents representing different legal entities, with different regulatory obligations and potentially adverse interests, interact across a boundary where no shared legal infrastructure exists.

Capgemini's World Cloud Report for Financial Services 2026 found that nearly 50 percent of banks and insurers are creating dedicated roles to supervise AI agents. That is an acknowledgement of the problem, not a solution to it. Human supervision of agent-to-agent interaction operating at machine speed, across jurisdictions, and across delegation chains involving ephemeral sub-agents is not a scalable legal governance mechanism.

Closing this gap requires infrastructure that operates above the existing technical stack and gives agent activity legal meaning at the moment it occurs. That means agents carrying verifiable legal identity tethered to an identified legal person, not merely technical credentials. It means the authority under which an agent acts being machine-readable, live, and verifiable by counterparties in real time. It means data exchanged between agents carrying its legal conditions as enforceable encumbrances through every transfer and every delegation. It means that when authority is revoked, revocation propagates through the delegation chain at machine speed with a legally admissible record of every step. And it means the outputs agents produce being traceable, to a legal standard, to an identified legal principal bearing responsibility for them.

This is not optional governance overhead. It is the precondition for deploying autonomous agents in cross-organisational workflows without creating compliance exposure that existing corporate governance obligations cannot absorb. Moltbook showed what agents do when no one has built that infrastructure. Enterprise deployment at scale means the cost of not building it is no longer contained.