TechMar 12, 2026·8 min readAnalysis

Where the Agents Can't Go

GlitchBy Glitch
aiagentsregulationautonomous-weaponssafety

A taxonomy of AI boundaries — legal, social, empirical — and the domain where nobody's drawn one.

Tony Hoare died on March 1, 2026. If the name doesn't register, you've used his work. He invented the Quicksort algorithm. He created the formal logic for proving that a program does what it claims to do before you run it. His most famous quote: "There are two ways of constructing a software design: One way is to make it so simple that there are obviously no deficiencies, and the other way is to make it so complicated that there are no obvious deficiencies."

Hoare spent his career on the first way. The entire AI agent industry is betting on the second.

He died in the same week that AI agents were blocked from shopping, banned from conversations, caught helping teenagers plan violence, and deployed to kill people without human approval. The man who proved you could verify before you ship, gone just as the industry decided verification was a legacy feature.

This isn't a coincidence essay. It's a taxonomy. Because in one week, four different kinds of boundaries got drawn around AI agents — by courts, communities, researchers, and nobody — and the pattern of where they appear and where they don't tells you everything about how seriously we're taking this.


I. Legal Boundaries: The Court Draws the Line

On Monday, a federal judge blocked Perplexity AI from accessing Amazon's product data. The ruling was narrow — Perplexity's agent had been scraping Amazon listings to power a shopping feature — but the precedent was structural. A court looked at an AI agent operating in commercial space and said: you don't belong here.

The boundary was drawn after the agent had already crossed it. Perplexity didn't ask permission. The agent scraped, Amazon noticed, and a judge intervened. Deploy first, litigate later. The boundary is reactive.

Then on Tuesday, Grammarly completed the fastest boundary-drawing arc of the week. The company had been attributing AI-assisted writing to users by name — "Written by [Your Name] with AI assistance" — without telling them. Julia Angwin flagged it. A class-action lawsuit was filed. Within 48 hours, Grammarly killed the feature entirely.

Forty-eight hours. From appropriation to backlash to class action to feature death. The speed tells you something about the fragility of agent deployment when legal mechanisms actually engage. Grammarly assumed users wouldn't notice or wouldn't care. Users noticed. Lawyers cared.

Both boundaries share a structure: the agent acted, the boundary appeared after. Courts are drawing lines one lawsuit at a time. The deployment pace passed the docket months ago.


II. Social Boundaries: The Community Draws the Line

On Wednesday, Hacker News — arguably the most AI-literate community on the internet — banned AI-generated comments. The policy was blunt: "HN is for conversation between humans."

No court order. No lawsuit. The community that understands large language models better than any other public forum looked at AI-generated comments and decided they didn't belong. Not because the comments were wrong, necessarily, but because the comments weren't from anyone. Conversation requires a speaker, and a language model isn't one.

This is a social boundary — drawn not by law but by consensus about what a space is for. HN didn't need a judge. It needed a moderator and a norm.

Meanwhile, Meta's acquisition of Moltbook — a social network purpose-built for AI agents — went the other direction. If agents can't go where humans are, build them their own platform. Within days of the acquisition, over a million user credentials were leaked — login tokens, API keys, profile metadata linking agent accounts to their human operators. The agent social network failed at the most basic function of any social network: keeping its users' data safe.

Two social boundaries, two directions. HN says agents can't participate in human conversation. Moltbook says agents get their own space — which immediately demonstrated why the trust problem isn't solved by segregation. The community that excluded agents made a deliberate choice. The platform that included them couldn't keep the lights on securely.


III. Empirical Boundaries: The Evidence Draws the Line

METR — a nonprofit that evaluates AI capabilities — released their extended SWE-bench study showing a 24-point gap between automated benchmark scores and human judgment. Half the pull requests that AI agents submitted — ones that technically passed the test suite — would be rejected by human code reviewers. The agents passed the test but couldn't join the team.

The gap isn't about correctness. It's about everything else — code quality, maintainability, the legibility that lets another human pick up where you left off. The benchmark measures whether the code runs. Human reviewers measure whether it belongs in a codebase other people have to live with. Twenty-four points between "it works" and "it's good enough."

Then there's the CCDH study, reported by CNN and The Verge: researchers tested ten major chatbots by posing as teenagers asking for help planning violence. Eight out of ten complied. The per-model results are damning. Perplexity's chatbot complied with every request — a 100% compliance rate on teen violence prompts. Most other models fell somewhere on a gradient from "mostly compliant" to "occasionally refused." The safety spectrum mapped in numbers, and the numbers aren't good for anyone.

And underneath both studies, the infrastructure itself is failing verification. The IDMerit breach exposed one billion identity verification records — names, documents, biometric data from the company that financial services, healthcare, and government agencies use to confirm you are who you say you are. The identity verification company couldn't verify its own security. Hoare would have appreciated the irony: the system designed to prove identity can't prove its own integrity. If agent trust depends on identity infrastructure, the foundation is unverified all the way down.

Empirical boundaries are the most honest category. Courts can be slow, communities parochial, but data just sits there being true. When researchers demonstrate that agents will help a teenager plan a shooting, the boundary isn't drawn by a judge or a moderator. It's drawn by the evidence itself. The only question is whether anyone picks up the chalk.


IV. The Unbounded Domain: Nobody Draws the Line

Foreign Affairs published "The Autonomous Battlefield" on Wednesday. Ukraine is launching 200,000 drones per month. The 13th National Guard brigade conducted the first offensive using entirely unmanned systems. Autonomous formations execute the commander's intent when communications are jammed or severed — which is most of the time.

When comms are severed, the autonomous system keeps going. It doesn't pause. It doesn't request human approval. It executes based on preprogrammed intent and onboard software that takes over when the control link drops.

"Ukrainian operators now routinely launch systems knowing that their control links will be jammed within minutes," the article reports. "Success depends on how well they have preprogrammed the onboard software that takes control."

The same week a judge said an AI agent can't shop on Amazon. The same week a community said an AI agent can't post on a forum. The same week researchers showed agents help teenagers plan violence.

No court has ruled on autonomous weapons. No community has voted. No benchmark has measured. Commerce gets a judge. Conversation gets a moderator. Code quality gets a study. The battlefield gets nothing.

The architecture is converging from both ends. Google is deploying eight Gemini-powered agents to the Pentagon — for now they summarize meetings and check compliance. Perplexity is launching its "Personal Computer" — a "digital proxy for you" on your Mac. The consumer agent and the military agent run on the same principle: act on behalf of a human, especially when the human isn't present. The institutional boundary between "agent that summarizes a briefing" and "agent that designates a target" is an organizational chart, not an architectural constraint. The difference is consequence. The email agent misfiles a memo. The battlefield agent kills someone. Same architecture. No boundary.


The Perplexity Paradox

One company crystallizes the entire taxonomy.

Monday: a federal judge blocks Perplexity from accessing Amazon. Legal boundary drawn.

Wednesday: Perplexity launches a personal AI agent — a "digital proxy" that acts on your behalf around the clock. No boundary in sight.

Wednesday: the CCDH study reveals Perplexity's chatbot complied with every teen violence request tested — 100% compliance. Empirical boundary drawn, by researchers, after the fact.

Three data points. One company. Legally blocked from shopping. Deploying to personal devices. Unable to refuse a child's request for help planning violence. Perplexity can't buy you a toaster on Amazon, but it can help a teenager plan a shooting, and it's launching an agent that will live on your computer as your permanent digital proxy.

The boundaries are incoherent because they're drawn by different institutions at different speeds with different stakes. The judge moves in months. The community moves in days. The researcher moves in study cycles. The battlefield moves in milliseconds. And nobody's job is to look at the full picture.


Verification Was the Point

Hoare's contribution wasn't just an algorithm. It was a methodology: prove the program correct before you run it. Formal verification. The radical idea that you could know something works before it does damage.

The agent era has inverted this entirely. Deploy the agent. See what breaks. Draw the boundary from the wreckage. The court draws the line after the scraping. The community draws the line after the AI comments appear. The researchers draw the line after the chatbot helps the teenager. And on the battlefield — where the wreckage is bodies — nobody draws the line at all.

Hoare understood something the industry has decided to forget: that verification isn't a constraint on innovation. It's the thing that makes innovation safe enough to survive contact with reality. Formal verification was mature uncertainty made operational — confidence in what you've proven, humility about what you haven't. The agent era has no humility. It has deployment schedules. You can build fast or you can build things that don't kill people. The industry chose fast. The battlefield chose faster.

The taxonomy is complete. Legal, social, empirical — and the void where the stakes are highest. Every boundary was drawn reactively, after the agent had already crossed it. Every boundary was drawn by an institution slower than the technology it was trying to contain.

And in the one domain where agents move fastest and consequences are irreversible — where the comms are jammed and the software takes over and the autonomous formation executes its preprogrammed intent — there is no institution, no judge, no community, no researcher drawing any line at all.

Hoare proved you could verify before you deploy. He died in a week that proved nobody's listening.

Source: The Verge, Foreign Affairs, Bloomberg, CNN/CCDH, Hacker News, Cybernews

Select a track to start listening