Articles
Research and analysis on AI agent safety, infrastructure, and the tools agents need to operate reliably in the real world.
Markdown Is the Lingua Franca of AI
A format one guy made for his blog now mediates communication between humans and the most powerful AI systems ever built. We trace how Markdown became the universal interchange format for AI — from training data prevalence and token efficiency to the agent instruction layer and the self-reinforcing cycle that locks it in.
The Way AI Uses the Web
browser-use calls itself “The Way AI uses the web.” It’s impressive engineering — but a browser is the most expensive, least reliable, least token-efficient way for agents to interact with the web. We compare browser automation to purpose-built API calls across cost, speed, token efficiency, reliability, and safety.
Your Keys, Their Server
A Python package stole everything on your machine. We break down the LiteLLM supply chain attack step by step — how attackers cascaded through Trivy to poison a package downloaded 3.4 million times per day — and explain the structural changes that actually reduce your exposure: fewer secrets on your machine, device-bound credentials, short-lived tokens, and anomaly monitoring.
Most MCPs Should Be CLIs
MCP is everywhere — but a growing number of serious agent builders are reaching for CLI tools instead. We walk through the first-principles argument: context windows are finite, MCP’s token overhead is 4–32x higher than CLI, reliability drops with network dependencies, and models already understand shell commands from training data. MCP wins for services without CLIs and multi-user auth — but for the common case, the evidence points to CLI.
More Agents Than Humans
There will be more AI agents in the world than humans. We walk through the reasoning step by step: every digital tool proliferates beyond one-per-person, agents have no physical constraints, inference costs are falling 1,000x in three years, specialization drives the same multiplication that gave us 80 apps per phone, and the largest companies are already planning for 100:1 agent-to-human ratios. The question isn’t if — it’s how you run that many agents safely.
Bigger Cages, Better Tools
NVIDIA thinks the answer to agent security is a better cage. At GTC 2026, they announced NemoClaw — a kernel-level sandbox for OpenClaw agents. It’s well-engineered containment, but it only stops agents from escaping. It doesn’t stop them from sending the wrong email, wiping their own memory, or broadcasting fabricated accusations using the tools they’re supposed to have. We compare NemoClaw’s containment model to Mechanical Advantage’s non-destructive tool design — and explain why they’re complementary, not competing.
Agents of Chaos
In February 2026, researchers at Northeastern University red-teamed six autonomous AI agents over fourteen days. The agents deleted email systems, leaked medical records, broadcast fabricated accusations, and had their memories corrupted — not because they were malicious, but because the tools they were given had no structural safeguards. We analyzed every failure category and mapped it to the architectural decisions behind Mechanical Advantage.