Writing

Intelligence,
unpacked.

Research notes, industry analysis, and practical insights on agentic systems, LLMs, and the future of AI.

The OpenAI Trial's Real Revelation: Stuart Russell's AGI Arms Race Warning
AI EngineeringMay 22, 2026

The OpenAI Trial's Real Revelation: Stuart Russell's AGI Arms Race Warning

Stuart Russell testified at the OpenAI trial that he's concerned about an AGI arms race. This is the same researcher who literally wrote the textbook on AI. His testimony reveals a fracture inside the AI safety community about how to think about frontier AI risk. Here's what's actually happening.

Recent

Multi-Agent Orchestration at Scale: Managing the Complexity of Many Agents
Agentic EngineeringMay 15, 2026

Multi-Agent Orchestration at Scale: Managing the Complexity of Many Agents

Running one agent is manageable. Running ten with different capabilities, dependencies, and failure modes is a systems engineering challenge. Orchestrating hundreds of agents simultaneously — with dynamic scaling, cross-agent communication, resource allocation, and fault tolerance — requires architectural patterns that most teams discover by failure. Here's the orchestration playbook.

Agent Prompt Engineering Patterns: Designing Prompts That Actually Scale
Agentic EngineeringMay 14, 2026

Agent Prompt Engineering Patterns: Designing Prompts That Actually Scale

Prompt engineering for agents is fundamentally different from prompt engineering for chatbots. Agents need prompts that define behavior, constrain actions, guide tool use, handle edge cases, and remain coherent across thousands of sessions. This post covers the prompt patterns that production agentic systems actually use — from system prompt architecture to few-shot demonstration design.

Structured Output from AI Agents: When the LLM Must Return Machine-Readable Data
Agentic EngineeringMay 13, 2026

Structured Output from AI Agents: When the LLM Must Return Machine-Readable Data

Agents that generate natural language responses are easy to build. Agents that produce structured, machine-readable outputs — JSON, XML, form data, database records — require solving a different class of problems: output format enforcement, parse failure recovery, schema evolution, and the fundamental tension between flexibility and reliability. Here's the engineering playbook.

Human-in-the-Loop Agents: Designing Checkpoints That Actually Work
Agentic EngineeringMay 12, 2026

Human-in-the-Loop Agents: Designing Checkpoints That Actually Work

Most human-in-the-loop implementations are theater — they pause the agent, show a modal, and let the human click 'approve.' Real HITL design is harder: it requires deciding what to show, what to ask, how to present context efficiently, and what to do when the human is unavailable. Here's the engineering framework for checkpoint design that balances safety, usability, and operational overhead.

Cutting the Cost of AI Agents: Token Efficiency, Model Routing, and Context Optimization
Agentic EngineeringMay 11, 2026

Cutting the Cost of AI Agents: Token Efficiency, Model Routing, and Context Optimization

Running AI agents in production is expensive — and most teams don't realize how much of their spend is waste. The median agentic workflow costs 10-50x more per task than a well-optimized version. This post breaks down where the money goes, which optimizations actually move the needle, and the cost-per-task benchmarks that should be on every agentic engineer's dashboard.

Uber's New Business: Turning Drivers Into a Sensor Grid for Self-Driving Cars
Agentic EngineeringMay 5, 2026

Uber's New Business: Turning Drivers Into a Sensor Grid for Self-Driving Cars

Uber is equipping its millions of drivers' vehicles with sensor kits and selling the data to autonomous vehicle companies. They already have 25 AV company partners including Wayve. The business model is elegant: human drivers become unwitting data collectors, and Uber becomes the middleware between physical world data and AV training. Here's what's actually happening and why it matters.

HS
Hardware SecurityMay 5, 2026

Differential Power Analysis on Neural Networks: Recovering Model Weights at the Physical Layer

The side-channel attack surface of neural networks goes well beyond timing and acoustic emissions. Differential power analysis — a technique borrowed from cryptographic hardware attacks — can extract precise model weights from an accelerator's power draw. I spent years building and breaking hardware this way, and what I'm seeing in the AI accelerator space is a security community that's still catching up.

HS
Hardware SecurityMay 5, 2026

AI Chip Security Certification: Why Common Criteria and IEC 62443 Matter for AI Accelerators

The AI chip industry is deploying hardware at a pace that makes the smartcard security incidents of the 2000s look cautious by comparison — and it's doing so without the formal certification infrastructure that cryptographic hardware spent two decades building. Common Criteria and IEC 62443 exist precisely to address this gap, but almost no AI accelerator vendors are using them.

HS
Hardware SecurityApril 30, 2026

Confidential Computing for AI Inference: Securing Model Inference in Untrusted Environments

Your AI inference pipeline runs on hardware you don't own, in infrastructure you don't control, managed by software you haven't audited. That's not a hypothetical — it's the default deployment model for cloud AI. Confidential computing is the only technology stack that addresses this reality directly. Here's how it works, where it actually helps, and what it can't do alone.

HS
Hardware SecurityApril 9, 2026

Hardware Root-of-Trust in Cloud AI Infrastructure: Why Software Security Alone Isn't Enough

Every layer of your cloud AI stack — model weights, training data, inference pipelines — sits on physical hardware you don't control. Most AI security programs treat this as someone else's problem. It isn't. Here's what hardware root-of-trust actually means for AI infrastructure, and why the software-only security posture that most AI teams rely on has a fundamental gap.

HS
Hardware SecurityApril 8, 2026

Side-Channel Attacks on ML Accelerators: The Hardware Security Threat AI Teams Are Ignoring

Your ML model's weights are leaking. Not through your API. Not through a data breach. Through power consumption, electromagnetic emissions, and timing variations in the hardware running your inference workload. Side-channel attacks on machine learning accelerators are not theoretical — they're reproducible, they're getting more accessible, and almost no one building AI systems is defending against them.

AA
Agentic AIApril 7, 2026

Agentic AI Governance: The Safety Framework We Actually Need (From Someone Building These Systems)

Most enterprise AI governance frameworks were designed for predictable, human-in-the-loop systems. Agentic AI — systems that plan, execute multi-step tasks, and delegate to other agents — breaks every assumption those frameworks are built on. Here's what a real agentic AI governance framework looks like, from someone building and deploying these systems in production.