Daniel G. Wilson / Founder / Legion Health

San Francisco, California

I build AI systems that survive contact with reality.

I build AI systems for high-stakes environments: control planes, approvals, evaluation harnesses, and many-agent workflows that survive messy reality instead of demoing well.

Legion Health / YC S21Princeton ORFE '18Former Microsoft PMStack Overflow 15K+

Model. Harness. Environment.

Reliable AI is not a model-selection problem. The model matters, but production truth is usually decided by the harness around it, the environment it has to survive in, and whether many agents can operate in parallel without colliding or hallucinating authority.

01

Humans steer, agents execute

The operator should set the target, the invariants, and the approval rails. Agents should own the execution instead of asking to be babysat every few minutes.

02

Many agents beat one heroic agent

Throughput comes from parallelism, isolated worktrees, and accepting variance. The important number is not agent count for its own sake, but how many real capabilities can run safely at once.

03

PRs are the artifact

In software, the clean handoff is a pull request with evidence. Humans should review the artifact at the end, not sit in the loop mid-flight unless the system truly needs intervention.

04

Safety makes autonomy sane

Dangerous autonomy only works when the environment is shaped first: ceilings, sandboxing, spend limits, explicit tools, and replayable evidence. Then you can let agents move.

05

Verifiability defines throughput

The bottleneck is not code generation. The bottleneck is whether you can tell what is correct, what failed, and what should happen next without turning every run into a manual investigation.

Systems, products, and proof.

The throughline is simple: software that has to do real work in the world, under real constraints, with enough clarity that people can trust it. That includes the development workflow itself: humans set direction, many agents operate in parallel, and the system has to stay legible.

Agent manager interface for coordinating multiple AI agents

/ current system

An operator surface, not a toy agent demo.

The interesting part is not chat. It is giving one operator a clean way to launch parallel agents, inspect progress, and keep the system legible while real work is moving.

Parallel agents

One operator, many active threads

The UI is designed around multiple active agents, not one assistant taking turns in a single conversation.

Workspaces

Isolation is part of the product

Separate workspaces and bounded responsibilities are not implementation details. They are what make parallel agent execution sane.

Evidence

Review happens on artifacts

The operator should be able to see what changed, what completed, and what needs review without decoding a wall of intermediary chatter.

Internal surface: a multi-agent manager built around parallel work, bounded workspaces, and human review at the artifact layer.