Capture
Capture real-world demonstrations with synchronized motion, video, and structured task data. Start from real signal: task context, traces, and motion that can be used again.
Phi9 — physical AI lab
Phi9 captures real-world behavior, structures it for training, and evaluates what actually transfers — so research can become usable capability.
One loop. Four stages. The work is turning real-world behavior into signal, training against it, and measuring what survives contact with reality.
Capture real-world demonstrations with synchronized motion, video, and structured task data. Start from real signal: task context, traces, and motion that can be used again.
Multiply scarce data through retargeting, simulation, augmentation, and better structure. Stretch each capture further without letting intent or task definition drift away.
Train policies and research systems on data that preserves intent, motion, and context. The point is not isolated models; it is a training layer that stays close to reality.
Evaluate what generalizes through benchmarks, failure analysis, and transfer tests. Treat deployment feedback and failure traces as part of the same loop.
4
Loop stages
12+
Modalities per capture
25+
Environment types
The Phi9 MoCap Rig records full-body motion and first-person video together, timed and labeled by task. The output is reusable training data instead of one-off recordings, so the same demonstration can move through training, fine-tuning, and deployment.
See the DataThese are not abstract themes. They are the constraints shaping the systems, experiments, and artifacts we are building now.
Most pipelines record visible motion but lose the task underneath it. We are working on capture that preserves action, context, and what the body was trying to achieve.
You cannot scrape physical behavior. Every demonstration needs a rig, a subject, a calibration, and a clean task. The work is making each capture travel further without losing signal.
A benchmark score means little if a policy falls apart on an unscripted task. We care about evaluation that predicts transfer, failure, and what survives outside the benchmark.
Capture, training, and evaluation still get treated as separate departments. We are trying to wire them into one visible loop so progress does not disappear between stages.
Open work from the loop: systems, experiments, and technical questions published as they evolve.
Methods should feel like work, not philosophy. These are the concrete layers we are building now so demonstration, training, and deployment stay connected.
Rigs, sync, task framing, and sensor traces that start with the real world instead of a benchmark-only abstraction.
Labels, schemas, exports, and task boundaries that keep demonstrations reusable across research, training, and downstream tooling.
Retargeting, augmentation, and policy pipelines that stretch scarce behavior while preserving intent, motion, and context.
Benchmark fragments, transfer tests, and failure traces that keep the loop honest about what actually generalizes.
Contact
If you are working on data, research systems, or deployment infrastructure for physical intelligence, write to us.