Physical Intelligence dropped π0.7 on April 16. The viral demo: a UR5e bimanual arm folding laundry with zero training data on that exact hardware. The same weights also pull espresso shots, bag groceries, and figure out an air fryer that appeared in training exactly twice.
Dumb arms, cloud brain
π0.7 isn’t a robot — it’s a brain that ships actions to robots. The supported physical rigs are bimanual tabletop setups, Franka research arms, the open-source DROID Franka platform, and Physical Intelligence’s own PI Droid humanoid. None of them run inference locally. The arm hits a cloud endpoint over WebSocket, gets back ~100ms of joint actions, executes the chunk, and pre-fetches the next one mid-motion. Physical Intelligence calls it Real-Time Action Chunking. The hardware stays cheap; the intelligence stays in a data center.
The API any agent can call
The inference stack is open-sourced as openpi. One Python call — policy.infer({images, prompt})["actions"] — returns a chunk of joint trajectories. Send “pick up the fork,” or wire Claude or GPT to emit high-level goals like “clean the kitchen” and let π0.7 stream out the next 100ms of motion. Pre-trained base checkpoints cover 10k+ hours of robot data; per-platform expert checkpoints layer on top for fine-tuning.
On UR5e laundry folding, the generalist π0.7 matched first-attempt human teleoperators and beat the prior π*0.6 RL specialist trained only for that task.
You Might Also Like
- Notebooklm py the Unofficial Python sdk That Finally Gives Google Notebooklm a Real api
- Parlor Puts Real Time Voice and Vision ai on Your Laptop 2 6 gb no Cloud no api Keys
- Parlor Runs Real Time Voice Vision ai on a Macbook 2 6 gb Zero Cloud Zero api Keys
- Googles Gemini 3 1 Flash Live Scores 90 8 on Audio Benchmarks Real Time Voice ai Gets Serious
- Simile Just Raised 100m to Build ai Copies of Real People and its Wild

Leave a comment