Russ Tedrake
VP, Robotics Research (and MIT prof)
+ Amazing university partners
"Dexterous Manipulation" Team
(founded in 2016)
For the next challenge:
For the next challenge:
Levine*, Finn*, Darrel, Abbeel, JMLR 2016
perception network
(often pre-trained)
policy network
other robot sensors
learned state representation
actions
x history
We've been exploring, and found something good in...
Image source: Ho et al. 2020
great tutorial: https://chenyang.co/diffusion.html
Image backbone: ResNet-18 (pretrained on ImageNet)
Total: 110M-150M Parameters
Training Time: 3-6 GPU Days ($150-$300)
(when training a single skill)
e.g. to deal with "multi-modal demonstrations"
Learning categorial distributions already worked well (e.g. AlphaGo)
Diffusion helped extend this to high-dimensional continuous trajectories
Standard LQR:
Optimal actor:
Training loss:
stationary distribution of optimal policy
Optimal denoiser:
(deterministic) DDIM sampler:
Straight-forward extension to LQG:
Diffusion Policy learns (truncated) unrolled Kalman filter.
converges to LQR solution
with TRI's Soft Bubble Gripper
Open source:
large language models
visually-conditioned language models
large behavior models
\(\sim\) VLA (vision-language-action)
\(\sim\) EFM (embodied foundation model)
Why actions (for dexterous manipulation) could be different:
should we expect similar generalization / scaling-laws?
Success in (single-task) behavior cloning suggests that these are not blockers
Big data
Big transfer
Small data
No transfer
robot teleop
(the "transfer learning bet")
Open-X
simulation rollouts
novel devices
In both ACT and Diffusion Policy, predicting sequences of actions seems very important
Thought experiment:
To predict future actions, must learn
dynamics model
task-relevant
demonstrator policy
dynamics
Cumulative Number of Skills Collected Over Time
w/ Chelsea Finn and Sergey Levine
Big data
Big transfer
Small data
No transfer
robot teleop
(the "transfer learning bet")
Open-X
simulation rollouts
novel devices
w/ Shuran Song
Big data
Big transfer
Small data
No transfer
robot teleop
(the "transfer learning bet")
Open-X
simulation rollouts
novel devices
w/ Dorsa Sadigh
Fine-grained evaluation suite across a number of different visual reasoning tasks
Prismatic VLMS \(\Rightarrow\) Open-VLA
w/ Carl Vondrick
Enough to make robots useful (~ GPT-2?)
\(\Rightarrow\) get more robots out in the world
\(\Rightarrow\) establish the data flywheel
Then we get into large-scale distributed (fleet) learning...
"Graphs of Convex Sets" (GCS)
Example: we asked the robot to make a salad...
Rigorous hardware eval (Blind, randomized testing, etc)
But in hardware, you can never run the same experiment twice...
w/ Chenfanfu Jiang
NVIDIA is starting to support Drake (and MuJoCo):
A foundation model for manipulation, because...
Some (not all!) of these basic research questions require scale
There is so much we don't yet understand... many open algorithmic challenges
https://www.tri.global/careers
We are hiring! (a little)
tri.global/careers