28 April 2023
By Shao Yuan Chew Chia
Joint work with: Terry Suh, Pang Tao, Sizhe Li
Motion planning for high dimensional contact-rich manipulation
Classes of Methods | Trajectory Optimization | Sampling based motion planning | Learning based methods |
---|---|---|---|
Examples | Continuous trajectory optimization |
Rapidly exploring Random Tree (RRT) | Reinforcement Learning, Behavior cloning |
How contact sampling can help | Provide good initial guesses | Increase state space coverage of "extend" step | Improve sample efficiency |
Uses optimal transport to define a heuristic for which particles need to move the most
Input: \(x\), \(x_g\)
Output: \(N\) best \(q\)'s
\(x_g\) = [0.3, 0, 0]
\(x_g\) = [2, 0, 0]
std = 0.01
std = 0.1
std = 0.01
std = 0.1
std = [0.01 , ... , 1]
Min
Mean
Max
Andrews, S., Erleben, K., & Ferguson, Z. (2022). Contact and friction simulation for computer graphics. In ACM SIGGRAPH 2022 Courses (pp. 1-172).
Incremental steps = 10
Incremental steps = 5
Incremental steps = 1
Incremental steps = 10
Incremental steps = 5
Incremental steps = 1
We want to solve
To use zeroth-order methods, we can instead choose a surrogate objective
where
which has the unbiased estimator of the sample mean, i.e.
Thus the natural estimator of the gradient is
\(x_g\) = [2,0,0], q_std = 0.1, u_std = 0.1, h=0.01
\(x_g\) = [2,0,0], q_std = 0.1, u_std = 0.1, h=0.01
\(x_g\) = [-1.5,-1.5,1], q_std = 0.1, u_std = 0.1, h=0.01
\(x_g\) = [-1.5,-1.5,1], q_std = 0.1, u_std = 0.1, h=0.01
\(x_g\) = [-1.5,-1.5,1], q_std = 0.1, u_std = 0.1, h=0.01
\(x_g\) = [2,0,0], q_std = 0.1, h=0.01
u_std = 0.1
u_std = 0.3
u_std = 0.5
\(x_g\) = [2,0,0], u_std = 0.3, h=0.01
q_std = 0.05
q_std = 0.1
q_std = 0.3
\(x_g\) = [2,0,0], u_std = 0.3, h=0.05
q_std = 0.05
q_std = 0.1
q_std = 0.3
\(x_g\) = [2,0,0]
Log barrier weight = 1000
(0.001N at 1m)
Log barrier weight = 100
(0.01N at 1m)
Log barrier weight = 10
(0.1N at 1m)
\(x_g\) = [0.3,0,0]
Log barrier weight = 1000
(0.001N at 1m)
Log barrier weight = 100
(0.01N at 1m)
Log barrier weight = 10
(0.1N at 1m)
\(x_g\) = [-1.5,-1.5,1]
Log barrier weight = 1000
(0.001N at 1m)
Log barrier weight = 100
(0.01N at 1m)
Log barrier weight = 10
(0.1N at 1m)
Full state dynamics
Just next object state
Just next object state, but analytically smoothed dynamics
We want to calculate
Analytically smoothed dynamics that takes a relative position command \(u_r\) and only returns \(x\)
Absolute position command as a function of robot position and relative position command
Weighted L2 Norm Cost
\(x_g\) = [-1.5,-1.5,1], log barrier weight = 100, h = 0.01
\(x_g\) = [-1.5,-1.5,1], log barrier weight = 100, h = 0.01
\(x_g\) = [-1.5,-1.5,1], log barrier weight = 100, h = 0.01
\(x_g\) = [-1.5,-1.5,1]
Log barrier weight = 1000
(0.001N at 1m)
Log barrier weight = 100
(0.01N at 1m)
Log barrier weight = 10
(0.1N at 1m)
\(x_g\) = [2, 0, 0]
Log barrier weight = 1000
(0.001N at 1m)
Log barrier weight = 100
(0.01N at 1m)
Log barrier weight = 10
(0.1N at 1m)
Log barrier weight = 1000
(0.001N at 1m)
Log barrier weight = 100
(0.01N at 1m)
Log barrier weight = 10
(0.1N at 1m)
\(x_g\) = [0.3, 0, 0]
Log barrier weight = 1000
(0.001N at 1m)
Log barrier weight = 100
(0.01N at 1m)
Log barrier weight = 10
(0.1N at 1m)