Jun 19, 2025
Adam Wei
Recall: Scaling up sim data improves performance up until a plateau...
Question: How much more real data would you need to collect to match the cotraining plateau?
cotraining plateau \(\approx\) 2-3x more deal data
Recall that:
Questions:
1. Can/should we align the sim and real representations?
2. Can we improve performance by reducing the sim2real gap at the representation level?
Denoising objective
Encourage aligned representations
Denoising objective
Encourage aligned representations
Goal: Learn a representation for sim and real that cannot be discerned by a classifier \(d_\phi\)
Issues: Adversarial training is unstable and challenging...
Given samples \(x_i \sim p^S_\theta\) and \(y_i \sim p^R_\theta\):
Ex: \(k(x, y)=e^{-\frac{\lVert x-y\rVert_2^2}{\sigma^2}}\)
Adding an alignment objective does not improve performance and adds complexity
Learning from "clean" and "corrupt" data
Won't be going over the details of the algorithm...
... only the high-level ideas
\(t=0\)
\(t=T\)
\(t=t_n\)
\(\mathbb E[\lVert h_\theta(x_t, t) - x_0 \rVert_2^2]\)
\(\mathbb E[\lVert \frac{\sigma_t^2 - \sigma_{t_n}^2}{\sigma_t^2} h_\theta(x_t,t) + \frac{\sigma_{t_n}^2}{\sigma_t^2}x_t - x_{t_n} \rVert_2^2]\)
Use clean data to learn denoisers for \(t\in(0,T]\)
Use corrupt data to learn denoisers for \(t\in(t_n,T]\)
Current results are worse than mixing the data
Intuition:
GCS
(clean)
RRT
(clean)
Task: Cotrain on GCS and RRT data
Goal: Sample clean and smooth GCS plans