Deep Generative models (VAEs) for ODEs with knows and unknowns
Harshavardhan Kamarthi
Data Seminar
Variational Auto-Encoder

- Goal: Learn distribution p(X)
- Encode x towards a known distribution on latent variable z. Decode z to x
Variational Auto-Encoder

Encoder
Prior on latent variable . Usually set as standard Gaussian.
Decoder
Variational ELBO Loss
Physics based VAEs
- Model data from complex ODEs
- Learn low-dimensional latent representations from data
- Learn relation between observed state and latent state where latent states are governed by ODE
- Learn parameters of the latent ODE
ODE2VAE: Deep generative second order ODEs with Bayesian neural networks

Yidiz et al (NIPS '19)
ODE2VAE: Deep generative second order ODEs with Bayesian neural networks
No information about the functional form
Goals:
- Learn low lank representations from sequential data
- Extrapolate/forecast
Second-order Bayesian ODE

- Second-order: to model higher-order dynamics
- is a Bayesian Neural network with prior
- To capture uncertainty

ODE2VAE Model

ODE2VAE Encoder

Note:
- Transition function is stochastic
- 3rd equation derived by "Change of variable" theorem (approximated by Runge-Kutta method)
ODE2VAE Decoder

is deterministic NN decoder
ODE2VAE Learning


ODE2VAE Learning


- Dynamic loss dominates VAE loss due to long trajectories
- Data may underfit to encoder distribution
- Add regularizer to make ode and encoder distribution similar
Extrapolation Results

CMU-Walking data (Extrapolate last 1/3 rd trajectory)
Rotating MNIST


Bouncing Ball
Generative ODE Modelling with Known and Unknowns

Linial et al (CHIL '21)
Generative ODE Modelling with Known and Unknowns
Given the functional form:
Estimate
1. Parameters
2. Extrapolate
Encoder

Use RNNs to encode latent variables and then transform to "physics grounded form"
Decoder

Transition probabilities on latent space is deterministic:
defined from
Transformation to observation space is deterministic via neural network:
defined from
Learning
Generative Distribution
Minimize ELBO
Results


Similarly outperformed baselines significantly for double pendulum a CVS model


What if part of functional is unknown?
Unknown
Then add NN to learn the unknown:
What if part of functional is unknown?


Physics-Integrated Variational Autoencoders for Robust and Interpretable Generative Modeling
- Functional is partially known

- known part of functional
- Model and extrapolate
Takeishi et al '21
Encoder

- are neural networks, are also trainable parameters

- are defined using domain information
Divide latent encoding into known and unknown parts
Decoder
- is neural network
- is from solution of ODE on
Focusing on Physics-based component
- can dominate training ignoring
- Define a Physics-only reduced model
- Where , the baseline function is another NN with trainable parameter
- Then add regularizer


Physics-based data augmetation
- Goal: ground outputs of which are parameters of
- Idea: Use physics model as data augmentation to train
- Sample from
- Generate
- Train to estimate from
Results


Physice ODE
By Harshavardhan Kamarthi
Physice ODE
- 135