Justine Zeghal
Supervisors: François Lanusse, Alexandre Boucaud, Eric Aubourg
Tri-state Cosmology x machine learning journal club
January 19, Paris, France
How to extract all the information embedded in our data?
Standard analysis relies on 2-point statistics and use a Gaussian analytic likelihood.
But these summary statistics is suboptimal at small scales..
Explicit joint likelihood
And then run an MCMC to get the posterior:
And then run an MCMC to get the posterior:
Explicit joint likelihood
It provides exact results but necessitates the BHM to be differentiable and requires a lot of simulations.
And then run an MCMC to get the posterior:
And then run an MCMC to get the posterior:
Explicit joint likelihood
Simulator
Summary statistics
Simulator
Simulator
And use neural-based likelihood-free approaches to get the posterior
using only:
Summary statistics
And use neural-based likelihood-free approaches to get the posterior
using only:
Simulator
Summary statistics
And use neural-based likelihood-free approaches to get the posterior
using only:
Simulator
Summary statistics
And use neural-based likelihood-free approaches to get the posterior
using only:
Simulator
Summary statistics
Simulator
Summary statistics
Simulator
Summary statistics
Simulator
Summary statistics
Denise Lanzieri, Justine Zeghal
T. Lucas Makinen, Alexandre Boucaud, François Lanusse, and Jean-Luc Starck
It is only a matter of the loss function you use to train your compressor..
We developed a fast and differentiable (JAX) log-normal mass maps simulator
1. We compress using one of the 5 losses.
Benchmark procedure:
2. We compare their extraction power by comparing their posteriors.
For this, we use a neural-based likelihood-free approach, which is fixed for all the compression strategies.
Simulator
Summary statistics
Justine Zeghal
Denise Lanzieri, Alexandre Boucaud, François Lanusse, and Eric Aubourg
Log-normal LSST Y10 like
differentiable
simulator
For our benchmark
With a few simulations it's hard to approximate the posterior distribution.
→ we need more simulations
BUT if we have a few simulations
and the gradients
(also know as the score)
then it's possible to have an idea of the shape of the distribution.
Normalizing flows are trained by minimizing the negative log likelihood:
Normalizing flows are trained by minimizing the negative log likelihood:
But to train the NF, we want to use both simulations and gradients
Normalizing flows are trained by minimizing the negative log likelihood:
But to train the NF, we want to use both simulations and gradients
Normalizing flows are trained by minimizing the negative log likelihood:
But to train the NF, we want to use both simulations and gradients
Normalizing flows are trained by minimizing the negative log likelihood:
Problem: the gradient of current NFs lack expressivity
But to train the NF, we want to use both simulations and gradients
Normalizing flows are trained by minimizing the negative log likelihood:
Problem: the gradient of current NFs lack expressivity
But to train the NF, we want to use both simulations and gradients
Normalizing flows are trained by minimizing the negative log likelihood:
Problem: the gradient of current NFs lack expressivity
But to train the NF, we want to use both simulations and gradients
Normalizing flows are trained by minimizing the negative log likelihood:
Problem: the gradient of current NFs lack expressivity
But to train the NF, we want to use both simulations and gradients
Normalizing flows are trained by minimizing the negative log likelihood:
Problem: the gradient of current NFs lack expressivity
But to train the NF, we want to use both simulations and gradients
→ On our toy Lotka Volterra model, the gradients helps to constrain the distribution shape
(from the simulator)
(requires a lot of additional simulations)
→For this particular problem, the gradients from the simulator are too noisy to help.
Log-normal LSST Y10 like
differentiable
simulator
For our benchmark
Focus on implicit inference methods
simulations
simulations
more than
Simulator
Summary statistics
contact: zeghal@apc.in2p3.fr
slides at: https://slides.com/justinezgh