Summary Presentation of:
Hugo Hadfield
Hyper-prior entropy model learns a family of entropy models, side information in z parameterises where in distribution an image is
Context model allows local context of previously decoded latents to help decode later ones, reduces required amount of information to transmit
Entropy of the latent image bitstream under the model given by the quantised entropy model parameters
Entropy of the hyper-latent bitstream
Error in the reconstructed image
rate-distortion tradeoff weight
Key trick is to substitute quantisation for the addition of uniform noise: convolution with a uniform distribution
This allows us to write down
Where \(c\) is the cumulative density of the underlying model, ie. a normal distribution