Daniel Yukimura
Result: Kernels derived from DNNs have the same approximation properties than shallow networks.
Consequence: The Kernel framework doesn't seem to explain the benefits of deep architectures.
Random feature kernels:
orth. for the Gaussian meas.
Step func.
ReLU
Neural Tangent Kernels:
Spherical harmonics and description of the RKHS
Consequences for ReLU networks:
What else?
some DNNs might have lazy behavior, but the good ones are not on this regime.
What else?
A learner learns a complicated target by decomposing it into a sequence of simpler functions.