A theory of representation learning in deep neural networks gives a deep generalisation of kernel methods
[article]
Adam X. Yang, Maxime Robeyns, Edward Milsom, Nandi Schoots, Laurence Aitchison
2022
arXiv
pre-print
We therefore develop a new infinite width limit, the representation learning limit, that exhibits representation learning mirroring that in finite-width networks, yet at the same time, remains extremely ...
Finally, we use this limit and objective to develop a flexible, deep generalisation of kernel methods, that we call deep kernel machines (DKMs). ...
This contrasts with a deep linear neural network, which has infinitely many optimal settings for the weights. Note that for the objective to be well-defined, we need K(G) to be full-rank. ...
arXiv:2108.13097v4
fatcat:jtnn3ftdifgqnfyo6em4l2ckze