A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2022; you can also visit the original URL.
The file type is application/pdf
.
On Random Weights and Unsupervised Feature Learning
2011
International Conference on Machine Learning
Recently two anomalous results in the literature have shown that certain feature learning architectures can yield useful features for object recognition tasks even with untrained, random weights. In this paper we pose the question: why do random weights sometimes do so well? Our answer is that certain convolutional pooling architectures can be inherently frequency selective and translation invariant, even with random weights. Based on this we demonstrate the viability of extremely fast
dblp:conf/icml/SaxeKCBSN11
fatcat:pk47gvvf25drtcmf7g3ddpsozy