Two recent algorithms, World Models by @hardmaru and Schmidhuber (2018) and Curiosity by @pathak2206 et al. (2018), have approx. equal performance when a learned module (RNN and embedding in ICM, respectively) is instead left as a fixed randomly initialized module.
@_brohrer_
-
Show this thread
-
Replying to @LiamFedus @hardmaru and
Random weights have often been a surprisingly strong baseline. Back in 2007-2011 we often saw convolutional nets with random weights getting state of the art performance on object recognition tasks like CalTech-101 and CIFAR-10
11 replies 25 retweets 160 likes -
Replying to @goodfellow_ian @LiamFedus and
Feature extraction via projection on a large family of random filters. Works especially well if your weights have an orthogonal initialization (so you don't lose to much information in the process).
3 replies 2 retweets 43 likes -
Replying to @fchollet @goodfellow_ian and
I was going to add a tweet but it didn't go through. What's interesting here is not that a random network would work, it's that training this random network does not improve its performance. It's a sign there's a fundamental flaw with how we think about the setup.
1 reply 0 retweets 15 likes -
Replying to @fchollet @goodfellow_ian and
If random projections work at all, it should be very easy to tune these random projections to get a better network. No reason to believe the random weights were a local minimum.
1 reply 1 retweet 15 likes -
Replying to @fchollet @goodfellow_ian and
Here they remove the last FC layer of many conv networks architectures and substitute it w/ a random projection and achieve similar results to the FC onehttps://arxiv.org/abs/1801.04540
1 reply 0 retweets 0 likes
If you remove a trained layer, add a random one, and then train a new layer on top, what you're looking at is just whether the random layer is destructive or not (with enough units, it's likely overcomplete, so not destructive). This is very different and not surprising
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.