Two recent algorithms, World Models by @hardmaru and Schmidhuber (2018) and Curiosity by @pathak2206 et al. (2018), have approx. equal performance when a learned module (RNN and embedding in ICM, respectively) is instead left as a fixed randomly initialized module.
@_brohrer_
-
-
Yes, thanks! Another friend also referred us to the Johnson-Lindenstrauss lemma in random projections. Would your expectation be that projection on a large family of random filters would work as well as a random fixed CNN, both with the same output dimension?
-
I'd expect the random CNN to work better. Would be an interesting experiment.
End of conversation
New conversation -
-
-
I was going to add a tweet but it didn't go through. What's interesting here is not that a random network would work, it's that training this random network does not improve its performance. It's a sign there's a fundamental flaw with how we think about the setup.
-
If random projections work at all, it should be very easy to tune these random projections to get a better network. No reason to believe the random weights were a local minimum.
- Show replies
New conversation -
-
-
orthogonal projection makes prefer sense. You may look for directions close to whatever, possible, rotation and reflection.
@abursuc the paper that you recently posted.Thanks. Twitter will use this to make your timeline better. UndoUndo
-
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.