In this light, the "intelligence" of the network comes purely from its training data. The network is sample-inefficient and only performs local generalization. The next frontier is abstraction & reasoning, which will enable extreme generalization and decent sample efficiency.
-
-
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo
-
-
-
O boy! I agree but what a way to deflate hype. I remember pushing fast generalized classifications using hash tables in my PhD... That was 30 years ago...
-
Yup, all the ideas are old. The clothing is new :)
End of conversation
New conversation -
-
-
Is there any research/litterature on this interpretation?
- Show replies
New conversation -
-
-
@fchollet nice interpretation.ThanksThanks. Twitter will use this to make your timeline better. UndoUndo
-
-
-
Yes!!!! So glad to see this phrasing. There was a ams article talking about the math of NN’s that had like one sentence of this and I spent like a week dreaming about the interpretation.
Thanks. Twitter will use this to make your timeline better. UndoUndo
-
-
-
Think of it this way too. A dictionary w. a similarity hash. This is easiest to see with any NN after an argmax on softmax layer (ie sends to a bucket). For scalar valued outputs,it's function approximation. More points give more coverage for better interpolation/bucket selection
-
Autoencoders and GANs don't quite fit in this story though. Autoencoders, lossy compressors. GANs closer tho. GANs apply transformations, but better thought of in terms of ODEs they define and buckets become basins.
- Show replies
New conversation -
-
-
Which in a sense why they can “abstract” in a sense, but can’t learn. The learning is in the original paring of inputs and targets.
Thanks. Twitter will use this to make your timeline better. UndoUndo
-
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.