...But they have existed for a very long time: since the early 90s for Q1 and since the early 80s for Q2. Now that the DL machinery works, and that so many people are working on both Qs, we have a shot at making real progress.
-
-
Replying to @ylecun @GaryMarcus
I guess the remaining question for your position are: GQ1. Will DL be part of the solution (you said yes) GQ2. Do you agree with "vectors, not symbols; diff functions, not hard logic" GQ3. If not, how do you propose we make reasoning compatible with DL?
3 replies 3 retweets 23 likes -
Replying to @ylecun
GQ1. DL is part of the solution GQ2. "vectors, not symbols” is false dichotomy diff functions: yes, in part Operations a la logic, we do need (contra your view) GQ3. Outputs of deep learning may serve as input to reasoning; symbolic techniques needed for some inferences.
3 replies 2 retweets 14 likes -
Replying to @GaryMarcus
Most humans don't actually do much that resemble your answer to GQ3, except a small number of humans using pen and paper, and only in the last couple of millennia. Right now, we need to get machines to the level of a house cat. Never mind symbolic mathematics and formal logic.
4 replies 1 retweet 23 likes -
Replying to @ylecun
The entire field of generative linguistics would beg to differ. The standard presumption there is that our comprehension of language revolves around manipulations of strings of structured symbols.
1 reply 1 retweet 8 likes -
Replying to @GaryMarcus
Yes, and that's precisely what's wrong with it. That view has been a complete and utter failure in NLP. That was the main point of Hinton's remarks. Noam Chomsky's birthday was yesterday. Geoff Hinton's was the day before yesterday.
2 replies 4 retweets 31 likes -
Replying to @ylecun
No approach has been a resounding success in NLU; that’s why I keep calling for hybrid vigor. Anyone who thinks we don’t need compositionality for language comprehension isn’t paying attention.
3 replies 0 retweets 8 likes -
Replying to @GaryMarcus @ylecun
The symbolic representations of formal logic and generative linguistics exhibit compositionality, but compositionality does not need them, as complex body plan assembly in evo-devo shows
1 reply 0 retweets 2 likes -
-
Replying to @GaryMarcus @ylecun
Body plans are composed by repeated and hierarchical activation of developmental processes. No less compositional than sentence structure. The protein/RNA mechanisms that achieve this are all based on graded molecular dynamics, not on symbol processing.
1 reply 1 retweet 6 likes
what makes them not symbolic processing? pax6 for example can serve as a signal inducing a whole cascade of genes. lac operon basically serves as an if-then instruction. looks like a stochastic or probabilistic grammar/cellular automata, evolved rather than programmed.
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.