GQ1. DL is part of the solution GQ2. "vectors, not symbols” is false dichotomy diff functions: yes, in part Operations a la logic, we do need (contra your view) GQ3. Outputs of deep learning may serve as input to reasoning; symbolic techniques needed for some inferences.
-
-
Replying to @GaryMarcus
Most humans don't actually do much that resemble your answer to GQ3, except a small number of humans using pen and paper, and only in the last couple of millennia. Right now, we need to get machines to the level of a house cat. Never mind symbolic mathematics and formal logic.
4 replies 1 retweet 23 likes -
Replying to @ylecun
I don’t argue that people can’t use formalisms that involve derivatives just because most people can’t explicitly explain what a derivative is. I think you are confusing formal, conscious use of a certain kind of machinery with what brain does unconsciously.
1 reply 0 retweets 8 likes -
Replying to @GaryMarcus
If you see the reasoning engine as *separate* (and qualitatively different) from the deep learning system that provides it with inputs, then we disagree. Unless this reasoning "system" is a pen and paper to do math/logic. And much of human intelligence functions without it.
1 reply 2 retweets 7 likes -
Replying to @ylecun
That may distill a second point of disagreement; I see nothing wrong with (for some purposes) having separate systems for (eg) image classification vs reasoning. Certainly it is possible in principle to engineer systems that way; what’s your objection? Efficiencies of training?
2 replies 0 retweets 4 likes -
Replying to @GaryMarcus @ylecun
I wonder if we are mixing different methodological questions here. If I need to build a robust and reliable system today, I would mix deep learning, probabilistic programming, and symbolic methods. 1/
1 reply 6 retweets 27 likes -
But at the same time, I think it is critically important to see if we can push DL methods to provide a unified solution to perception, reasoning, and action (with robustness and safety). 2/
5 replies 4 retweets 31 likes -
Deep learning, in principle, provides grounding for the semantics of internal representations that current symbolic methods lack. Other learning approaches might do this too. 3/
3 replies 0 retweets 8 likes -
Replying to @tdietterich @ylecun
But not for compositionality or causal reasoning in any obvious way (cc
@yudapearl)2 replies 0 retweets 5 likes -
I wish I could join this discussion but I can't parse Tom's "DL provides grounding for the semantics of internal representations that current symbolic methods lack." In my simple world, it is symbolic representation that provides semantics, not DL. What am I missing?
6 replies 3 retweets 32 likes
@yudapearl makes a really good point here, but crucially i think he is referring to what I would call sentence-level semantics; the stuff @tdietterich is talking about is at the level of individual symbols. deep learning might help there, but not with what Judea rightfully seeks.
-
-
Replying to @GaryMarcus @yudapearl and
"Language production must rely on careful planning of meaning and this meaning needs to be the basis of sentence formulation as opposed to retrieving from memory sentence fragments that have tended to occur in similar circumstances." https://csl.sony.fr/wp-content/themes/sony/uploads/pdf/steels-12c.pdf …
0 replies 0 retweets 1 likeThanks. Twitter will use this to make your timeline better. UndoUndo
-
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.