Deep learning more and more looks like a scientific revolution in the sense of Thomas Kuhn. As exciting as seeing engineering successes, is to experience first hand how a research field goes through a “paradigm shift”.
-
Show this thread
-
E.g., for a long time, the dominant view in AI/cogsci/linguistics was that ‘classical’ neural networks cannot learn to represent rules, variables, hierarchical structure, compositionality. Which was a reason to avoid these models. Here’s Jackendoff (2002) summarizing
@GaryMarcuspic.twitter.com/KLCtfdmjqZ
2 replies 0 retweets 8 likesShow this thread -
Replying to @wzuidema
i stand by these claims. Importantly though the term classical is not mine, and interested readers should read the original source, The Algebraic Mind.
1 reply 1 retweet 1 like -
Replying to @GaryMarcus
I'm happy that you do; for one thing, your work helps to make the point that it is not obvious that neutral networks can learn such things, and this helps explain the relevance of some of our results. But I disagree that NNs are fundamentally unable to learn rules, variables, 1/n
1 reply 0 retweets 1 like -
Replying to @wzuidema @GaryMarcus
tree structure, or compositionality, at least under reasonable definitions. Not reasonable: picking your favorite symbolic system, checking whether a NN learns exactly that, and interpreting failure as evidence that the whole class cannot be learned.
2 replies 0 retweets 0 likes -
Replying to @wzuidema @GaryMarcus
I feel like this raises as many concerns as it addresses... shouldn’t a properly compositional system be able to do quite a large range of tasks? So for any given task, why would an apparently compositional net fail?
1 reply 0 retweets 0 likes -
Replying to @beausievers @GaryMarcus
We looked at a simple arithmetic task with addition, substraction and brackets. Simple, but an infinite domain and clearly compositional. Networks approximate answers within training range almost perfectly, generalize quite well, with errors increasing with length of expressions.
2 replies 0 retweets 1 like -
Replying to @wzuidema @beausievers
Isn’t that pretty much as I anticipated? Cc
@rgalhama1 reply 0 retweets 0 likes -
I read the prediction of the '99 ABA/ABB paper (and Algebraic Mind) as anticipating the failure of non-symbolic-NNs in generalizing outside of the training space (i.e. not being able to account for rules/hierarchy/composition)
2 replies 0 retweets 1 like -
Replying to @rgalhama @GaryMarcus and
While it is certainly not trivial to have non-symbolic-NNs learn solutions that approximate a symbolic system, I think the body of work that Jelle summarizes shows it is not a fundamental limitation of this class of models.
1 reply 0 retweets 0 likes
How are you defining “this class of models” and why isn’t the result above an illustration that extrapolation is hard for them? I don’t follow.
-
-
Raquel G. Alhama Retweeted Raquel G. Alhama
I've posted the reply here https://twitter.com/rgalhama/status/1166008735602487296 … (ironically, it turns out that inducing the tree structure of a tweet thread is a hard task for my human neural net
)Raquel G. Alhama added,
0 replies 0 retweets 1 likeThanks. Twitter will use this to make your timeline better. UndoUndo
-
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.