It may be fruitful & fashionable to say we dont understand anything about the brain & so cant model it. But that's simply not how modeling works. Its iterative and you go with your best guess until contradictory evidence provides you with a better one.
-
-
But the problem Grace is that the advocates of ANNs do not take your position: They do not want to iterate to a better (biologically) model, but instead double down on this particular abstraction at this particular point in time.
1 reply 0 retweets 3 likes -
With all due respect, I don't know who these practitioners you speak of are. Grace's position is the mainstream, in my opinion.
3 replies 1 retweet 22 likes -
Replying to @tyrell_turing @tarinziyaee and
Think this might also just be a diff of positions due to diff end goals? Those who might view AI as way to understand our brain & intel & leverage it have Graces position. Versus those who see AI as a blanket term wrt to data science+ML & focused on solns to specific problems.
1 reply 0 retweets 1 like -
Replying to @GaneshNatesh @tyrell_turing and
Certainly there are people who use & design ANNs w/o any interest in interating on models of the brain. And that's fine. But that doesnt mean their artificial neurons arent abstractions of real neurons (unless they really take them far off in some other direction)
2 replies 0 retweets 5 likes -
Replying to @neurograce @GaneshNatesh and
And people who do want to iterate and make better brain models are on solid enough ground when they start with ANNs
1 reply 0 retweets 2 likes -
Replying to @neurograce @GaneshNatesh and
I think I understand where you're coming from. I'm less concerned with the model of neurons in play than the logic implemented in the 'circuits' in use today. The only reason I argue about this topic is that the abstractions in use don't appear to be converging on the subject.
1 reply 0 retweets 1 like -
Replying to @chophshiy @GaneshNatesh and
The success of CNNs in predicting the response of real visual neurons (and similar findings for audition) suggest they are going in the right direction wrt getting the logic right. But yea just cuz you hook some artificial neurons up doesnt mean they'll always work like the brain
1 reply 0 retweets 7 likes -
Replying to @neurograce @chophshiy and
agreed with
@neurograce, e only additional point I am making here is that we ought have some humility here, 33 years after White et al's 1986 landmark worm connectome, which in honesty we must recognize we have not yet fully understood. read it, and weep: https://royalsocietypublishing.org/doi/abs/10.1098/rstb.1986.0056 …1 reply 0 retweets 2 likes -
Replying to @GaryMarcus @neurograce and
@neurograce has written this excellent article on the DNNs + visual system comparisons:https://neurdiness.wordpress.com/2018/05/17/deep-convolutional-neural-networks-as-models-of-the-visual-system-qa/ …2 replies 3 retweets 10 likes
this is a good QA but the query about what visual system has that CNNs lack also include compositionality, ability to recognize line drawings, silhouettes, and other abstractions even without explicit training data, ability to infer relationships between objects, etc
-
-
Replying to @GaryMarcus @kateiyas and
Sure it could include those too, which people are also working on. I wrote it before the bulk of the work, eg, showing that CNNs respond more to texture than shape, but that's something people have already found ways of countering.
2 replies 0 retweets 3 likes -
Replying to @neurograce @GaryMarcus and
And to be clear, a standard AlexNet is not *catastrophic* on silhouettes or line drawings, it isnt as good as humans (chance here is 1/16) and this paper introduces training that can help with that https://arxiv.org/abs/1811.12231 pic.twitter.com/zRHHTiKuYX
3 replies 0 retweets 9 likes - 2 more replies
New conversation -
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.