you are kidding right?
-
-
Replying to @filippie509 @MaxALittle and
@abhijitysharma i would urge you to try this out and to read carefully Alcorn et als new arXiv paper, to understand the strengths and weaknesses of data augmentation.
1 reply 1 retweet 8 likes -
Replying to @GaryMarcus @filippie509 and
That paper doesn't study data augmentation at all AFAICT. They're using a pre-trained imagenet baseline, which explicitly avoids the kind of data augmentation necessary to recognize these synthetic 3d images in unusual poses.
3 replies 0 retweets 5 likes -
Replying to @jeremyphoward @filippie509 and
I wasn’t saying it did, I was saying that one could learn something by trying to use data augmentation as a candidate solution. Simple tricks like translation won’t work, and even 3-d/6d rotation may not work and may be prohibitively cost/not viable.
1 reply 0 retweets 2 likes -
Replying to @GaryMarcus @filippie509 and
OK. I don't think this paper helps show the "strengths and weaknesses of data augmentation". It's already been seen that data augmentation and convnets can give good pose invariance. You can help it along a bit using stuff like Group Equivariant Convolutional Networks
1 reply 0 retweets 6 likes -
Replying to @jeremyphoward @GaryMarcus and
The Alcorn paper simply shows that you can't expect to use different poses at inference time than you had in your data or used in data augmentation at training time, unless you force appropriate symmetry in your architecture
2 replies 0 retweets 7 likes -
Replying to @jeremyphoward @filippie509 and
Or, put differently, it highlights how fragile deep learning is when tested outside of distribution, and shows how (their word) “naive” DNN’s understanding of objects is.
6 replies 1 retweet 10 likes -
Replying to @GaryMarcus @jeremyphoward and
1/2 Pretty much anything (humans too) is fragile when tested "out of distribution". I'd rephrase it as DNNs have limited generalisation capabilities compared to AGI say (but still amazing compared to what we had before). Not being fussy, just I think it helps framing the problem.
2 replies 0 retweets 0 likes -
Replying to @larosaandrea @GaryMarcus and
2/2 Tbf, the paper seems to use "DNNs" as a proxy for "some specific architectures trained on a certain way on ImageNet and COCO". To me its main message is that those are not robust benchmarks. And then yes, that current DNNs might have limited generalisation capabilities
1 reply 0 retweets 0 likes
yes: limited generalization capacities, and it’s important to understand and acknowledge those limits in order to build better tools
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.