2/ I'm a big believer in optimization and intelligent use of limited resources, so I'm in agreement! I have to embrace this really, given my resource limited position. But "Hit the Wall" in context means a lot less than I think this headline conveys.
-
Show this thread
-
3/ As far as this apparent broader lack of civility between deep learning advocates and their discontents: I see a whole lot of arguing over nothing. I think what Jerome Pesenti says in the article is a pretty normal sentiment among most researchers and practitioners:
1 reply 0 retweets 2 likesShow this thread -
4/ "Deep learning and current AI, if you are really honest, has a lot of limitations. We are very very far from human intelligence, and there are some criticisms that are valid: It can propagate human biases, it’s not easy to explain, it doesn't have common sense, it’s more on
1 reply 0 retweets 3 likesShow this thread -
5/ the level of pattern matching than robust semantic understanding. But we’re making progress in addressing some of these, and the field is still progressing pretty fast. You can apply deep learning to mathematics, to understanding proteins,
1 reply 0 retweets 3 likesShow this thread -
6/ there are so many things you can do with it." Yes. Exactly. It's a tool. The vast majority of us using it aren't parading it around calling it one step from AGI. We know better esp after dealing with it the realities of daily trial/error/nudging of hyperparameters, etc.
1 reply 0 retweets 7 likesShow this thread -
7/ Those who are overselling deep learning seem to be those who write books/articles that need views, or do PR, or are trying to sell snake oil. Yes that's bad. But creating drama out of thin air is just a waste of time and emotion (lots of Twitter fights over this lately!).
1 reply 1 retweet 7 likesShow this thread -
Replying to @citnaj
I think you know, things are maybe a little less ideal than how you've painted them. First, on hitting a wall on compute. Many results have depended on a large investment in compute to be achievable: top pretrained LMs, stylegan, RL game bots etc. Unless something changes, andpic.twitter.com/CoHcIVwlUE
3 replies 0 retweets 0 likes -
Replying to @sir_deenicus
The something I expect to change is a focus on using the same hardware more effectively (optimization!). I think there's a lot of reasons to believe that there's a lot to be gained here: 1. Historically, software efficiency gains often dwarf anything you can get out of hardware
1 reply 0 retweets 2 likes -
Replying to @citnaj @sir_deenicus
example- EfficientNet came out recently and showed just how inefficient previous vision networks were (and there's a lot of attention being paid to these!) We're talking order of magnitude here. And it does work (I use it!)pic.twitter.com/SOJ9iOJtWO
1 reply 0 retweets 3 likes -
Replying to @citnaj @sir_deenicus
In my own work, I've been able to shrink down DeOldify 10x and get better performance with this mindset. There were simply just bad decisions made at the software level previously (mostly stuff I grabbed from elsewhere that was considered pretty much SOTA!).
1 reply 0 retweets 2 likes
2. The whole specialized hardware route really still seems to be under explored. We're still using general purpose GPUs.
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.