It seems to me this is based on a bad model of how scientific progress gets made.
-
Show this thread
-
You can see this in part by looking at historically important discoveries: what do backpropagation, conv nets, Alexnet, GANS, LSTMs, ReLUs all have in common? All were developed with relatively small compute, and small data sets.
3 replies 19 retweets 94 likesShow this thread -
In science, the most important progress often comes from better questions and better ideas, not better equipment (in this case, more computational power and data).
6 replies 74 retweets 230 likesShow this thread -
In biology, the government-run genome project cost 10 times as much as Venter's private project. Much of the reason for the cost difference is that Venter adopted a clever hack (pairwise end shotgun sequencing) the government project didn't use until the end.
4 replies 3 retweets 46 likesShow this thread -
In particle physics, Freeman Dyson found that, contrary to conventional wisdom, only a small fraction of the most important progress comes from building bigger accelerators. Much of it comes from much harder-to-control improvements in detectors and the like.pic.twitter.com/J2ejGimM0t
4 replies 8 retweets 82 likesShow this thread -
Why is big science so seductive? In part, because it seems guaranteed: you can plan, you can see success from the start. That's much less nerve-inducing (and _seems_ less uncertain) than needing to have clever ideas along the way.
2 replies 18 retweets 78 likesShow this thread -
Back to neural nets: a danger in scaling up your computational power is that you start to focus _only_ on questions that require that computational power. You hire specialists who thrive in that environment, but who aren't so good at playing with basic, fundamental questions...
2 replies 18 retweets 112 likesShow this thread -
... and your culture starts to tilt that way, driving out people who do like to play with basic, fundamental questions.
1 reply 2 retweets 58 likesShow this thread -
Take all this with a grain of salt. Neural nets are a side interest, not my main interest. Maybe I'm wrong. But I don't think so. This dynamic has played out in genome sequencing, in particle physics, & in many other areas. Big science is attractive, but often small science wins
7 replies 16 retweets 83 likesShow this thread -
Replying to @michael_nielsen
https://www.reddit.com/r/MachineLearning/comments/2lmo0l/ama_geoffrey_hinton/clw9pj6/ … Geoff Hinton seems to agree with youpic.twitter.com/DCEwGaSS7x
1 reply 0 retweets 2 likes
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.