this was a joke but this is also true what deepmind here took very little compute compared to what is commonly being done with text and from what I can tell nothing new on the architecture side all these tools were around, it "just" took good engineeringhttps://twitter.com/alth0u/status/1333498820508807169 …
-
-
Replying to @alth0u
yup, repurpose self attention transformer to do graph learning and profit none of these are new methods
1 reply 0 retweets 2 likes -
all of the low hanging fruit left in applied ML is safely hanging in the trees behind a shield of difficult engineering, data set gen, and tuning self supervision
1 reply 0 retweets 4 likes -
Replying to @WillManidis @alth0u
haha maybe so if you prune down the space to "applied ML" but i think there are lots of theory advances left in Deep RL, novel architectures etc some of these will become clearer on the next compounding of compute power
1 reply 0 retweets 1 like -
i continue to believe that the interplay between cognitive neuroscience and deep learning will be one of the great flywheels of 21st century science and i'm not sure that we're close to the end
1 reply 0 retweets 3 likes -
agree — just a literal generation of applied results have been within striking distance for a generation if any ml eng could also pipeline something
1 reply 0 retweets 2 likes -
Replying to @WillManidis @alth0u
agreed tbh i got a whole neurips paper out of this insight and doing some engineering that no one else would do
2 replies 0 retweets 1 like
stop doxxing yourself. there's only like 4 neurips papers a year that are predicated on good eng
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.