sample from @tdietterich:pic.twitter.com/f1FIkuFiEP
You can add location information to your Tweets, such as your city or precise location, from the web and via third-party applications. You always have the option to delete your Tweet location history. Learn more
another example, in the very title of the paper that introduced BERT. not ealso discussion of QA and language inference, not just prediction by word, @tdietterichpic.twitter.com/zVP6PNcozn
And anyone remember what the U in the GLUE benchmark stands for? [Generalized Language Understanding Evaluation]
It’s not all or nothing. There are just degrees of understanding. These large language models do have some surprising level of understanding: they can figure out entity types or do phrase translations. But they also lack common sense and make silly coherence mistakes.
Which makes the concept of “understanding” not very interesting and overused. Better to refer back to tasks and benchmarks and not extrapolate from that as @tdietterich suggested in his recent medium piece.
If that's a direct quote it seems wrong to me, but I also think the power of prediction is only starting to reveal itself (as models get more stuff to predict about). I think it's remarkable the knowledge that prediction can impart.
That is a direct quote (in screenshot). IMO, Ilya is absolutely correct in that point. However, when Gary points out current language models can't do this, people backtrack to such an extent that they pretend no one ever claimed language models are even related to understanding.pic.twitter.com/DdlpoW4FrM
This speaks to a level of general befuddlement that even leading AI/ML folks have gotten into about what current deep learning-based algorithms are actually doing. Of course, it's OK for experts to be confused, but the press keeps amplifying this as if DL is somehow intelligent.
To me the hype surrounding the current wave of #AI is in it self a phenomenon a lot more interesting than any of the actual developments in the field. Frequent antropomorphisations of these relatively primitive statistical engines and media buying it all is simply perplexing.
This is probably due to the fact that ML researchers have to ridiculously oversell their papers if they are to get published, let alone read and cited. The field is just so saturated that people have to pass off incremental improvements as if they're worthy of a Nobel prize.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.