I know each person and organization and they're all well intentioned. Some are getting aggravated at each other but I believe that's as they all see this as important and are reacting to a live-fire exercise split across the already contentious areas of {ML, PR, AI hype, ...}.
-
-
Show this thread
-
The work that caused the kerfuffle was a large scale language model from
@OpenAI. Think of it as a super powered version of the predictive text on your phone that has read more data and can generate fairly coherent text.https://twitter.com/Smerity/status/1096189350268686337 …Show this thread -
The
@OpenAI team released the paper and code for training the LM but not the largest of the pretrained models. The reasoning for not releasing the underlying model was concerns around dual use (dual use tl;dr good tech that can easily be used for bad)https://twitter.com/jackclarkSF/status/1096214382772514816 …
Show this thread -
Whilst
@OpenAI have no requirement to open source the code or the pretrained models many in the community are troubled. Some are focused on reproducibility of science, others on dual use, others on the PR narrative, ... As there's no consensus anywhere there's no agreement.Show this thread -
The code for this work was released today and the technology hasn't changed substantially since last year. Larger groups are likely able to replicate the results - or at least an appreciably similar result - quickly. This impacts reproducibility for many smaller groups though.
Show this thread -
No release aligns with
@OpenAI's charter but leaves many in the community unclear as to considered risks. The ML community can't necessarily discuss the threat model as in the worst case you may be saying "Oh, that house has a key under the doormat".https://twitter.com/catherineols/status/1096211302249201664 …Show this thread -
Releasing a "restricted" model in this way has other (intentional or not) consequences - primarily that "AI is scary" is cat nip to reporters. Hence by acting like a good dual use citizen you can accidentally provoke the AI hype beast.https://twitter.com/mark_riedl/status/1096129834927964160 …
Show this thread -
This also runs in to the issue that research from big labs often ends up Press Release'd rather than released to the research community. As noted, few in the community knew it was coming out. I was a researcher that journalists reached out to early but...https://twitter.com/mark_riedl/status/1096214448807858176 …
Show this thread -
...even if a journalist has good underlying sources, they're in a zero sum game. Multiple journalists are publishing on the research and if they cat nip their story with "AI is scary" they can likely win more audience. This hype escalation is not pretty
https://twitter.com/Smerity/status/892248098692399104 …
Show this thread -
This isn't the first time that
@OpenAI ended up in this predicament. Their 1v1 DotA bot caused a great deal of angst in the ML community as it was released with nearly no details and a grand PR reception. This time they're explicitly better but the modern era has new issues.Show this thread -
As everyone has a different point of view, it's just collisions everywhere :S - When does a model go from "safe" to "dual use"? - How much of a "dual use" delay do we need to add? - Should we release to journalists first or researchers? - How can small labs participate in PR?
Show this thread -
None of these questions have black and white answers and we don't even know what the shades of gray look like.
Show this thread
End of conversation
New conversation -
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.
in SF.