serious question - can any GPT-2 experts answer? has anyone studied this?https://twitter.com/bradpwyble/status/1188155924902502400 …
-
-
This should be a required baseline in testing any language model that purports to generate meaningful text.
-
See also: https://old.reddit.com/r/MachineLearning/comments/dfky70/discussion_exfiltrating_copyright_notices_news/ … IMO problem is not memorization--when you get down to it, generation is not that useful--you need facts. The problem is the amount of required data is anti selective curation--it motivates a stance of vacuum everything to blindly feed the model
End of conversation
New conversation -
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.