Can differentiable programming learn arbitrarily complex behaviors, up to superhuman performance, given a dense sampling of the behavior manifold (e.g. an infinite training data generator, such as self-play)? Yes. We knew that. Arguably AlphaGo was the concrete proof-of-concept.
-
-
Note that the question of model choice (the structure of the differentiable architecture used) is rather secondary, because as long as memorization capacity is sufficient, *any* model will do -- if there is training data to match. Even a single hidden Dense layer.
Show this thread -
It just moves the threshold for "have we drawn enough data yet" by a bit -- by no more than a few orders of magnitude. But when infinite data is available, this is "just" the size of your cloud computing bill.
Show this thread
End of conversation
New conversation -
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.