Can differentiable programming learn arbitrarily complex behaviors, up to superhuman performance, given a dense sampling of the behavior manifold (e.g. an infinite training data generator, such as self-play)? Yes. We knew that. Arguably AlphaGo was the concrete proof-of-concept.
-
Show this thread
-
Thus, the question of failure or success boils down to "have we drawn enough training data yet?" (or in some cases, "does our architecture have sufficient capacity & appropriate structure wrt the training data available?"). Whether the answer is yes or no, it teaches us nothing.
1 reply 0 retweets 13 likesShow this thread -
Note that the question of model choice (the structure of the differentiable architecture used) is rather secondary, because as long as memorization capacity is sufficient, *any* model will do -- if there is training data to match. Even a single hidden Dense layer.
2 replies 0 retweets 19 likesShow this thread -
It just moves the threshold for "have we drawn enough data yet" by a bit -- by no more than a few orders of magnitude. But when infinite data is available, this is "just" the size of your cloud computing bill.
1 reply 1 retweet 16 likesShow this thread -
This Tweet is unavailable.
Yes, that would be new and significant, but also almost certainly not doable with differentiable programming
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.