Central planning is unbeatable as a resource allocation mechanism for small systems. As you attempt to scale it up, it becomes terribly ineffective, to such an extent that decentralized control algorithms wastly outperform it, despite being inherently wasteful
-
Show this thread
-
You don't need to look at country-scale economies to observe this effect (obviously free markets >> central planning), it applies even to large companies. Past a certain scale you need teams that compete against each other with overlapping products. But why?
5 replies 9 retweets 75 likesShow this thread -
I think it's a technology problem, not an intrinsic issue. Better cybernetics will enable us to scale efficient central control (while taking into account uncertainty and exploration) to increasingly large systems in the future
9 replies 9 retweets 47 likesShow this thread -
This is tightly related to AI as well: it's the question of local vs. global optimization in a non-differentiable system, open-ended goal-setting, objective propagation from one module to its neighbors, parent, and internal submodules...
1 reply 5 retweets 65 likesShow this thread -
Backprop is a centralized answer, that only applies to differentiable functions where "control" means adjusting some parameters and where the optimization objective is already known. Its range of applicability is minuscule
6 replies 10 retweets 85 likesShow this thread -
Replying to @fchollet
The objective function isn't always known e.g. GANs. RL problems also involve objectives that can only be sampled from. I get the local vs global argument, but it's unclear what you mean about backprop being limited by objectives.
1 reply 0 retweets 0 likes -
Replying to @Zergylord
"known" objective doesn't mean "can be expressed analytically as a func of the data", it means you know what you're optimizing for from the start. Your GAN implementation itself is an expression of a known optimization objective. "Unknown objective" refers to open-ended problems.
2 replies 0 retweets 4 likes -
Replying to @fchollet
So that would imply this isn't a backprop problem, but rather a problem with optimization as a field? Could you give an example of an "open ended problem"
1 reply 0 retweets 0 likes -
Replying to @Zergylord
With GANs, you know what the end product is going to be: a model that has learned the latent distribution of your data and that you can sample from. You can train multiple GANs and get approximately the same results. So you know B and you know how to get from A to B
1 reply 0 retweets 0 likes
In most real world problems, you don't know B and don't know how to get there, e.g. "what product should I build next?". The space of possible answers is infinite and wildly diverse. Humans are inherently open-ended problem solvers: they set their own goals
-
-
Replying to @fchollet
To the extent historical outcomes allow humans to make these decisions, it's just a very hard RL problem. The non stationarity would surely make existing approaches fail, but noone is saying that the field is solved. Again, this doesn't sound like a problem with backprop per se.
0 replies 0 retweets 1 likeThanks. Twitter will use this to make your timeline better. UndoUndo
-
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.