Some reflections on the learning-to-control renaissance.https://www.argmin.net/2020/06/29/tour-revisited/ …
-
-
Replying to @beenwrekt
Great post. There is also adaptive control of partially observable LQR, i.e. LQG, in
@beenwrekt’s work sqrt T regret is proven. Our recent work shows logarithmic regret via new closed-loop system id method and online learning https://bit.ly/2ZKTnGX@SahinLale@kazizzad@Caltech1 reply 2 retweets 17 likes
Replying to @AnimaAnandkumar @beenwrekt and
Continuous exploration is achieved by noise in output observations, not present in LQR + continuous model refinement achieved by closed-loop system-id allows faster rates in LQG. In regret sense, adaptive control of LQG easier than LQR. https://arxiv.org/abs/2003.11227
3:19 PM - 4 Jul 2020
0 replies
0 retweets
2 likes
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.