Tweetovi

Blokirali ste korisnika/cu @ofirnachum

Jeste li sigurni da želite vidjeti te tweetove? Time nećete deblokirati korisnika/cu @ofirnachum

  1. 8. sij

    Many of our group's recent papers - DualDICE, ValueDICE, GenDICE, AlgaeDICE - can be framed as applications of this duality. Still, lots of potential remaining applications for others to explore, and lingering questions of how these formulations interplay with stoch. opt methods.

    Poništi
  2. 8. sij

    Fenchel-Rockafellar duality is a powerful tool that more people should be aware of, especially for RL! Straightforward applications of it enable offpolicy evaluation, offpolicy policy gradient/imitation learning, among others

    Poništi
  3. 17. pro 2019.

    Code is here: Hope this will serve as a strong baseline for future work to improve on!

    Prikaži ovu nit
    Poništi
  4. 17. pro 2019.

    <-- my most recent paper with & J. Tompson introducing *ValueDICE* - an off-policy imitation learning alg. We set new SOTA for online imitation learning, and for the 1st time (afaik) beat behavior cloning in the totally offline regime.

    Prikaži ovu nit
    Poništi
  5. 4. pro 2019.

    Very excited about my new paper! We formulate the on-policy max-return RL objective w.r.t *arbitrary* offline data and without *any* explicit importance correction. Amazingly, the gradient of the objective w.r.t pi is exactly the on-policy policy gradient!

    Poništi
  6. 27. stu 2019.

    Offline RL -what do you need to know about this notoriously difficult regime? Although recent papers propose a variety of algorithmic novelties, we find many of these unnecessary in practice. Extensive studies will hopefully guide future research &practice

    Poništi
  7. 8. lis 2019.

    [new paper] Adding a simple linear constraint to a loss function (commonly used in ML fairness) can have strange and unintuitive effects on the resulting model. While such constraints seem natural/harmless, we don't fully understand their consequences yet

    Poništi
  8. 24. ruj 2019.

    new paper! We investigate the underlying reasons for success of hierarchical RL, finding that (surprisingly) much of it is due to exploration, and that this benefit can be achieved *without* explicit hierarchies of policies

    Poništi
  9. 29. kol 2019.

    A few people had asked for this so we open-sourced our code for "Identifying and Correcting Label Bias in Machine Learning". Hope people can find it useful and build on top of it!

    Poništi
  10. 14. kol 2019.

    I'm excited to announce the release of my most recent work: applying HRL to robotic "manipulation via locomotion" tasks with impressive real-world results! w/ amazing collaborators - and more!

    Poništi
  11. 29. lip 2019.
    Poništi
  12. 25. lip 2019.

    New paper out! An advancement in properly estimating off-policy occupancy ratios. We apply it to off-policy policy evaluation with great results, but we believe it should be useful in many more off-policy settings!

    Poništi
  13. proslijedio/la je Tweet
    20. lip 2019.

    Great thanks to for the honor to kickoff the Deep Reinforcement Learning Summit with the important topic on "Secure Deep Reinforcement Learning"! Amazing talks from fellow speakers and others!

    Poništi
  14. 14. lip 2019.

    Our submission won best paper at the RL4RL workshop at ICML!

    Poništi
  15. 9. lip 2019.

    Nice work with great collaborators - we'll be presenting this at ICML! Tue Jun 11th 03:05PM @ Room 104 and Tuesday poster #108.

    Poništi
  16. proslijedio/la je Tweet
    24. ožu 2019.

    Fast & Simple Resource-Constrained Learning of Deep Network Structure By Suppose you have a working CNN, MorphNet will adjust number of channels in each layer to satisfy memory/latency constraints

    Poništi
  17. 22. ožu 2019.

    Open source library for "MorphNets: Fast & Simple Resource-Constrained Learning of Deep Network Structure" has been released . Based on work (CVPR 2018) with and others at ().

    Poništi
  18. 10. velj 2019.

    Great summary of our recent work about safe RL, a collaboration between researchers at Google, DeepMind, and FAIR:

    Poništi
  19. proslijedio/la je Tweet
    29. sij 2019.
    Odgovor korisniku/ci

    Hey, a really interesting and elegant paper! I wrote a summary of the main ideas, aiming to make them accessible to a broad audience. I hope you like it!

    Poništi
  20. 16. sij 2019.

    My new paper on learning fair machine learning classifiers: We frame the problem as trying to learn with respect to unknown (and true) labels despite only having access to observed (and biased) labels. We find a surprisingly simple solution for doing so!

    Poništi

Čini se da učitavanje traje već neko vrijeme.

Twitter je možda preopterećen ili ima kratkotrajnih poteškoća u radu. Pokušajte ponovno ili potražite dodatne informacije u odjeljku Status Twittera.

    Možda bi vam se svidjelo i ovo:

    ·