Skip to content
By using Twitter’s services you agree to our Cookies Use. We and our partners operate globally and use cookies, including for analytics, personalisation, and ads.
  • Home Home Home, current page.
  • About

Saved searches

  • Remove
  • In this conversation
    Verified accountProtected Tweets @
Suggested users
  • Verified accountProtected Tweets @
  • Verified accountProtected Tweets @
  • Language: English
    • Bahasa Indonesia
    • Bahasa Melayu
    • Català
    • Čeština
    • Dansk
    • Deutsch
    • English UK
    • Español
    • Filipino
    • Français
    • Hrvatski
    • Italiano
    • Magyar
    • Nederlands
    • Norsk
    • Polski
    • Português
    • Română
    • Slovenčina
    • Suomi
    • Svenska
    • Tiếng Việt
    • Türkçe
    • Ελληνικά
    • Български език
    • Русский
    • Српски
    • Українська мова
    • עִבְרִית
    • العربية
    • فارسی
    • मराठी
    • हिन्दी
    • বাংলা
    • ગુજરાતી
    • தமிழ்
    • ಕನ್ನಡ
    • ภาษาไทย
    • 한국어
    • 日本語
    • 简体中文
    • 繁體中文
  • Have an account? Log in
    Have an account?
    · Forgot password?

    New to Twitter?
    Sign up
fchollet's profile
François Chollet
François Chollet
François Chollet
Verified account
@fchollet

Tweets

François CholletVerified account

@fchollet

Deep learning @google. Creator of Keras. Author of 'Deep Learning with Python'. Opinions are my own.

United States
fchollet.com
Joined August 2009

Tweets

  • © 2021 Twitter
  • About
  • Help Center
  • Terms
  • Privacy policy
  • Cookies
  • Ads info
Dismiss
Previous
Next

Go to a person's profile

Saved searches

  • Remove
  • In this conversation
    Verified accountProtected Tweets @
Suggested users
  • Verified accountProtected Tweets @
  • Verified accountProtected Tweets @

Promote this Tweet

Block

  • Tweet with a location

    You can add location information to your Tweets, such as your city or precise location, from the web and via third-party applications. You always have the option to delete your Tweet location history. Learn more

    Your lists

    Create a new list


    Under 100 characters, optional

    Privacy

    Copy link to Tweet

    Embed this Tweet

    Embed this Video

    Add this Tweet to your website by copying the code below. Learn more

    Add this video to your website by copying the code below. Learn more

    Hmm, there was a problem reaching the server.

    By embedding Twitter content in your website or app, you are agreeing to the Twitter Developer Agreement and Developer Policy.

    Preview

    Why you're seeing this ad

    Log in to Twitter

    · Forgot password?
    Don't have an account? Sign up »

    Sign up for Twitter

    Not on Twitter? Sign up, tune into the things you care about, and get updates as they happen.

    Sign up
    Have an account? Log in »

    Two-way (sending and receiving) short codes:

    Country Code For customers of
    United States 40404 (any)
    Canada 21212 (any)
    United Kingdom 86444 Vodafone, Orange, 3, O2
    Brazil 40404 Nextel, TIM
    Haiti 40404 Digicel, Voila
    Ireland 51210 Vodafone, O2
    India 53000 Bharti Airtel, Videocon, Reliance
    Indonesia 89887 AXIS, 3, Telkomsel, Indosat, XL Axiata
    Italy 4880804 Wind
    3424486444 Vodafone
    » See SMS short codes for other countries

    Confirmation

     

    Welcome home!

    This timeline is where you’ll spend most of your time, getting instant updates about what matters to you.

    Tweets not working for you?

    Hover over the profile pic and click the Following button to unfollow any account.

    Say a lot with a little

    When you see a Tweet you love, tap the heart — it lets the person who wrote it know you shared the love.

    Spread the word

    The fastest way to share someone else’s Tweet with your followers is with a Retweet. Tap the icon to send it instantly.

    Join the conversation

    Add your thoughts about any Tweet with a Reply. Find a topic you’re passionate about, and jump right in.

    Learn the latest

    Get instant insight into what people are talking about now.

    Get more of what you love

    Follow more accounts to get instant updates about topics you care about.

    Find what's happening

    See the latest conversations about any topic instantly.

    Never miss a Moment

    Catch up instantly on the best stories happening as they unfold.

    1. Julian Togelius‏ @togelius 6 May 2020

      Reinforcement learning is a paradigm that will eventually be superseded. We just haven't figured out what the new, more generally useful, paradigm is yet. When we do, there's going to be a revolution. It will be very interesting.

      33 replies 108 retweets 673 likes
      Show this thread
    2. Thomas G. Dietterich‏ @tdietterich 6 May 2020
      Replying to @togelius

      Is your issue with the problem formulation or with the current set of algorithms? Most of your critique seems aimed at the algorithms, not the formalism.

      1 reply 0 retweets 38 likes
    3. Julian Togelius‏ @togelius 6 May 2020
      Replying to @tdietterich

      It's really about the problem formulation. I think it's holding us back.

      3 replies 0 retweets 13 likes
    4. François Chollet‏Verified account @fchollet 6 May 2020
      Replying to @togelius @tdietterich

      The algos are inadequate to address the problem formulation, and the problem is misleading & counter-productive (in particular reward maximization, static separation between agent & environment, lack of distinction between behavior programs & behavior program generation)

      1 reply 1 retweet 39 likes
    5. Thomas G. Dietterich‏ @tdietterich 6 May 2020
      Replying to @fchollet @togelius

      @fchollet please say more about your last two items: Why is separation between the agent and the environment a problem? What is the distinction between behavior programs and behavior program generation that we need?

      2 replies 2 retweets 10 likes
    6. François Chollet‏Verified account @fchollet 6 May 2020
      Replying to @tdietterich @togelius

      1) An agent is not a static set of possible actions and reward variables. Its affordances change over time: the environment becomes part of the agent. A clever agent will actively seek to gradually *own (absorb) more of the environment* over time.

      3 replies 8 retweets 73 likes
      François Chollet‏Verified account @fchollet 6 May 2020
      Replying to @fchollet @tdietterich @togelius

      2) this emphasizes task-specific skill, which is useless (no one needs a program that plays Pacman) as opposed to algos capable of acquiring arbitrary skills. Which is why deep RL still achieves close to 0 generalization after all these years: generalization was never encouraged.

      10:05 PM - 6 May 2020
      • 4 Retweets
      • 27 Likes
      • Sadih ifrit Sam Wizer Chomba Bupe F.J. Valverde Nikita Voloboev oliver rosario Vikram Shakti Akshay Dev
      5 replies 4 retweets 27 likes
        1. New conversation
        2. Thomas G. Dietterich‏ @tdietterich 6 May 2020
          Replying to @fchollet @togelius

          You must have a more general (!) notion of generalization in mind. AlphaGo certainly can generalize to Go positions that it has never seen during training. That is the form of generalization we have been trying to achieve for many years.

          3 replies 1 retweet 15 likes
        3. Thomas G. Dietterich‏ @tdietterich 6 May 2020
          Replying to @tdietterich @fchollet @togelius

          I don't think task-specific skills are useless. I would love to have an RL system that could make decisions for long-term wildfire management. It is easy to formulate as a (mostly stationary) MDP, but still not practical to solve because the simulator is too expensive

          2 replies 1 retweet 15 likes
        4. Show replies
        1. Sean S. Yi / 이석원 (李碩遠)‏ @seanswyi 7 May 2020
          Replying to @fchollet @tdietterich @togelius

          I'm not anywhere near qualified to butt into this conversation, but wouldn't expecting RL to achieve the generalization that you're implying be analogous to expecting, say, a history professor to also be a physics expert?

          0 replies 0 retweets 0 likes
          Thanks. Twitter will use this to make your timeline better. Undo
          Undo
        1. New conversation
        2. Denny Britz‏ @dennybritz 7 May 2020
          Replying to @fchollet @tdietterich @togelius

          Don’t think anything in the RL formulation forces you to have a static reward fn. or action space, and the definition of “agent” is up to you too. Doesn’t need to be static. I agree that the algorithms are inadequate, but don’t see how the problem formulation is limiting.

          1 reply 0 retweets 8 likes
        3. Denny Britz‏ @dennybritz 7 May 2020
          Replying to @dennybritz @fchollet and

          For example, if you defined a meta-learning problem over a huge range of possible environments with fluid agent definitions, that would still be an RL problem. Perhaps not a very useful benchmark for current algorithms, but totally valid.

          0 replies 0 retweets 1 like
        4. End of conversation
        1. Jose Mora‏ @j_mora 7 May 2020
          Replying to @fchollet @tdietterich @togelius

          Jose Mora Retweeted Jürgen Schmidhuber

          Novelty seeking explainable-model-based RL ~ science: model ~ theory novelty ~ experiment Problems: 1. Find ways to perceive unknown reality, qualitatively: new dimensions 2. Be less conformist, challenge and simplify model: "God does not play dice"https://twitter.com/SchmidhuberAI/status/1255755680855793665 …

          Jose Mora added,

          Jürgen Schmidhuber @SchmidhuberAI
          GANs are special cases of Artificial Curiosity (1990) and also closely related to Predictability Minimization (1991). Now published in Neural Networks 127:58-66, 2020. #selfcorrectingscience #plagiarism Open Access: https://authors.elsevier.com/a/1axlv3BBjKdbNZ … Preprint: https://arxiv.org/abs/1906.04493  pic.twitter.com/5phJUmsYEJ
          0 replies 0 retweets 1 like
          Thanks. Twitter will use this to make your timeline better. Undo
          Undo
        1. Sadih‏ @fkachav 8 May 2020
          Replying to @fchollet @tdietterich @togelius

          Couldn't an ensemble of goals and rewards converge to a generalizable solution? As the number of goals grows so does the "awareness" of the environment.

          0 replies 0 retweets 0 likes
          Thanks. Twitter will use this to make your timeline better. Undo
          Undo

      Loading seems to be taking a while.

      Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.

        Promoted Tweet

        false

        • © 2021 Twitter
        • About
        • Help Center
        • Terms
        • Privacy policy
        • Cookies
        • Ads info