Skip to content
By using Twitter’s services you agree to our Cookies Use. We and our partners operate globally and use cookies, including for analytics, personalisation, and ads.
  • Home Home Home, current page.
  • About

Saved searches

  • Remove
  • In this conversation
    Verified accountProtected Tweets @
Suggested users
  • Verified accountProtected Tweets @
  • Verified accountProtected Tweets @
  • Language: English
    • Bahasa Indonesia
    • Bahasa Melayu
    • Català
    • Čeština
    • Dansk
    • Deutsch
    • English UK
    • Español
    • Filipino
    • Français
    • Hrvatski
    • Italiano
    • Magyar
    • Nederlands
    • Norsk
    • Polski
    • Português
    • Română
    • Slovenčina
    • Suomi
    • Svenska
    • Tiếng Việt
    • Türkçe
    • Ελληνικά
    • Български език
    • Русский
    • Српски
    • Українська мова
    • עִבְרִית
    • العربية
    • فارسی
    • मराठी
    • हिन्दी
    • বাংলা
    • ગુજરાતી
    • தமிழ்
    • ಕನ್ನಡ
    • ภาษาไทย
    • 한국어
    • 日本語
    • 简体中文
    • 繁體中文
  • Have an account? Log in
    Have an account?
    · Forgot password?

    New to Twitter?
    Sign up
GaryMarcus's profile
Gary Marcus
Gary Marcus
Gary Marcus
@GaryMarcus

Tweets

Gary Marcus

@GaryMarcus

CEO/Founder of http://Robust.AI ; cognitive scientist, and best-selling author. New book: http://Rebooting.AI : Building Artificial Intelligence We Can Trust

garymarcus.com
Joined December 2010

Tweets

  • © 2019 Twitter
  • About
  • Help Center
  • Terms
  • Privacy policy
  • Imprint
  • Cookies
  • Ads info
Dismiss
Previous
Next

Go to a person's profile

Saved searches

  • Remove
  • In this conversation
    Verified accountProtected Tweets @
Suggested users
  • Verified accountProtected Tweets @
  • Verified accountProtected Tweets @

Promote this Tweet

Block

  • Tweet with a location

    You can add location information to your Tweets, such as your city or precise location, from the web and via third-party applications. You always have the option to delete your Tweet location history. Learn more

    Your lists

    Create a new list


    Under 100 characters, optional

    Privacy

    Copy link to Tweet

    Embed this Tweet

    Embed this Video

    Add this Tweet to your website by copying the code below. Learn more

    Add this video to your website by copying the code below. Learn more

    Hmm, there was a problem reaching the server.

    By embedding Twitter content in your website or app, you are agreeing to the Twitter Developer Agreement and Developer Policy.

    Preview

    Why you're seeing this ad

    Log in to Twitter

    · Forgot password?
    Don't have an account? Sign up »

    Sign up for Twitter

    Not on Twitter? Sign up, tune into the things you care about, and get updates as they happen.

    Sign up
    Have an account? Log in »

    Two-way (sending and receiving) short codes:

    Country Code For customers of
    United States 40404 (any)
    Canada 21212 (any)
    United Kingdom 86444 Vodafone, Orange, 3, O2
    Brazil 40404 Nextel, TIM
    Haiti 40404 Digicel, Voila
    Ireland 51210 Vodafone, O2
    India 53000 Bharti Airtel, Videocon, Reliance
    Indonesia 89887 AXIS, 3, Telkomsel, Indosat, XL Axiata
    Italy 4880804 Wind
    3424486444 Vodafone
    » See SMS short codes for other countries

    Confirmation

     

    Welcome home!

    This timeline is where you’ll spend most of your time, getting instant updates about what matters to you.

    Tweets not working for you?

    Hover over the profile pic and click the Following button to unfollow any account.

    Say a lot with a little

    When you see a Tweet you love, tap the heart — it lets the person who wrote it know you shared the love.

    Spread the word

    The fastest way to share someone else’s Tweet with your followers is with a Retweet. Tap the icon to send it instantly.

    Join the conversation

    Add your thoughts about any Tweet with a Reply. Find a topic you’re passionate about, and jump right in.

    Learn the latest

    Get instant insight into what people are talking about now.

    Get more of what you love

    Follow more accounts to get instant updates about topics you care about.

    Find what's happening

    See the latest conversations about any topic instantly.

    Never miss a Moment

    Catch up instantly on the best stories happening as they unfold.

    1. Gary Marcus‏ @GaryMarcus Oct 22
      • Report Tweet

      Gary Marcus Retweeted Alex Rose

      Many people have tried to defend pure DRL w Nature Machine Intelligence article that actually is a hybrid model; the below youtube pointer IS to a pure convnet - but read fine print: “network works decently well for any position less than 6 moves away from solved” #symbolphobiahttps://twitter.com/AlexRoseGames/status/1186571935611850752 …

      Gary Marcus added,

      Alex Rose @AlexRoseGames
      Replying to @ArcusCoTangens @GaryMarcus and 2 others
      for a start, camera detection of a rubik's cube is trivial and has been possible for well over a decade in grandiose fashion https://www.youtube.com/watch?v=jkft2qaKv_o&t=34s … solving a cube using covnet learning is also trivial: https://www.youtube.com/watch?v=QLUca-x2ZVo … nothing you are complaining about is hard
      3 replies 0 retweets 14 likes
    2. Thomas G. Dietterich‏ @tdietterich Oct 22
      • Report Tweet
      Replying to @GaryMarcus

      To solve Rubik's cube in generality, you need to approach it as a reinforcement learning problem. You must explore the search space somehow, e.g., MCTS, Thompson sampling, etc. Is all search in discrete spaces "symbolic"? Not in an interesting way

      5 replies 0 retweets 7 likes
    3. Brad Wyble‏ @bradpwyble Oct 22
      • Report Tweet
      Replying to @tdietterich @GaryMarcus

      I doubt this is how most people learn it. One reads books (or watches youtube) and learns the heuristics (mostly if-then rules) by memorization.

      1 reply 0 retweets 1 like
    4. Brad Wyble‏ @bradpwyble Oct 22
      • Report Tweet
      Replying to @bradpwyble @tdietterich @GaryMarcus

      As to learning the solve without instruction, I would expect that people use a mixture of forward modelling and RL. It can't just be state space exploration.

      1 reply 0 retweets 0 likes
    5. Thomas G. Dietterich‏ @tdietterich Oct 22
      • Report Tweet
      Replying to @bradpwyble @GaryMarcus

      And the forward model is learned somehow. Perhaps through exploring the state space? It's interesting that the model can be factored into the rotations of the cube and the locations of the colored patches. AFAIK that's beyond the state of the art for RL

      1 reply 0 retweets 0 likes
    6. Thomas G. Dietterich‏ @tdietterich Oct 22
      • Report Tweet
      Replying to @tdietterich @bradpwyble @GaryMarcus

      Rubik's cube is a great challenge for creating an agent that can start by learning a forward model and then compile it (through practice == RL) into pattern->action rules. Do we have any learning method that can do all of this in a general way?

      1 reply 0 retweets 1 like
    7. Brad Wyble‏ @bradpwyble Oct 22
      • Report Tweet
      Replying to @tdietterich @GaryMarcus

      I'm genuinely not sure if using a forward-model with discrete states (e.g. realize that rotating CCW will achieve desirable cube state X) is RL. In some sense that might be RL but at the risk of making the definition of RL so broad that it includes all mental activity

      2 replies 0 retweets 0 likes
    8. Thomas G. Dietterich‏ @tdietterich Oct 22
      • Report Tweet
      Replying to @bradpwyble @GaryMarcus

      Understanding what constitutes a "legal move" (rotating CCW) and its effects is what I mean by learning a forward model. Assessing whether a move is good requires reasoning about sequences of moves and their long-term effects. RL is a way to do that (for Markov Decision Problems)

      2 replies 0 retweets 0 likes
      Gary Marcus‏ @GaryMarcus Oct 22
      • Report Tweet
      Replying to @tdietterich @bradpwyble

      it's a way to try, but it's notable that those who have tried fail or use MCTS and sometimes A* as well, presumably because RL on its own fails.

      9:13 PM - 22 Oct 2019
      1 reply 0 retweets 0 likes
        1. New conversation
        2. Thomas G. Dietterich‏ @tdietterich Oct 22
          • Report Tweet
          Replying to @GaryMarcus @bradpwyble

          Every RL algorithm must include an exploration algorithm. MCTS is such an exploration algorithm. A* is not, because it requires a heuristic function, which is not one of the givens for RL problems

          1 reply 0 retweets 1 like
        3. Thomas G. Dietterich‏ @tdietterich Oct 22
          • Report Tweet
          Replying to @tdietterich @GaryMarcus @bradpwyble

          There is no such thing as "RL on its own" without also specifying an exploration method.

          0 replies 0 retweets 3 likes
        4. End of conversation

      Loading seems to be taking a while.

      Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.

        Promoted Tweet

        false

        • © 2019 Twitter
        • About
        • Help Center
        • Terms
        • Privacy policy
        • Imprint
        • Cookies
        • Ads info