Rajat Gupta

@GuptaRajat033

Software Engineer;Udacity Mentor for NLP,AI and DRL NDs, Student ; dally on twitter, reddit, arxiv, kaggle and MOOCs in DL

Vrijeme pridruživanja: lipanj 2013.

Tweetovi

Blokirali ste korisnika/cu @GuptaRajat033

Jeste li sigurni da želite vidjeti te tweetove? Time nećete deblokirati korisnika/cu @GuptaRajat033

  1. proslijedio/la je Tweet
    28. sij

    New paper: Towards a Human-like Open-Domain Chatbot. Key takeaways: 1. "Perplexity is all a chatbot needs" ;) 2. We're getting closer to a high-quality chatbot that can chat about anything Paper: Blog:

    Prikaži ovu nit
    Poništi
  2. proslijedio/la je Tweet
    14. sij

    BigGAN samples are famously photo-realistic but limited in diversity for some classes. Slightly modifying only the class embeddings (network unchanged) can reduce the diversity gap by ~50%! Work with Long Mai and led by fantastic !! Paper & video:

    Prikaži ovu nit
    Poništi
  3. proslijedio/la je Tweet
    10. sij

    Very happy to share our latest work accepted at : we prove that a Self-Attention layer can express any CNN layer. 1/5 📄Paper: 🍿Interactive website : 🖥Code: 📝Blog:

    Prikaži ovu nit
    Poništi
  4. proslijedio/la je Tweet
    5. sij

    Drifting in an autonomous vehicle. Uses rotation rate of the vehicle’s velocity vector to track the path, while yaw acceleration is used to stabilize sideslip. Could help autonomous vehicles in emergencies. Fun results.

    Poništi
  5. proslijedio/la je Tweet
    2. sij

    The 2010s were an eventful decade for NLP! Here are ten shocking developments since 2010, and 13 papers* illustrating them, that have changed the field almost beyond recognition. (* in the spirit of and , exclusively from other groups :)).

    Prikaži ovu nit
    Poništi
  6. proslijedio/la je Tweet

    Solo papers are going back into style! They are great because they are like ambitious manifestos that describe a unique idea and boldly broadcast that this is my idea and my idea alone! Here are the latest ones from prominent AI researchers.

    Prikaži ovu nit
    Poništi
  7. proslijedio/la je Tweet

    Still slowly making my way through this year's NeurIPS talks. Esp like to stumble by good talks from slightly different areas, e.g. tonight liked "ML Meets Single-Cell Biology" Incredible that we're mapping out cell state markov chains for tissues.

    Poništi
  8. proslijedio/la je Tweet

    The press about surveillance capitalism focuses on unintended effects like hacks and data theft. That’s important, but we must resist the *intended* effect—the cultivation of a consumerist society whose behavior can be manipulated at a massive scale to suit commercial interests.

    Prikaži ovu nit
    Poništi
  9. proslijedio/la je Tweet
    24. pro 2019.

    i can't make this up

    Poništi
  10. proslijedio/la je Tweet
    23. pro 2019.

    BackPACK: Packing more into backprop "we introduce BackPACK, an efficient framework built on top of PyTorch, that extends the backpropagation algorithm to extract additional information from first- and second-order derivatives"

    Poništi
  11. proslijedio/la je Tweet
    13. pro 2019.

    Can we simulate both a user👩‍🦰 and system 🤖 and learn to autocomplete in an unsupervised way? Yes! We frame the autocomplete task as a cooperative communication game. Talk: Dec 14 9:45-10AM (West 118)

    Prikaži ovu nit
    Poništi
  12. proslijedio/la je Tweet
    4. pro 2019.

    Our new paper, Deep Learning for Symbolic Mathematics, is now on arXiv We added *a lot* of new results compared to the original submission. With (1/7)

    Prikaži ovu nit
    Poništi
  13. proslijedio/la je Tweet
    28. stu 2019.

    Excited to share our work on Contrastive Learning of Structured World Models! C-SWMs learn object-factorized models & discover objects without supervision, using a simple loss inspired by work on graph embeddings Paper: Code: 1/5

    Prikaži ovu nit
    Poništi
  14. proslijedio/la je Tweet
    10. lip 2019.

    My most recent residency project is out: "Hamiltonian Neural Networks" Blog: Paper: Starting from noisy (pixel) data, we can learn _exact_ conservation of energy-like quantities.

    Prikaži ovu nit
    Poništi
  15. proslijedio/la je Tweet
    2. pro 2019.

    I've been working with for a long time to create something that combines the best of Notebooks with the best of traditional software development approaches. It's called nbdev. We're releasing it today as open source. 1/

    Prikaži ovu nit
    Poništi
  16. proslijedio/la je Tweet
    1. pro 2019.

    Large language models are starting to capture larger swaths of English Grammar, and several of us at NYU have gotten interested in trying to get a broad overview of where models are succeeding and failing. [new dataset alert; thread]

    Prikaži ovu nit
    Poništi
  17. proslijedio/la je Tweet
    27. stu 2019.

    In our recent collaboration with we show how to generate realistic complex scenes from scratch! While the problem is extremely challenging, we show how to achieve SOTA in unconditional generation and improve conditional generation using SPADE

    Poništi
  18. proslijedio/la je Tweet
    26. stu 2019.

    Introducing the SHA-RNN :) - Read alternative history as a research genre - Learn of the terrifying tokenization attack that leaves language models perplexed - Get near SotA results on enwik8 in hours on a lone GPU No Sesame Street or Transformers allowed.

    The SHA-RNN is composed of an RNN, pointer based attention, and a “Boom” feed-forward with a sprinkling of layer normalization. The persistent state is the RNN’s hidden state h as well as the memory M concatenated from previous memories. Bake at 200◦F for 16 to 20 hours in a desktop sized oven.
    The attention mechanism within the SHA-RNN is highly computationally efficient. The only matrix multiplication acts on the query. The A block represents scaled dot product attention, a vector-vector operation. The operators {qs, ks, vs} are vectorvector multiplications and thus have minimal overhead. We use a sigmoid to produce {qs, ks}. For vs see Section 6.4.
    Bits Per Character (BPC) onenwik8. The single attention SHA-LSTM has an attention head on the second last layer and hadbatch size 16 due to lower memory use. Directly comparing the head count for LSTM models and Transformer models obviously doesn’tmake sense but neither does comparing zero-headed LSTMs against bajillion headed models and then declaring an entire species dead.
    Poništi
  19. proslijedio/la je Tweet
    25. stu 2019.

    Something i've been working on for the last couple of months: Train practical models on your data, download a pip installable model, with the idea of reducing the amount of time between idea and implementation.

    Prikaži ovu nit
    Poništi
  20. proslijedio/la je Tweet
    24. stu 2019.

    I wrote an 8k word doc on machine learning systems design. This covers: 1. Project setup 2. Data pipeline 3. Training & debugging 4. Serving with case studies, resources, and 27 exercises. This is the 1st draft so feedback is much needed. Thank you!

    Poništi

Čini se da učitavanje traje već neko vrijeme.

Twitter je možda preopterećen ili ima kratkotrajnih poteškoća u radu. Pokušajte ponovno ili potražite dodatne informacije u odjeljku Status Twittera.

    Možda bi vam se svidjelo i ovo:

    ·