Tweetovi

Blokirali ste korisnika/cu @clark_kev

Jeste li sigurni da želite vidjeti te tweetove? Time nećete deblokirati korisnika/cu @clark_kev

  1. proslijedio/la je Tweet

    🔥“we train a model on one GPU for 4 days that outperforms GPT (trained using 30x more compute) on the GLUE natural language understanding ... we match the performance of RoBERTa, the current state-of-the-art pre-trained transformer, while using less than 1/4 of the compute.” 🤯

    Poništi
  2. proslijedio/la je Tweet
    27. stu 2019.

    New analysis paper from my group! We zoom in on some of et al.'s on syntax-sensitive attention heads in BERT (+RoBERTa, +...), and find interestingly mixed results.

    Poništi
  3. proslijedio/la je Tweet
    3. stu 2019.

    Excited to share new work!!! “Generalization through Memorization: Nearest Neighbor Language Models” We introduce kNN-LMs, which extend LMs with nearest neighbor search in embedding space, achieving a new state-of-the-art perplexity on Wikitext-103, without additional training!

    Prikaži ovu nit
    Poništi
  4. proslijedio/la je Tweet
    30. lis 2019.

    Excited to share our work on BART, a method for pre-training seq2seq models by de-noising text. BART outperforms previous work on a bunch of generation tasks (summarization/dialogue/QA), while getting similar performance to RoBERTa on SQuAD/GLUE

    Prikaži ovu nit
    Poništi
  5. proslijedio/la je Tweet
    10. ruj 2019.

    How do we design probes that give us insight into a representation? In paper with , our "control tasks" help us understand the capacity of a probe to make decisions unmotivated by the repr. paper: blog:

    Prikaži ovu nit
    Poništi
  6. proslijedio/la je Tweet

    best paper award went to: What does BERT look at? An Analysis of BERT’s Attention. Kevin Clark, Urvashi Khandelwal, Omer Levy and Christopher D. Manning.

    Poništi
  7. 10. srp 2019.

    BAM! Our new paper presents "Born-Again Multi-Task Networks," a simple way to improve multi-task learning using knowledge distillation. With . Paper: Code:

    Poništi
  8. 27. lip 2019.

    Code for our paper "What Does BERT Look At? An Analysis of BERT's Attention" () has been released!

    Poništi
  9. 12. lip 2019.

    Check out our new paper "What Does BERT Look At? An Analysis of BERT's Attention" with ! Among other things, we show that BERT's attention corresponds surprisingly well to aspects of syntax and coreference.

    Poništi

Čini se da učitavanje traje već neko vrijeme.

Twitter je možda preopterećen ili ima kratkotrajnih poteškoća u radu. Pokušajte ponovno ili potražite dodatne informacije u odjeljku Status Twittera.

    Možda bi vam se svidjelo i ovo:

    ·