Tweets
- Tweets, current page.
- Tweets & replies
- Media
You blocked @EzraWu
Are you sure you want to view these Tweets? Viewing Tweets won't unblock @EzraWu
-
The thesis and slides will be available online shortly after the finalization of the thesis and one under review chapter getting cleared
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
Life update: I have successfully defended my thesis and will join Bloomberg AI in Januarypic.twitter.com/k42htVyvky
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
Resources we released: * Large English-Arabic encoder https://huggingface.co/jhu-clsp/roberta-large-eng-ara-128k … * Data projection https://github.com/shijie-wu/crosslingual-nlp … * Arabic denormalization toolkit https://github.com/KentonMurray/ArabicDetokenizer … [4/4]
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
We also conducted multilingual experiments with 8 target languages. This is a joint work with Mahsa Yarmohammadi
@ruyimarone@fe1ixxu@_t_sne@hiaoxui@YunmoChen Jialiang Guo@charman@kentonmurray@aaronsteven@mdredze@ben_vandurme@jhuclsp [3/4]pic.twitter.com/zPWC5nOlAl
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
We conducted extensive experiments on EN-->AR as test case including:
data projection pipeline with various MT models and aligners;
impact of encoder on MT and aligner;
impact of fine-tuning on aligner;
self-training as controlled comparison of data projection. [2/4]pic.twitter.com/vZnOovo2FU
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
Q: What's the best choice to improve zero-shot cross-lingual transfer performance if u can’t manually annotate any data? A: The best setup is task dependent, so try multiple setups!
Check out our #EMNLP2021 paper “Everything Is All It Takes” https://arxiv.org/abs/2109.06798 [1/4]pic.twitter.com/24cwwjxByV
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
Shijie Wu Retweeted
We are excited to share new #UniMorph languages! As a part of@sigmorphon shared task on morphological inflection, we added data for
32 languages from 13 families
. We analyzed systems' predictions on them, conducting an extensive error analysis: https://bit.ly/3CpwivC 1/8pic.twitter.com/67Om3Dc10O
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
Happy to share that our paper on "Applying the Transformer to Character-level Transduction" won the
#EACL2021 honorable mention for short paper awards! Congrats@ryandcotterell and Mans! Talk https://slideslive.com/38954605/applying-the-transformer-to-characterlevel-transduction … Paper https://arxiv.org/abs/2005.10213 Code https://github.com/shijie-wu/neural-transducer …pic.twitter.com/VZTzQfVWXG
Thanks. Twitter will use this to make your timeline better. UndoUndo -
Shijie Wu Retweeted
"Differentiable Generative Phonology", in collaboration with
@EzraWu and@ryandcotterell, is finally out! https://arxiv.org/abs/2102.05717 Tired: Asking linguists to posit discrete underlying forms Wired: learning continuous underlying forms end-to-endShow this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
Shijie Wu Retweeted
Which *BERT (and how can we improve *BERT-science)? Come to the Zoom QA session 13 at 11am-12 EST. See thread for a tl;dr: http://slideslive.com/38939146 And chat about constant memory coreference resolution at Gather session 5, tomorrow 1-3pm EST: http://slideslive.com/38938797
#EMNLP2020https://twitter.com/nlpaxia/status/1312917389235957760 …Thanks. Twitter will use this to make your timeline better. UndoUndo -
Summary of "Which *BERT? A Survey Organizing Contextualized Encoders" https://twitter.com/nlpaxia/status/1312917389235957760?s=20 …
#EMNLP2020Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
Summary of "Do Explicit Alignments Robustly Improve Multilingual Encoders?" https://twitter.com/EzraWu/status/1314575418649968641?s=20 …
#EMNLP2020
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
Q: Do Explicit Alignments Robustly Improve Multilingual Encoders? My talk: https://slideslive.com/38939127/ Code: https://github.com/shijie-wu/crosslingual-nlp … Meeting: Gather town in 10 mins Q: Which *BERT should I use?
@nlpaxia's talk: https://slideslive.com/38939146/ Meeting: Zoom, Nov 18, 11-12 EST#EMNLP2020https://twitter.com/EzraWu/status/1305868107114831872 …Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
Shijie Wu Retweeted
DATASET RELEASE: "CC100", the CommonCrawl dataset of 2.5TB of clean unsupervised text from 100 languages (used to train XLM-R) is now publicly available. You can find below the Data: http://data.statmt.org/cc-100/ Script: https://bit.ly/3oC6aXy By @VishravC et al.pic.twitter.com/KrjcdeRG6P
Thanks. Twitter will use this to make your timeline better. UndoUndo -
Shijie Wu Retweeted
Ok here we go. Look at the electoral maps by county for the last few decades of US presidential elections. You’ll notice that the South goes almost uniformly Republican red every time. Duh. But if you look closer, there’s something else there ...pic.twitter.com/3LMu9lsIPO
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
Interesting work showing word translation from mBERT with template: "The word ‘SOURCE’ in LANGUAGE is: [MASK]."https://twitter.com/hila_gonen/status/1318465935104245760 …
Thanks. Twitter will use this to make your timeline better. UndoUndo -
Shijie Wu Retweeted
New EMNLP paper with
@PeterHndrsn@ukhndlwl@robinomial@kmahowald and@jurafsky -- With Little Power Comes Great Responsibility -- https://arxiv.org/abs/2010.06595 (1/3)Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
… (3) quality of bitext has small impact on downstream performance (4) while alignment helps mBERT in some cases, none of the methods we considered improve XLMR (5) a bigger model leads to much bigger gain compared to ad-hoc alignment, without any need for bitext (3/3)
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
... (1) contrastive alignment consistently outperforms L2 alignment and performs more robustly than linear mapping (2) zero/few-shot cross-lingual transfer (includ. benchmarks) needs to report mean & variance w/ different seeds, similar to the few-shot learning community .. (2/3)
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo -
Happy to share our
#EMNLP2020 paper (w/@mdredze) on contrastive cross-lingual alignment titled “Do Explicit Alignments Robustly Improve Multilingual Encoders?” Expand thread for TL;DR ... (1/3) Paper https://arxiv.org/abs/2010.02537 Code https://github.com/shijie-wu/crosslingual-nlp …pic.twitter.com/nbkD6NZnBx
Show this threadThanks. Twitter will use this to make your timeline better. UndoUndo
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.