Remove 2020 Remove BERT Remove Computational Linguistics
article thumbnail

The State of Multilingual AI

Sebastian Ruder

Research models such as BERT and T5 have become much more accessible while the latest generation of language and multi-modal models are demonstrating increasingly powerful capabilities. Models that allow interaction via natural language have become ubiquitious. The size of the gradient circle represents the number of languages in the class.

article thumbnail

Multi-domain Multilingual Question Answering

Sebastian Ruder

Reading Comprehension assumes a gold paragraph is provided Standard approaches for reading comprehension build on pre-trained models such as BERT. Using BERT for reading comprehension involves fine-tuning it to predict a) whether a question is answerable and b) whether each token is the start and end of an answer span.

BERT 52
professionals

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

All Languages Are NOT Created (Tokenized) Equal

Topbots

70% of research papers published in a computational linguistics conference only evaluated English.[ In Findings of the Association for Computational Linguistics: ACL 2022 , pages 2340–2354, Dublin, Ireland. Association for Computational Linguistics. Are All Languages Created Equal in Multilingual BERT?

article thumbnail

Explosion in 2019: Our Year in Review

Explosion

The update fixed outstanding bugs on the tracker, gave the docs a huge makeover, improved both speed and accuracy, made installation significantly easier and faster, and added some exciting new features, like ULMFit/BERT/ELMo-style language model pretraining. Dec 9: Ines’ key thoughts on trends in AI from 2019 and looking into 2020.

NLP 52
article thumbnail

ML and NLP Research Highlights of 2021

Sebastian Ruder

6] such as W2v-BERT [7] as well as more powerful multilingual models such as XLS-R [8]. For each input chunk, nearest neighbor chunks are retrieved using approximate nearest neighbor search based on BERT embedding similarity. Advances in Neural Information Processing Systems, 2020. What happened?   Why is it important?  

NLP 52
article thumbnail

Reward Isn't Free: Supervising Robot Learning with Language and Video from the Web

The Stanford AI Lab Blog

Conference of the North American Chapter of the Association for Computational Linguistics. ↩ Devlin, J., BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. RoBERTa: A Robustly Optimized BERT Pretraining Approach. Florence: A New Foundation Model for Computer Vision. Neumann, M.,

article thumbnail

2022: We reviewed this year’s AI breakthroughs

Applied Data Science

In our review of 2019 we talked a lot about reinforcement learning and Generative Adversarial Networks (GANs), in 2020 we focused on Natural Language Processing (NLP) and algorithmic bias, in 202 1 Transformers stole the spotlight. Just wait until you hear what happened in 2022. What happened?