Remove 2017 Remove BERT Remove Knowledge Model
article thumbnail

Commonsense Reasoning for Natural Language Processing

Probably Approximately a Scientific Blog

Traditionally, language models are trained to predict the next word in a sentence (top part of Figure 2, in blue), but they can also predict hidden (masked) words in the middle of the sentence, as in Google's BERT model (top part of Figure 2, in orange). So knowledge in language models is not the most accurate and reliable.