BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

Review

review of the paper 2018 paper introducing BERT, a pre-trained language model
Attention
Bidirectional LSTM
Deep learning
Embeddings
LLM
NLP
Paper
Review
Transformer
Podcast
Author

Oren Bochman

Published

Sunday, May 9, 2021

Keywords

GLUE tasks, SQuAD, SWAG, MultiNLI, NSP, MLM, Contextual embeddings