BERT (Bidirectional Encoder Representations from Transformers) was proposed by researchers at Google AI Language in 2018. It has caused a stir in the Machine Learning community by presenting state-of-the-art results in a variety of NLP tasks. In this post, we will get an overview of BERT, how it works and what makes BERT so effective.

An Introduction to BERT