Working of BERT

Learn how BERT is bidirectional with detailed examples.

Bidirectional Encoder Representation from Transformer (BERT), as the name suggests, is based on the transformer model. We can perceive BERT as the transformer, but only with the encoder.

In transformers, we feed the sentence as input to the transformer's encoder, and it returns the representation for each word in the sentence as an output. Well, that's exactly what BERT is—an encoder representation from transformer. Okay, so what about the term Bidirectional?

How BERT is bidirectional

Get hands-on with 1200+ tech skills courses.