Beyond Word Embeddings

how-to-understand-bert 🤗

This is a collection of Bert Learning resources created by Roman Kryvokhyzha. The tokenization in context of BERT provides great details of the embeddings used in Bert.


Theory 🙌

So, I have started my long way in BERT with blog posts and videos produced by Chris McCormick. After watching the first video, I clearly understood that it is necessary to read more about Attention and Transformer.

  1. Attention
  2. Transformer

Next, let’s talk about tokenization in context of BERT.

  • [en] WordPiece Embeddings - II part of BERT Research by Chris McCormick ❤️

  • [en] Tokenizers: How machines read ❤️

  • also, in vocabulary folder you can find different examples ❤️‍🔥


And finally, let’s read more about BERT.


Examples 💪


Additional sources 📚