A review of state of the art vision-language models such as CLIP, DALLE, ALIGN and SimVL
Learn everything there is to know about the attention mechanisms of the infamous transformer, through 10+1 hidden insights and observations
Understand how positional embeddings emerged and how we use the inside self-attention to model highly structured data such as images
An intuitive understanding on Transformers and how they are used in Machine Translation. After analyzing all subcomponents one by one such as self-attention and positional encodings , we explain the principles behind the Encoder and Decoder and why Transformers work so well
New to Natural Language Processing? This is the ultimate beginner’s guide to the attention mechanism and sequence learning to get you started
Use unsupervised learning to cluster documents based on their content