This week we’re looking into Transformers.
Transformers were introduced a couple of years ago with the paper Attention is All You Need by Google Researchers. Since its introduction, Transformers has been widely adopted in the industry. Models like BERT, GPT-3 made groundbreaking improvements in the world of NLP and ASR using Transformers. Since then, model libraries like HuggingFace made it possible for everyone to use Transformer based models in their projects.
But what are Transformers and how do they work? How are they different from other Deep Learning models like RNNs, LSTMs? Why are they better? In this video, we learn about it all!