About Generative AI Language Modeling with Transformers course
This course provides you with an overview of how to use transformer-based models for natural language processing (NLP).
In this course, you will learn to apply transformer-based models for text classification, focusing on the encoder component.
You’ll learn about positional encoding, word embedding, and attention mechanisms in language transformers and their role in capturing contextual information and dependencies.
Additionally, you will be introduced to multi-head attention and gain insights on decoder-based language modeling with generative pre-trained transformers (GPT) for language translation, training the models, and implementing them in PyTorch.
Further, you’ll explore encoder-based models with bidirectional encoder representations from transformers (BERT) and train using masked language modeling (MLM) and next sentence prediction (NSP).
Finally, you will apply transformers for translation by gaining insight into the transformer architecture and performing its PyTorch implementation.
The course offers practical exposure with hands-on activities that enables you to apply your knowledge in real-world scenarios.
This course is part of a specialized program tailored for individuals interested in Generative AI engineering.
This course requires a working knowledge of Python, PyTorch, and machine learning.