Generative AI Language Modeling with Transformers
This course is part of multiple programs. Learn more
Instructors: Joseph Santarcangelo +2 more
Instructor ratings
We asked all learners to give feedback on our instructors based on the quality of their teaching style.
What you'll learn
Skills you'll gain
There are 2 modules in this course
In this course, you will learn to apply transformer-based models for text classification, focusing on the encoder component. You’ll learn about positional encoding, word embedding, and attention mechanisms in language transformers and their role in capturing contextual information and dependencies. Additionally, you will be introduced to multi-head attention and gain insights on decoder-based language modeling with generative pre-trained transformers (GPT) for language translation, training the models, and implementing them in PyTorch. Further, you’ll explore encoder-based models with bidirectional encoder representations from transformers (BERT) and train using masked language modeling (MLM) and next sentence prediction (NSP). Finally, you will apply transformers for translation by gaining insight into the transformer architecture and performing its PyTorch implementation. The course offers practical exposure with hands-on activities that enables you to apply your knowledge in real-world scenarios. This course is part of a specialized program tailored for individuals interested in Generative AI engineering. This course requires a working knowledge of Python, PyTorch, and machine learning.
Advanced Concepts of Transformer Architecture
Explore more from Machine Learning
©2025 ementorhub.com. All rights reserved