This course will guide you through the essential concepts of Transformer Neural Networks and their implementation using PyTorch. Starting with an introduction to Transformers, you will learn to build and train Transformer models from scratch. Additionally, you will explore the advantages of using pre-trained Transformer models and how to leverage them effectively in your projects. By the end of this course, you will have a solid foundation in programming Transformer Neural Networks with PyTorch.
Overview
Syllabus
- Introduction to Transformer Neural Networks
- Explore Transformer neural networks, their architecture, and applications like ChatGPT. Delve into NLP basics, tokenization, and model training using PyTorch for AI advancements.
- Building Transformer Neural Networks with PyTorch
- Learn to build a Transformer model with PyTorch, covering tokenization, embeddings, multi-head attention, training, and text generation for NLP tasks.
- Using Pre-Trained Transformers
- Master the use of pre-trained transformers, including their training, fine-tuning, limitations, and applying them to NLP tasks like text generation and QA.
Taught by
Ivan Mushketyk