Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Building a Word-Level Language Model Using RNNs in PyTorch - Lab 4.4

Donato Capitella via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn to implement a word-level language model using Recurrent Neural Networks (RNNs) in PyTorch through a hands-on tutorial that guides you through working with fairy tales datasets, tokenization techniques, and vocabulary creation. Master the process of preparing data for self-supervised learning, constructing and training the model architecture, and explore advanced concepts like temperature and top-p sampling for text generation. Dive into word embeddings while following along with the provided Colab notebook, which includes practical examples and comprehensive code implementations. Progress through key concepts including model creation, text generation techniques, and the training process, all demonstrated using a fairy tales dataset to build a functional language model from scratch.

Syllabus

- Introduction
- Fairy Tales Dataset
- Tokenization
- Creating a Vocabulary of Tokens
- Preparing Data for Self-Supervised Learning
- Creating the Model
- Text Generation
- Training the Model
- Temperature and Top-p Sampling
- Word Embeddings

Taught by

Donato Capitella

Reviews

Start your review of Building a Word-Level Language Model Using RNNs in PyTorch - Lab 4.4

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.