Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Training DeepSeek R1 1.5B Model for Enhanced Mathematical Reasoning Through Coldstart Method

Chris Hay via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore a 42-minute technical video demonstrating how to enhance mathematical reasoning capabilities in smaller language models through the DeepSeek R1 coldstart method. Learn the process of structuring and generating chain-of-thought prompts based on the DeepSeek R1 paper, and discover how to create synthetic training data using a custom mathematics compiler. Follow along with the complete fine-tuning pipeline that enables a 1.5B parameter model to achieve superior mathematical problem-solving performance compared to larger models. Dive into key concepts including think tags, synthetic chain-of-thought generation, natural language translation, self-reflection mechanisms, and practical implementation using Qwen2.5-1.5B. Access companion GitHub repositories for the math compiler and verifiers while following a detailed walkthrough from initial setup through testing and final model evaluation.

Syllabus

- intro
- DeepSeek R1 Chat
- DeepSeek R1 Ollama
- Think Tags
- Deep Seek R1 paper
- Generating synthetic long chains of thought
- Translating the CoT to natural language
- Self Reflection and Self Correction
- Generating sample data
- Testing the Qwen2.5-1.5B
- Fine Tuning Qwen2.5-1.5B with our Coldstart data
- Chatting with our Fine Tuned Model
- Conclusion

Taught by

Chris Hay

Reviews

Start your review of Training DeepSeek R1 1.5B Model for Enhanced Mathematical Reasoning Through Coldstart Method

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.