Master Finance Tools - 35% Off CFI (Code CFI35)
PowerBI Data Analyst - Create visualizations and dashboards from scratch
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore a 42-minute technical video demonstrating how to enhance mathematical reasoning capabilities in smaller language models through the DeepSeek R1 coldstart method. Learn the process of structuring and generating chain-of-thought prompts based on the DeepSeek R1 paper, and discover how to create synthetic training data using a custom mathematics compiler. Follow along with the complete fine-tuning pipeline that enables a 1.5B parameter model to achieve superior mathematical problem-solving performance compared to larger models. Dive into key concepts including think tags, synthetic chain-of-thought generation, natural language translation, self-reflection mechanisms, and practical implementation using Qwen2.5-1.5B. Access companion GitHub repositories for the math compiler and verifiers while following a detailed walkthrough from initial setup through testing and final model evaluation.
Syllabus
- intro
- DeepSeek R1 Chat
- DeepSeek R1 Ollama
- Think Tags
- Deep Seek R1 paper
- Generating synthetic long chains of thought
- Translating the CoT to natural language
- Self Reflection and Self Correction
- Generating sample data
- Testing the Qwen2.5-1.5B
- Fine Tuning Qwen2.5-1.5B with our Coldstart data
- Chatting with our Fine Tuned Model
- Conclusion
Taught by
Chris Hay