Whisper Data Preparation and Fine-Tuning with Unsloth

Whisper Data Preparation and Fine-Tuning with Unsloth

Trelis Research via YouTube Direct link

1:23 One-click GPU and Jupyter Notebook Setup

3 of 20

3 of 20

1:23 One-click GPU and Jupyter Notebook Setup

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

Whisper Data Preparation and Fine-Tuning with Unsloth

Automatically move to the next video in the Classroom when playback concludes

  1. 1 0:00 Whisper preparation and fine-tuning with Unsloth
  2. 2 0:40 Resources: Trelis.com/ADVANCED-audio
  3. 3 1:23 One-click GPU and Jupyter Notebook Setup
  4. 4 3:37 Whisper vs Voxtral vs Kyutai
  5. 5 4:48 Installation of Unsloth and Whisper Timestamped
  6. 6 7:52 Using Whisper Large versus Turbo
  7. 7 8:53 Video Overview / Layout - How to prepare data and train
  8. 8 11:33 Audio recording and transcription with whisper timestamped
  9. 9 13:06 Whisper vs Whisper-Timestamped and the motivation for word timestamps
  10. 10 15:34 Creating text/audio segments using word-timestamped transcripts
  11. 11 18:46 Segment time-stamps using whisper not easy to then chunk to less than 30s!!!
  12. 12 19:50 Word time-stamps with Whisper Timestamped
  13. 13 20:56 Automated vs manual transcript cleanup techniques
  14. 14 28:48 Dataset creation from audio and text segments
  15. 15 20:53 Fine-tuning with Unsloth
  16. 16 33:26 Word Error Rate - Teacher Force versus predict_with_generate
  17. 17 36:11 Training hyperparameters and losses / results
  18. 18 37:33 Evaluating base and fine-tuned model performance
  19. 19 39:15 Merging, pushing to hub and preparing for inference see also https://www.youtube.com/watch?v=qXtPPgujufI
  20. 20 40:21 Conclusion

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.