Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Qwen Image Models Training - 0 to Hero Level Tutorial - LoRA and Fine Tuning - Base and Edit Model

Software Engineering Courses - SE Courses via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Master Qwen image model training through this comprehensive step-by-step tutorial covering both LoRA training and full fine-tuning/DreamBooth techniques for Qwen Image base and Edit Plus 2509 models. Learn to train locally on Windows computers with GPUs requiring as little as 6GB VRAM using a custom-developed Gradio application that simplifies the legendary Kohya Musubi Tuner trainer. Discover the complete workflow from dataset preparation and quality guidelines to monitoring training performance, testing checkpoints, and generating high-quality images with specialized presets. Explore advanced techniques including automatic image cropping and resizing, captioning strategies, VRAM optimization, checkpoint comparison using grid generators, and upscaling workflows with automatic face fixing through segmentation inpainting. Master specialized training scenarios for style and product training, learn prompt engineering with Google AI Studio Gemini, and understand the differences between BF16 and FP8 scaled training tiers. Gain expertise in converting large fine-tuned models, resuming incomplete training sessions, and implementing the trained models in SwarmUI for practical applications. The tutorial includes detailed coverage of dataset structuring, variety importance in training data, performance optimization techniques including overclocking, and batch image generation with wildcards for efficient workflow management.

Syllabus

0:00 Introduction & Tutorial Goals
0:59 Showcase: Realistic vs. Style Training GTA 5 Example
1:26 Showcase: High-Quality Product Training
1:40 Showcase: Qwen Image Edit Model Capabilities
1:57 Effort & Cost Behind The Tutorial
2:19 Introducing The Custom Training Application & Presets
3:09 Power of Qwen Models: High-Quality Results from a Small Dataset
3:58 Detailed Tutorial Outline & Chapter Flow
4:36 Part 4: Dataset Preparation Critical Section
5:05 Part 5: Monitoring Training & Performance
5:23 Part 6: Generating High-Quality Images with Presets
5:44 Part 7: Specialized Training Scenarios
6:07 Why You Should Watch The Entire Tutorial
7:15 Part 1 Begins: Finding Resources & Downloading The Zip File
7:50 Mandatory Prerequisites Python, CUDA, FFmpeg
8:30 Core Application Installation on Windows
9:47 Part 2: Downloading The Qwen Training Models
10:28 Features of The Custom Downloader Fast & Resumable
11:24 Verifying Model Downloads & Hash Check
12:41 Part 3 Begins: Starting The Application & UI Overview
13:16 Crucial First Step: Selecting & Loading a Training Preset
13:43 Understanding The Preset Structure LoRA/Fine-Tune, Epochs, Tiers
15:01 System & VRAM Preparation: Checking Your Free VRAM
16:07 How to Minimize VRAM Usage Before Training
17:06 Setting Checkpoint Save Path & Frequency
19:05 Saving Your Custom Configuration File
19:52 Part 4 Begins: Dataset Preparation Introduction
20:10 Using The Ultimate Batch Image Processing Tool
20:53 Stage 1: Auto-Cropping & Subject Focusing
23:37 Stage 2: Resizing Images to Final Training Resolution
25:49 Critical: Dataset Quality Guidelines & Best Practices
27:19 The Importance of Variety Clothing, Backgrounds, Angles
29:10 New Tool: Internal Image Pre-Processing Preview
31:21 Using The Debug Mode to See Each Processed Image
32:21 How to Structure The Dataset Folder For Training
34:31 Pointing The Trainer to Your Dataset Folder
35:19 Captioning Strategy: Why a Single Trigger Word is Best
36:30 Optional: Using The Built-in Detailed Image Captioner
39:56 Finalizing Model Paths & Settings
40:34 Setting The Base Model, VAE, and Text Encoder Paths
41:59 Training Settings: How Many Epochs Should You Use?
43:45 Part 5 Begins: Starting & Monitoring The Training
46:41 Performance Optimization: How to Improve Training Speed
48:35 Tip: Overclocking with MSI Afterburner
49:25 Part 6 Begins: Testing & Finding The Best Checkpoint
51:35 Using The Grid Generator to Compare Checkpoints
55:33 Analyzing The Comparison Grid to Find The Best Checkpoint
57:21 How to Resume an Incomplete LoRA Training
59:02 Generating Images with Your Best LoRA
1:00:21 Workflow: Generate Low-Res Previews First, Then Upscale
1:01:26 The Power of Upscaling: Before and After
1:02:08 Fixing Faces with Automatic Segmentation Inpainting
1:04:28 Manual Inpainting for Maximum Control
1:06:31 Batch Generating Images with Wildcards
1:08:49 How to Write Excellent Prompts with Google AI Studio Gemini
1:10:04 Quality Comparison: Tier 1 BF16 vs Tier 2 FP8 Scaled
1:12:10 Part 7 Begins: Fine-Tuning DreamBooth Explained
1:13:36 Converting 40GB Fine-Tuned Models to FP8 Scaled
1:15:15 Testing Fine-Tuned Checkpoints
1:16:27 Training on The Qwen Image Edit Model
1:17:39 Using The Trained Edit Model for Prompt-Based Editing
1:24:22 Advanced: Teaching The Edit Model New Commands Control Images
1:27:01 Performance Impact of Training with Control Images
1:31:41 How to Resume an Incomplete Fine-Tuning Training
1:33:08 Recap: How to Use Your Trained Models
1:35:36 Using Fine-Tuned Models in SwarmUI
1:37:16 Specialized Scenario: Style Training
1:38:20 Style Dataset Guidelines: Consistency & No Repeating Elements
1:40:25 Generating Prompts for Your Trained Style with Gemini
1:44:45 Generating Images with Your Trained Style Model
1:46:41 Specialized Scenario: Product Training
1:47:34 Product Dataset Guidelines: Proportions & Detail Shots
1:48:56 Generating Prompts for Your Trained Product with Gemini
1:50:52 Conclusion & Community Links Discord, GitHub, Reddit

Taught by

Software Engineering Courses - SE Courses

Reviews

Start your review of Qwen Image Models Training - 0 to Hero Level Tutorial - LoRA and Fine Tuning - Base and Edit Model

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.