Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn to dramatically reduce noise clusters in topic modeling through a rewards-guided, GPU-accelerated approach that improves BERTopic performance by over 20% data retention. Discover how to implement a novel optimization method that addresses the common problem where over 50% of documents get labeled as noise (-1) in traditional BERTopic implementations. Master the application of a scalarized objective function that simultaneously balances coherence for meaningful topics, diversity for distinct topic separation, and noise reduction to eliminate uninformative clusters. Explore hyperparameter optimization techniques using Optuna to fine-tune UMAP and HDBSCAN parameters for optimal results. Achieve 90%+ speedups through GPU acceleration with NVIDIA cuML while maintaining topic quality and interpretability. Gain hands-on experience with practical notebook walkthroughs covering BERTopic fundamentals, objective function implementation, hyperparameter optimization processes, and comprehensive results comparison between baseline and optimized approaches. Access the complete implementation through the provided GitHub notebook to immediately apply these techniques to your own NLP pipelines, large corpus analysis, or topic-driven insight projects.
Syllabus
00:00 Brief introduction of today’s topic: Minimizing Noise Cluster for Topic Modeling
00:20 Background
00:50 Our Method
02:26 BERTopic Introduction
03:29 BERTopic Notebook Walk Through
04:35 The Objective Function
06:04 The Objective Function Notebook Walk Through
07:37 Hyperparameter Optimization HPO
08:21 Hyperparameter Optimization Notebook Walk Through
10:06 Results Comparison Baseline & Our Approach
11:35 GPU Acceleration with NVIDIA cuML
12:48 How to Initiate GPU Acceleration
13:13 Summary and Conclusion
Taught by
NVIDIA Developer