Overview
Syllabus
[00:00:00] Introducing Jarek and DeepL’s mission
[00:01:46] Competing with Google Translate & LLMs
[00:04:14] Pretraining vs. proprietary model strategy
[00:06:47] Building GPU data centers in 2017
[00:08:09] The value of curated bilingual and monolingual data
[00:09:30] How DeepL measures translation quality
[00:12:27] Personalization and enterprise-specific tuning
[00:14:04] Why translation demand is growing
[00:16:16] ROI of incremental quality gains
[00:18:20] The role of human translators in the future
[00:22:48] Hallucinations in translation models
[00:24:05] DeepL’s work on speech translation
[00:28:22] The broader impact of global communication
[00:30:32] Handling smaller languages and language pairs
[00:32:25] Multi-language model consolidation
[00:35:28] Engineering infrastructure for large-scale inference
[00:39:23] Adapting to evolving LLM landscape & enterprise needs
Taught by
Weights & Biases