У нас вы можете посмотреть бесплатно Self-Distillation Enables Continual Learning или скачать в максимальном доступном качестве, видео которое было загружено на ютуб. Для загрузки выберите вариант из формы ниже:
Если кнопки скачивания не
загрузились
НАЖМИТЕ ЗДЕСЬ или обновите страницу
Если возникают проблемы со скачиванием видео, пожалуйста напишите в поддержку по адресу внизу
страницы.
Спасибо за использование сервиса ClipSaver.ru
🔄 Unlocking the Future of AI: Self-Distillation for Continual Learning! https://www.emergent-behaviors.com/so... In this engaging video, we delve into the innovative research paper "Self-Distillation Enables Continual Learning" by the talented team from MIT, Improbable AI Lab, and ETH Zurich. Discover how self-distillation fine-tuning (SDFT) offers a powerful solution to the static model problem, allowing AI systems to learn new skills while retaining existing knowledge. We explore the limitations of traditional supervised fine-tuning (SFT) and the phenomenon of catastrophic forgetting. Learn how self-distillation transforms the learning process by enabling models to teach themselves. This comprehensive breakdown will enhance your understanding of continual learning and its implications for AI development. 🌟 📌 What You'll Learn: • The static model problem and its implications for AI training • How self-distillation fine-tuning (SDFT) overcomes limitations of supervised fine-tuning • The mechanics of the distillation loop and its impact on continual learning • Key results demonstrating SDFT's advantages over traditional methods ⏳ Timestamps: 0:00 Introduction to Self-Distillation for Learning 0:45 The Static Model Problem: Frozen in Time 1:30 The Villain: Supervised Fine-Tuning as Imitation Only 2:15 Catastrophic Forgetting: Learning New Tasks Erases Old Skills 3:00 The Ideal and the Nightmare: Reinforcement Learning Needs a Reward 3:45 Epiphany: Self-Distillation Fine-Tuning (SDFT) 4:30 How It Works: Prompt Injection Creates the Teacher Distribution 5:15 The Distillation Loop: On-Policy Learning with Reverse KL 6:00 The Math Intuition: Implicit Reward from Teacher Probabilities 6:45 Results: Pareto Frontier of Old vs New Task Performance 7:30 Sequential Skill Learning: The Juggling Experiment Over Time 8:15 Knowledge Injection: The 2025 Myanmar Earthquake Test 9:00 Scaling Laws: SDFT Improves as the Base Model Becomes a Better Teacher 9:45 Hidden Superpower: Reasoning Recovery from Answer-Only Data 10:30 Conclusion: Continual Learning without Explicit Rewards SELF-DISTILLATION ENABLES CONTINUAL LEARNING https://arxiv.org/pdf/2601.19897 Idan Shenfeld, MIT; Mehul Damani, MIT; Jonas Hübotter, ETH Zurich; Pulkit Agrawal, MIT #AI #MachineLearning #SelfDistillation #ContinualLearning #SupervisedFineTuning #ReinforcementLearning #CatastrophicForgetting #InContextLearning #AIResearch #TechInnovation #MIT #ETHZurich #ImprobableAI #FutureOfAI #DeepLearning #ArtificialIntelligence ---