У нас вы можете посмотреть бесплатно Grokking, Generalization Collapse, and Dynamics of Training Deep Neural Nets [Charles Martin] - 734 или скачать в максимальном доступном качестве, видео которое было загружено на ютуб. Для загрузки выберите вариант из формы ниже:
Если кнопки скачивания не
загрузились
НАЖМИТЕ ЗДЕСЬ или обновите страницу
Если возникают проблемы со скачиванием видео, пожалуйста напишите в поддержку по адресу внизу
страницы.
Спасибо за использование сервиса ClipSaver.ru
Today, we're joined by Charles Martin, founder of Calculation Consulting, to discuss Weight Watcher, an open-source tool for analyzing and improving Deep Neural Networks (DNNs) based on principles from theoretical physics. We explore the foundations of the Heavy-Tailed Self-Regularization (HTSR) theory that underpins it, which combines random matrix theory and renormalization group ideas to uncover deep insights about model training dynamics. Charles walks us through WeightWatcher’s ability to detect three distinct learning phases—underfitting, grokking, and generalization collapse—and how its signature “layer quality” metric reveals whether individual layers are underfit, overfit, or optimally tuned. Additionally, we dig into the complexities involved in fine-tuning models, the surprising correlation between model optimality and hallucination, the often-underestimated challenges of search relevance, and their implications for RAG. Finally, Charles shares his insights into real-world applications of generative AI and his lessons learned from working in the field. 🗒️ For the full list of resources for this episode, visit the show notes page: https://twimlai.com/go/734. 🔔 Subscribe to our channel for more great content just like this: https://youtube.com/twimlai?sub_confi... 🗣️ CONNECT WITH US! =============================== Subscribe to the TWIML AI Podcast: https://twimlai.com/podcast/twimlai/ Follow us on Twitter: / twimlai Follow us on LinkedIn: / twimlai Join our Slack Community: https://twimlai.com/community/ Subscribe to our newsletter: https://twimlai.com/newsletter/ Want to get in touch? Send us a message: https://twimlai.com/contact/ 📖 CHAPTERS =============================== 00:00 - Introduction 4:08 - WeightWatcher 4:50 - Applying quant techniques to AI 7:03 - Overfitting and underfitting in models 11:40 Challenges in fine-tuning 17:00 - Degrees of fine-tuning 22:15 - Spiking neural networks 27:57 - Grokking 29:30 - Generalization collapse 30:00 - HTSR theory 34:17 - Data-centric AI and layer-specific training 38:45 - Renormalization group 39:29 - Challenges in data access and compliance 42:45 - Benchmarking 47:50 - The correlation between hallucination and model optimality 54:14 - Application of theoretical physics to AI 1:00:58 - Renormalization group, HTSR, and critical exponents 1:08:53 - Evaluation of grokking paper 1:13:25 - Real-world applications and lessons learned 🔗 LINKS & RESOURCES =============================== Calculation Consulting - https://calculationconsulting.com/ WeightWatcher - https://weightwatcher.ai/ Implicit Self-Regularization in Deep Neural Networks: Evidence from Random Matrix Theory and Implications for Learning (HTSR paper) - https://jmlr.org/papers/v22/20-410.html 📸 Camera: https://amzn.to/3TQ3zsg 🎙️Microphone: https://amzn.to/3t5zXeV 🚦Lights: https://amzn.to/3TQlX49 🎛️ Audio Interface: https://amzn.to/3TVFAIq 🎚️ Stream Deck: https://amzn.to/3zzm7F5