У нас вы можете посмотреть бесплатно How AI “Grokks” Reality | Geometry of Insight Explained (LLM Research Paper) или скачать в максимальном доступном качестве, видео которое было загружено на ютуб. Для загрузки выберите вариант из формы ниже:
Если кнопки скачивания не
загрузились
НАЖМИТЕ ЗДЕСЬ или обновите страницу
Если возникают проблемы со скачиванием видео, пожалуйста напишите в поддержку по адресу внизу
страницы.
Спасибо за использование сервиса ClipSaver.ru
Link to Substack Article: https://richardaragon.substack.com/p/... How AI “Grokks” Reality | Geometry of Insight Explained (LLM Research Paper) In this video, we break down the research paper: “Do Personality Traits Interfere? Geometric Limitations of Steering in Large Language Models.” Can large language models (LLMs) truly be steered to adopt stable personality traits — or are there geometric limits baked into their representations? We explore: • How LLMs represent concepts in high-dimensional space • The geometric structure behind “grokking” • Why personality steering may conflict internally • Mechanistic interpretability insights • What this means for alignment and controllability This paper provides a deep look into the geometry of representation in transformer models and reveals structural constraints that affect how AI systems encode meaning. Timestamps 00:00 Overview 02:10 What Does It Mean for AI to “Grok”? 05:40 Geometric Representation in LLMs 11:25 Personality Steering Experiments 18:50 Key Limitations 23:10 Implications for AI Alignment If you enjoy deep technical breakdowns of AI research papers with practical experiments, consider subscribing. What do you think — are personality traits fundamentally limited in LLMs? #LLM #AIResearch #MechanisticInterpretability #MachineLearning #TransformerModels #AIAlignment Join this channel to get access to perks: / @richardaragon8471