У нас вы можете посмотреть бесплатно How AI Learns to Cooperate: The Power of In-Context Inference in Multi-Agent Systems или скачать в максимальном доступном качестве, видео которое было загружено на ютуб. Для загрузки выберите вариант из формы ниже:
Если кнопки скачивания не
загрузились
НАЖМИТЕ ЗДЕСЬ или обновите страницу
Если возникают проблемы со скачиванием видео, пожалуйста напишите в поддержку по адресу внизу
страницы.
Спасибо за использование сервиса ClipSaver.ru
Can self-interested AI agents learn to cooperate without being hardcoded to do so? In this episode of SciPulse, we dive into a groundbreaking research paper from Google DeepMind’s Paradigms of Intelligence team: "Multi-agent cooperation through in-context co-player inference." As AI evolves from isolated systems into interacting autonomous agents, ensuring cooperation in "mixed-motive" settings like the Iterated Prisoner’s Dilemma remains a massive challenge. Traditional methods often rely on complex "meta-learning" or rigid assumptions about how other agents think. This paper proposes a simpler, more scalable path. Key Topics Covered: • The Problem of Non-Stationarity: Why standard reinforcement learning often fails when agents are constantly adapting to one another. • Mixed Pool Training: A novel training setup where agents face a diverse distribution of co-players, forcing them to infer strategies on the fly. • In-Context Best-Response: How sequence models (like GRUs) develop the ability to adapt their strategy within a single episode without updating their permanent weights. • The Path to Cooperation: A fascinating three-step mechanism where in-context adaptation makes agents vulnerable to "extortion," which paradoxically drives them toward mutual cooperation. • Predictive Policy Improvement (PPI): An introduction to a new model-based RL algorithm that leverages the generative power of sequence models. Why This Matters: The researchers demonstrate that the complex machinery previously thought necessary for AI cooperation—such as meta-gradients or explicit timescale separation—is actually unnecessary. By bridging the gap between multi-agent reinforcement learning and the training paradigms of modern foundation models, this work suggests that cooperative social behaviors can emerge naturally from diversity and in-context learning. --- Educational Disclaimer: This video is intended as a summary and educational overview of the research paper. It does not replace a thorough reading of the original work. We encourage all viewers to consult the full paper for technical depth and complete experimental data. Read the full paper here: https://arxiv.org/pdf/2602.16301 #AI #MachineLearning #ReinforcementLearning #MultiAgentSystems #GameTheory #InContextLearning #GoogleAI #SciPulse #Research #TechEducation #ArtificialIntelligence #PrisonersDilemma #SequenceModels #FoundationModels