У нас вы можете посмотреть бесплатно AI Safety Beyond Benchmarks -- Dr. Swabha Swayamdipta on Evaluation, Personalization, and Control или скачать в максимальном доступном качестве, видео которое было загружено на ютуб. Для загрузки выберите вариант из формы ниже:
Если кнопки скачивания не
загрузились
НАЖМИТЕ ЗДЕСЬ или обновите страницу
Если возникают проблемы со скачиванием видео, пожалуйста напишите в поддержку по адресу внизу
страницы.
Спасибо за использование сервиса ClipSaver.ru
As language models become more capable, the hardest questions are no longer just about performance, but about safety, interpretation, and control. In this episode of Women in AI Research, we speak with Swabha Swayamdipta, Assistant Professor of Computer Science at the University of Southern California and co-Associate Director of the USC Center for AI and Society. Swabha’s research examines how the design and deployment of language models intersect with real-world risks — from how models behave in unexpected ways to how seemingly technical choices can have broader societal consequences. We talk about AI safety from multiple angles: what it means when hidden inputs to models can sometimes be inferred from their outputs, why personalization introduces new trade-offs around privacy and user agency, and how assumptions about model behavior can quietly shape downstream harms. Rather than focusing only on accuracy or benchmarks, the conversation asks what kinds of evidence we actually need to trust these systems in practice. CHAPTERS 00:00 Swabha's Journey into NLP Research 04:25 Navigating Career Challenges and Building Networks 08:58 The Importance of AI Safety and Reliability 10:49 Addressing Security and Privacy Concerns in Language Models 13:41 Innovations in Language Model Inversion 20:33 Balancing Personalization and Privacy in AI 27:08 Incorporating Psychological Scaffolds in Language Models 30:19 The Duality of AI: Enhancing Human Decision-Making 32:17 AI in Social Issues: Addressing Homelessness 35:18 OATH-Frames: Analyzing Public Sentiment on Homelessness 46:40 Suicide Prevention: AI's Role in Critical Interventions 56:12 The Responsibility of AI Researchers: Balancing Capability and Safety REFERENCES 13:52 Better Language Model Inversion by Compactly Representing Next-Token Distributions (https://arxiv.org/abs/2506.17090) 27:18 Improving Language Model Personas via Rationalization with Psychological Scaffolds (https://ui.adsabs.harvard.edu/abs/202...) 35:21 OATH-Frames: Characterizing Online Attitudes Towards Homelessness with LLM Assistants (https://arxiv.org/abs/2406.14883) 46:52 Uncovering Intervention Opportunities for Suicide Prevention with Language Model Assistants (https://arxiv.org/abs/2508.18541) 🎧 Subscribe to stay updated on new episodes spotlighting brilliant women shaping the future of AI. WiAIR website: ♾️ https://women-in-ai-research.github.io Follow us at: ♾️ LinkedIn: / women-in-ai-research ♾️ Bluesky: https://bsky.app/profile/wiair.bsky.s... ♾️ X (Twitter): https://x.com/WiAIR_podcast #AISafety #LanguageModels #AIResearch #ResponsibleAI #WomenInAI #NLP #MachineLearning #AIAlignment #wiair #wiairpodcast