У нас вы можете посмотреть бесплатно Hacking AI is TOO EASY (this should be illegal) или скачать в максимальном доступном качестве, видео которое было загружено на ютуб. Для загрузки выберите вариант из формы ниже:
Если кнопки скачивания не
загрузились
НАЖМИТЕ ЗДЕСЬ или обновите страницу
Если возникают проблемы со скачиванием видео, пожалуйста напишите в поддержку по адресу внизу
страницы.
Спасибо за использование сервиса ClipSaver.ru
Want to deploy AI in your cloud apps SAFELY? Let Wiz help: https://ntck.co/wiz Can you hack AI? In this video I sit down with elite AI hacker Jason Haddix to unpack how attackers compromise AI-enabled apps—not just jailbreak chatbots, but exfiltrate customer data, abuse tool calls, and pivot across systems. We walk through his six-part AI pentest blueprint, play the Gandalf prompt-injection game, and demo wild techniques like emoji smuggling and link smuggling. You’ll see real-world cases (think Slack salesbots + Salesforce leaks), why MCP (Model Context Protocol) and agentic frameworks can widen the blast radius, and then we flip to defense: web-layer fundamentals, a “firewall for AI” on inputs/outputs, and least-privilege for data and tools—plus a hands-on demo you can try. If you’re building with AI in 2025, this is your wake-up call (and your roadmap). Educational content only—hack ethically and only with permission. 📌 Watch the Full Interview here: • The AI Attack Blueprint (Interview with Ja... Links and STUFF —-------------------------------------------------------- Practice Prompt Injection: https://gandalf.lakera.ai/baseline Pliney's Github: https://github.com/elder-plinius Follow Jason Everywhere: X: https://x.com/Jhaddix Linkedin: / jhaddix Instagram: / j.haddix56 Tiktok: / jhaddix56 Checkout Jason’s courses: Website: https://www.arcanum-sec.com/ Training Overview: Training: https://www.arcanum-sec.com/training-... Attacking AI course: https://www.arcanum-sec.com/training/... Hacking your career: https://www.arcanum-sec.com/training/... 🔥🔥Join the NetworkChuck Academy!: https://ntck.co/NCAcademy **Sponsored by Wiz.io 00:00 - Hack companies through AI? 00:58 - What does “hacking AI” really mean? 01:43 - AI pentest vs. red teaming (6-step blueprint) 02:42 - Prompt Injection 101 (why it’s so hard) 04:14 - Try it live: Gandalf prompt-injection game 05:09 - Jailbreak taxonomy: intents, techniques, evasions 05:55 - Emoji smuggling + anti-classifier demo 07:23 - Link smuggling (data exfiltration trick) 11:38 - Real-world leaks: Salesforce/Slack bot case 13:47 - MCP security risks & blast radius 16:55 - Can AI hack for us? Agents & bug bounties 20:52 - Defense in depth: web, AI firewall, least privilege 24:57 - Jason’s Magic Card: GPT-4o system prompt leak (wild story) SUPPORT NETWORKCHUCK --------------------------------------------------- ➡️NetworkChuck membership: https://ntck.co/Premium ☕☕ COFFEE and MERCH: https://ntck.co/coffee Check out my new channel: https://ntck.co/ncclips 🆘🆘NEED HELP?? Join the Discord Server: / discord STUDY WITH ME on Twitch: https://bit.ly/nc_twitch READY TO LEARN?? --------------------------------------------------- -Learn Python: https://bit.ly/3rzZjzz -Get your CCNA: https://bit.ly/nc-ccna FOLLOW ME EVERYWHERE --------------------------------------------------- Instagram: / networkchuck Twitter: / networkchuck Facebook: / networkchuck Join the Discord server: http://bit.ly/nc-discord AFFILIATES & REFERRALS --------------------------------------------------- (GEAR I USE...STUFF I RECOMMEND) My network gear: https://geni.us/L6wyIUj Amazon Affiliate Store: https://www.amazon.com/shop/networkchuck Buy a Raspberry Pi: https://geni.us/aBeqAL Do you want to know how I draw on the screen?? Go to https://ntck.co/EpicPen and use code NetworkChuck to get 20% off!! fast and reliable unifi in the cloud: https://hostifi.com/?via=chuck Prompt Injection explained with live demos: Gandalf game, emoji smuggling, and link smuggling exfiltration. AI Pentesting vs AI Red Teaming: a six-phase methodology for securing LLM apps end-to-end. LLM jailbreak taxonomy: intents, techniques, evasions, and utilities—how attackers actually think. RAG poisoning, tool-call abuse, and over-scoped API keys: the hidden risks in modern AI products. MCP (Model Context Protocol) security: tools/resources/prompts, server hardening, and blast-radius control. Agentic frameworks (LangChain, LangGraph, CrewAI) security pitfalls—and how to test them safely. Real-world case study: Slack salesbot + Salesforce data exposure and what went wrong. Defense in depth for AI: input/output validation, a firewall for AI (guardrails/classifiers), least privilege. Bug bounty + AI: why mid-tier vulns are getting automated while human creativity still wins. 2025 AI security blueprint: map your attack surface, prevent system-prompt leaks, and lock down data access. #promptinjection #aihacking #airedteaming