У нас вы можете посмотреть бесплатно Mac Studio M3 Ultra: what size LLMs will it be able to run and how fast? или скачать в максимальном доступном качестве, которое было загружено на ютуб. Для скачивания выберите вариант из формы ниже:
Если кнопки скачивания не
загрузились
НАЖМИТЕ ЗДЕСЬ или обновите страницу
Если возникают проблемы со скачиванием, пожалуйста напишите в поддержку по адресу внизу
страницы.
Спасибо за использование сервиса ClipSaver.ru
The Mac Studio M3 Ultra just landed in 2025! It will feature up to 512 GB of RAM and 800 GB/s memory bandwidth. In this video, I make some predictions about what the size of LLM will be which you can fit into this machine, and what speeds you will be able to run them at. Comment from the future, whether my predictions were right or not. Spoiler alert: do not get your hopes up too high for really big models like DeepSeek R1 natively and at full model size - it will not be able to fit into the RAM without quantization! Also, I predict that DeepSeek R1 will not run at decent speeds (10+ token/s) - otherwise I promise to publicly eat my hat here on this channel. Use the LLM RAM Calculator to estimate the size of your LLMs according with the size of RAM you have: https://llm-calc.rayfernando.ai/?ram=512 This useful article charts different inference options you have today, including Apple silicon: https://www.hardware-corner.net/guide... llama.cpp on Apple Silicon M-series inference performance: https://github.com/ggml-org/llama.cpp... Mac Studio 2022 (M1) specs: https://support.apple.com/en-us/111900 Mac Studio 2023 (M2) specs: https://support.apple.com/en-us/111835 Mac Studio 2025 (M3 Ultra / M4 Max) specs: https://www.apple.com/mac-studio/specs/ Exo (Clustering Macs together for inference): https://github.com/exo-explore/exo Let me know in the comments what burning questions you want to get answered! Subscribe to my channel for more tips on AI for managers, entrepreneurs and business people, upcoming AI tools which will save you time and make you money.