У нас вы можете посмотреть бесплатно EP001: How Transformers Smashed the Sequential Bottleneck или скачать в максимальном доступном качестве, видео которое было загружено на ютуб. Для загрузки выберите вариант из формы ниже:
Если кнопки скачивания не
загрузились
НАЖМИТЕ ЗДЕСЬ или обновите страницу
Если возникают проблемы со скачиванием видео, пожалуйста напишите в поддержку по адресу внизу
страницы.
Спасибо за использование сервиса ClipSaver.ru
Attention is All You Need (https://proceedings.neurips.cc/paper_...) • The Shift to Transformers: The overview discusses the move away from complex recurrent and convolutional neural networks toward the Transformer architecture, which relies entirely on attention mechanisms to draw global dependencies between inputs and outputs,. • Self-Attention & Multi-Head Attention: It explains how the model uses self-attention to relate different positions of a single sequence and multi-head attention to simultaneously attend to information from different representation subspaces,,. • Efficiency and Parallelization: A major highlight is the Transformer's ability to allow for significantly more parallelization during training, leading to a new state of the art in translation quality after training for a fraction of the time required by previous models,. • Architectural Components: The summary breaks down the encoder-decoder structure, the use of positional encodings to account for sequence order without recurrence, and the application of point-wise feed-forward networks,,. • State-of-the-Art Results: Finally, it reviews the impressive results on English-to-German and English-to-French translation tasks, where the Transformer outperformed existing best results and established new benchmarks