Do we need Attention? - Linear RNNs and State Space Models (SSMs) for NLP Published 2023-07-05 Download video MP4 360p Recommendations 1:06:35 MedAI #41: Efficiently Modeling Long Sequences with Structured State Spaces | Albert Gu 33:50 Do we need Attention? A Mamba Primer 27:14 But what is a GPT? Visual intro to Transformers | Chapter 5, Deep Learning 1:02:17 RWKV: Reinventing RNNs for the Transformer Era (Paper Explained) 36:16 The math behind Attention: Keys, Queries, and Values matrices 1:28:14 Talk 1: Nuts and Bolts of Modern State Space Models - Part I 57:19 Efficiently Modeling Long Sequences with Structured State Spaces - Albert Gu | Stanford MLSys #46 40:08 The Most Important Algorithm in Machine Learning 16:01 Mamba - a replacement for Transformers? 48:06 Transformers are RNNs: Fast Autoregressive Transformers with Linear Attention (Paper Explained) 43:26 xLSTM: Extended Long Short-Term Memory 40:40 Mamba: Linear-Time Sequence Modeling with Selective State Spaces (Paper Explained) 13:37 What are Transformer Models and How do they Work? 31:51 MAMBA from Scratch: Neural Nets Better and Faster than Transformers 45:46 Geoffrey Hinton | On working with Ilya, choosing problems, and the power of intuition 1:04:28 Structured State Space Models for Deep Sequence Modeling (Albert Gu, CMU) 59:48 [1hr Talk] Intro to Large Language Models 18:17 Can we reach AGI with just LLMs? 1:02:50 MIT 6.S191 (2023): Recurrent Neural Networks, Transformers, and Attention Similar videos 39:43 [한글자막] Do we need Attention? Linear RNNs and State Space Models SSMs for NLP 03:23 Data Science & Machine Learning: State-Space Models vs RNNs 45:48 MAMBA AI (S6): Better than Transformers? 05:29 Natural Language Processing In 5 Minutes | What Is NLP And How Does It Work? | Simplilearn 12:45 RNNs for Diffusion? Generating Images with DiffuSSM 1:19:06 Hardware-aware Algorithms for Sequence Modeling - Tri Dao | Stanford MLSys #87 14:06 Mamba Might Just Make LLMs 1000x Cheaper... 1:15:33 CMU Advanced NLP 2021 (7): Attention 23:03 Deep Learning Foundations by Soheil Feizi : Linear Attention 13:05 State-Space Models, Control, Time Series Analysis (Episode 19) More results