Transformers are RNNs: Fast Autoregressive Transformers with Linear Attention (Paper Explained)
Video Statistics and Information
Channel: Yannic Kilcher
Views: 19,459
Rating: undefined out of 5
Keywords: deep learning, machine learning, arxiv, explained, neural networks, ai, artificial intelligence, paper, nlp, natural language processing, attention, attention mechanism, linear, linear transformer, linformer, reformer, idiap, epfl, queries, keys, softmax, kernel, routing, inner product, rnn, recurrent neural network, transformer, bert, autoregressive, dimensions, topic modeling, language model
Id: hAooAOFRsYc
Channel Id: undefined
Length: 48min 5sec (2885 seconds)
Published: Sat Jul 04 2020
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.