Chapter 10: Sequence Models & Attention

290 concepts 7 sections Prerequisites: Training Deep Networks
Chapter Overview

RNN/LSTM/GRU · Seq2Seq & Attention · Transformer architecture · BERT & GPT · State Space Models · Efficient architectures

Sections