Translation Models (2/3/2022)
Lecture: (by Graham Neubig)
- Sequence-to-sequence models w/ attention
- Decoding strategies
- Transformers
Language in 10: Luo
Slides: Translation Models Slides
Discussion: No discussion, but code walk of the annotated transformer and introduction of assignment 2.
References:
- Highly Recommended: The Annotated Transformer
- Reference: Attentional NMT (Bahdanau et al. 2015)
- Reference: Effective Approaches to Attention (Luong et al. 2015)
- Reference: Self Attention (Cheng et al. 2016)
- Reference: Attention is All You Need (Vaswani et al. 2017)
- Reference: Coverage Embedding Models (Mi et al. 2016)