Self Attention
Simple RNN + Self Attention
Simple RNN:
Simple RNN + Self Attention:
Calculate Weights:
Summary
With self-attention, RNN is less likely to forget.
Pay attention to the context relevant to the new input.
Reference
Cheng, Dong, & Lapata. Long Short-Term Memory-Networks for Machine Reading. In EMNLP, 2016.
Last updated