Tweeted By @ml_review
Why Self-Attention? A Targeted Evaluation of Neural Machine Translation Architectures
— ML Review (@ml_review) August 31, 2018
w/ @RicoSennrich
1) self-attentional & CNNs do not outperform RNNs in subject-verb agreement
2) self-attentional outperform RNNs & CNNs on word sense disambiguationhttps://t.co/TIvuYMUuLe pic.twitter.com/Og01mkqs5K