Tweeted By @hardmaru
A nice thread that explains intuitive why self-attention mechanism is needed in order to properly represent the relationships of the inputs—something that MLP’s struggle with. Looking forward to the next thread! https://t.co/s5Zrl6cjbH
— hardmaru (@hardmaru) January 18, 2022