Tweeted By @ak92501
Less is More: Pay Less Attention in Vision Transformers
— AK (@ak92501) June 1, 2021
pdf: https://t.co/ydo2bFvxsH
abs: https://t.co/baTSDrBpEd
hierarchical vision transformer pays less attention in
early stages to ease huge computational cost of self-attention modules over high-resolution
representations pic.twitter.com/6J5xdAO0mc