Homepage
Close
Menu

Site Navigation

  • Home
  • Archive(TODO)
    • By Day
    • By Month
  • About(TODO)
  • Stats
Close
by ylecun on 2022-10-10 (UTC).

MaxViT : combines ConvNet modules and 2 types of self attention (local n'y block, and on a subsampled grid).
Since DETR (hi @alcinos26 !), I've become convinced that combining Conv and attention/dynamic routing was the Right Thing. https://t.co/DNOBsqL54Z

— Yann LeCun (@ylecun) October 10, 2022
cvresearch
by martin_gorner on 2022-10-11 (UTC).

This looks like the Vision Transformers architecture we have been waiting for: MaxViT https://t.co/WbzgJ50PjB
1/ State of the Art accuracy on ImageNet (no pre-training on huge datasets)
2/ Linear complexity wrt. image size (thanks to a clever attention design) pic.twitter.com/5bW0N7n3s5

— Martin Görner (@martin_gorner) October 11, 2022
researchcv

Tags

learning tutorial misc nlp rstats gan ethics research dataviz survey python tool security kaggle video thought bayesian humour tensorflow w_code bias dataset pytorch cv tip application javascript forecast swift golang rl jax julia gnn causal surey diffusion
© Copyright Philosophy 2018 Site Template by Colorlib