Tweeted By @ak92501
Billion-Scale Pretraining with Vision Transformers for
— AK (@ak92501) August 13, 2021
Multi-Task Visual Representations
pdf: https://t.co/ZPTagL3LzO
abs: https://t.co/TfhdXimw4s
a scalable approach for pretraining with over a billion images in order to improve a production Unified Visual Embedding model pic.twitter.com/bFmlbpD01e