Mingxing Tan   @tanmingxing   6/10/2021       

Happy to introduce CoAtNet: combining convolution and self-attention in a principled way to obtain better capacity and better generalization. 88.56% top-1 with ImageNet21K (13M imgs), matching ViT-huge with JFT (300M imgs). Paper: https://t.co/AQE33LuzSr

 Reply  0     Retweet   81      Like   299





Posted by Mingxing Tan