-
Thanks for making the pre-trained weights available. I understand that the converted weights for ViT-B and Vit-L are pretrained using ImageNet-21K. However for the weights provided for Vit-Small, are they pretrained using ImageNet-1K or ImageNet-21K ? |
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 2 replies
-
@aelnouby the converted weights are pretrained on ImageNet-21k and finetuned on ImageNet-1k by the Google group that wrote the paper / released official models. My 'small' model def was a reduced size model I threw together that was more practical to train on 2xGPU setup than the official 'base' model. It was just trained on ImageNet-1k from scratch, but achieved results comparable to the official base model training results on pure ImageNet-1k mentioned in the paper. I used heavier augmentation. |
Beta Was this translation helpful? Give feedback.
-
@aelnouby, @mmoayeri Did you happen to find any VIT (preferably VIT-B) with pretrained weights from Imagenet1k-only (not 21k+finetune)? Thanks! |
Beta Was this translation helpful? Give feedback.
@aelnouby the converted weights are pretrained on ImageNet-21k and finetuned on ImageNet-1k by the Google group that wrote the paper / released official models.
My 'small' model def was a reduced size model I threw together that was more practical to train on 2xGPU setup than the official 'base' model. It was just trained on ImageNet-1k from scratch, but achieved results comparable to the official base model training results on pure ImageNet-1k mentioned in the paper. I used heavier augmentation.