Reproducing ViT results fine tune imagenet 1K #1174
Unanswered
shairoz-deci
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm trying to train ViT_base_16 on imagenet 1K from the imagenet21K pretraining checkpoint and can't reproduce results.
The recipe from the paper should be batch=512, 20K steps, 500 warmup steps, lr=0.01, cosine scheduler and grad_clip of 1 which is similar to running:
on 8 gpus with the exception of the warmup which I resolved locally.
The pretrained checkpoint for vit_base_patch16_224 reaches an accuracy of 84.4 while the training above reaches 83.65
Did anyone manage to reproduce the reported accuracy of ViT?
Beta Was this translation helpful? Give feedback.
All reactions