Replies: 2 comments 3 replies
-
Face the same problem. I am concerning the reproducibility of these large models and thus would like to use the official implementations for research. |
Beta Was this translation helpful? Give feedback.
-
Do note that most of the models available in timm are not trained using this repository (though many rely on this repository to some extend, see, e.g., deit). As such, the training configurations to reproduce the results of a model using this repository do not in general exist. However, if you browse the source code for the models in this repository, you will find references to the papers and original code implementations. For example, if you want to access the code and paper for the original ViT paper, you can check out models/vision_transformer.py. This should give you all the information you need to reproduce the results of the respective papers. Further, you can find a number of configurations for this repository to reproduce results across many models at rigthman's gists. |
Beta Was this translation helpful? Give feedback.
-
Hi, I would like to know where could I find the default training recipe for different Transformer models like ViT, DeiT, ConViT, SwinViT, XCiT, such as the training epochs, learning rate, learning scheduler and other parameters. :)
Beta Was this translation helpful? Give feedback.
All reactions