I cannot reproduce results finetuning on ImageNet1k after pretraining on ImageNet21k. How long does it usually take to pretrain on ImageNet21K with 8 V100 GPUs to get the same performance on ImageNet1k that you currently have finetuning with ViT-B/S/T?
From DeiT III: Revenge of the ViT (https://arxiv.org/pdf/2204.07118.pdf), how long exactly (exact number of hours) does it take to pretrain for 90 epochs on ImageNet21k with 8 V100 GPUs.