When I load the pre-trained weights from ViT-base-patch16-384 (available on huggingface) to my model using load_state_dict(), I am getting the following error: size mismatch for pos_embed: copying a param with shape torch.Size ( [1, 577, 768]) from checkpoint, the shape in current model is torch.Size ( [1, 1601, 768]).