--- base_model: - openai/clip-vit-large-patch14 datasets: - mnist metrics: - accuracy --- # Model Card ## Model Details - Architecture: ViT-Large with patch size 14 - Training Data: MNIST dataset ## Training Details Adam Optimizer with a constant learning rate 1e-5 for 4000 steps training (batch_size=32). Only the vision encoder is fine-tuned. ## Evaluation Results - pre-trained: 0.7602328658103943 - fine-tuned: 0.9975429177284241