Model Architecture Details

Architecture Overview

  • Architecture: ViT Small

Configuration

Attribute Value
Patch Size 32
Image Size 224
Num Layers 3
Attention Heads 4
Objective Function CrossEntropy

Performance

  • Validation Accuracy (Top 5): 0.4179
  • Validation Accuracy: 0.2148

Additional Resources

The model was trained using the library: ViT-Prisma.
For detailed metrics, plots, and further analysis of the model's training process, refer to the training report.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Dataset used to train Prisma-Multimodal/ImageNet-Small-Attention-and-MLP-Patch32