Update README.md
Browse files
README.md
CHANGED
@@ -19,7 +19,7 @@ This repository contains pretrained ProtoViT models for interpretable image clas
|
|
19 |
|
20 |
## Model Description
|
21 |
|
22 |
-
ProtoViT combines Vision Transformers with prototype-based learning to create models that are both highly accurate and interpretable. Rather than functioning as a black box, ProtoViT learns interpretable prototypes that
|
23 |
|
24 |
### Supported Architectures
|
25 |
|
@@ -55,6 +55,11 @@ All models were trained and evaluated on the CUB-200-2011 fine-grained bird spec
|
|
55 |
- numpy
|
56 |
- pillow
|
57 |
|
|
|
|
|
|
|
|
|
|
|
58 |
## Citation
|
59 |
|
60 |
If you use this model in your research, please cite:
|
|
|
19 |
|
20 |
## Model Description
|
21 |
|
22 |
+
[ProtoViT](https://github.com/Henrymachiyu/ProtoViT) combines Vision Transformers with prototype-based learning to create models that are both highly accurate and interpretable. Rather than functioning as a black box, ProtoViT learns interpretable prototypes that explain its classification decisions through visual similarities.
|
23 |
|
24 |
### Supported Architectures
|
25 |
|
|
|
55 |
- numpy
|
56 |
- pillow
|
57 |
|
58 |
+
## Limitations and Bias
|
59 |
+
- Data Bias: These models are trained on CUB-200-2011, which may not generalize well to images outside this dataset.
|
60 |
+
- Resolution Constraints: The models are trained at a resolution of 224×224; higher or lower resolutions may impact performance.
|
61 |
+
- Location Misalignment: Same as the other CNN based models, these models are not perfectly immune to location misalignment under adversarial attack.
|
62 |
+
|
63 |
## Citation
|
64 |
|
65 |
If you use this model in your research, please cite:
|