AashishKumar commited on
Commit
6569276
·
verified ·
1 Parent(s): e6a4626

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +121 -0
README.md ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - Hemg/cifake-real-and-ai-generated-synthetic-images
5
+ language:
6
+ - en
7
+ metrics:
8
+ - accuracy
9
+ library_name: transformers
10
+ tags:
11
+ - Diffusors
12
+ - GanDetectors
13
+ - Cifake
14
+ ---
15
+ # Model Card for Model ID
16
+
17
+ <!-- Provide a quick summary of what the model is/does. -->
18
+
19
+ This model card provides comprehensive information about the model's architecture, training data, evaluation metrics, and environmental impact.
20
+
21
+ ## Model Details
22
+
23
+ ### Model Description
24
+
25
+ This model is a pre-trained model for image classification, specifically designed for detecting fake images, including both real and AI-generated synthetic images. It utilizes the ViT (Vision Transformer) architecture for image classification tasks.
26
+
27
+ - **Developed by:** [Author(s) Name(s)]
28
+ - **Funded by [optional]:** [Funding Source(s)]
29
+ - **Shared by [optional]:** [Organization/Individual(s) Sharing the Model]
30
+ - **Model type:** Vision Transformer (ViT)
31
+ - **Language(s) (NLP):** N/A
32
+ - **License:** Apache License 2.0
33
+ - **Finetuned from model [optional]:** [Base Pre-trained Model]
34
+
35
+ ### Model Sources [optional]
36
+
37
+ - **Repository:** https://github.com/AashishKumar-3002/AIGuardVision.git
38
+
39
+ ## Uses
40
+
41
+ ### Direct Use
42
+
43
+ This model can be directly used for classifying images as real or AI-generated synthetic images.
44
+
45
+ ### Downstream Use [optional]
46
+
47
+ This model can be fine-tuned for specific image classification tasks related to detecting fake images in various domains.
48
+
49
+ ### Out-of-Scope Use
50
+
51
+ The model may not perform well on tasks outside the scope of image classification, such as object detection or segmentation.
52
+
53
+ ## Bias, Risks, and Limitations
54
+
55
+ The model's performance may be influenced by biases in the training data, leading to potential inaccuracies in classification.
56
+
57
+ ### Recommendations
58
+
59
+ Users should be aware of potential biases and limitations when using the model for classification tasks, and additional data sources may be necessary to mitigate biases.
60
+
61
+ ## How to Get Started with the Model
62
+
63
+ Use the code below to get started with the model:
64
+
65
+ [Code Snippet for Model Usage]
66
+
67
+ ## Training Details
68
+
69
+ ### Training Data
70
+
71
+ The model was trained on the CIFake dataset, which contains real and AI-generated synthetic images for training the classification model.
72
+
73
+ ### Training Procedure
74
+
75
+ #### Preprocessing [optional]
76
+
77
+ Data preprocessing techniques were applied to the training data, including normalization and data augmentation to improve model generalization.
78
+
79
+ #### Training Hyperparameters
80
+
81
+ - **Training regime:** Fine-tuning with a learning rate of 0.0000001
82
+ - **Batch Size:** 64
83
+ - **Epochs:** 100
84
+
85
+ #### Speeds, Sizes, Times [optional]
86
+
87
+ - **Training Time:** 1 hr 36 min
88
+
89
+ ## Evaluation
90
+
91
+ ### Testing Data, Factors & Metrics
92
+
93
+ #### Testing Data
94
+
95
+ The model was evaluated on a separate test set from the CIFake dataset.
96
+
97
+ #### Factors
98
+
99
+ The evaluation considered factors such as class imbalance and dataset diversity.
100
+
101
+ #### Metrics
102
+
103
+ Evaluation metrics included accuracy, precision, recall, and F1-score.
104
+
105
+ ### Results
106
+
107
+ The model achieved an accuracy of [Accuracy] on the test set, with detailed metrics summarized in the following table:
108
+
109
+ [Metrics Table]
110
+
111
+ ## Model Examination [optional]
112
+
113
+ <!-- Relevant interpretability work for the model goes here -->
114
+
115
+ [Information on Model Examination, if available]
116
+
117
+ ## Technical Specifications [optional]
118
+
119
+ ### Model Architecture and Objective
120
+
121
+ The model architecture is based on the Vision Transformer (ViT) architecture, which uses self-attention mechanisms for image classification tasks.