tahayf commited on
Commit
4bc732f
1 Parent(s): 1c9704f

Upload 3 files

Browse files
Files changed (3) hide show
  1. config.json +44 -0
  2. model.safetensors +3 -0
  3. preprocessor_config.json +27 -0
config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/clip-vit-base-patch16",
3
+ "architectures": [
4
+ "CLIPForImageClassification"
5
+ ],
6
+ "id2label": {
7
+ "0": "LABEL_0",
8
+ "1": "LABEL_1",
9
+ "2": "LABEL_2",
10
+ "3": "LABEL_3",
11
+ "4": "LABEL_4",
12
+ "5": "LABEL_5",
13
+ "6": "LABEL_6",
14
+ "7": "LABEL_7"
15
+ },
16
+ "initializer_factor": 1.0,
17
+ "label2id": {
18
+ "LABEL_0": 0,
19
+ "LABEL_1": 1,
20
+ "LABEL_2": 2,
21
+ "LABEL_3": 3,
22
+ "LABEL_4": 4,
23
+ "LABEL_5": 5,
24
+ "LABEL_6": 6,
25
+ "LABEL_7": 7
26
+ },
27
+ "logit_scale_init_value": 2.6592,
28
+ "model_type": "clip",
29
+ "problem_type": "single_label_classification",
30
+ "projection_dim": 512,
31
+ "text_config": {
32
+ "bos_token_id": 0,
33
+ "dropout": 0.0,
34
+ "eos_token_id": 2,
35
+ "model_type": "clip_text_model"
36
+ },
37
+ "torch_dtype": "float32",
38
+ "transformers_version": "4.45.2",
39
+ "vision_config": {
40
+ "dropout": 0.0,
41
+ "model_type": "clip_vision_model",
42
+ "patch_size": 16
43
+ }
44
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:512c87864d1c393d1876c2c9096d3139b7bb1faf1a7c5970ef4c2c9eaf48015a
3
+ size 343246272
preprocessor_config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
+ "do_center_crop": true,
7
+ "do_convert_rgb": true,
8
+ "do_normalize": true,
9
+ "do_rescale": true,
10
+ "do_resize": true,
11
+ "image_mean": [
12
+ 0.48145466,
13
+ 0.4578275,
14
+ 0.40821073
15
+ ],
16
+ "image_processor_type": "CLIPImageProcessor",
17
+ "image_std": [
18
+ 0.26862954,
19
+ 0.26130258,
20
+ 0.27577711
21
+ ],
22
+ "resample": 3,
23
+ "rescale_factor": 0.00392156862745098,
24
+ "size": {
25
+ "shortest_edge": 224
26
+ }
27
+ }