benny1923 commited on
Commit
daacb81
·
1 Parent(s): 247bad4

upload model

Browse files
README.md CHANGED
@@ -1,3 +1,5 @@
1
  ---
2
  license: cc-by-nc-4.0
 
 
3
  ---
 
1
  ---
2
  license: cc-by-nc-4.0
3
+ tags:
4
+ - transformers.js
5
  ---
config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/metaclip-b16-fullcc2.5b",
3
+ "architectures": [
4
+ "CLIPModel"
5
+ ],
6
+ "initializer_factor": 1.0,
7
+ "logit_scale_init_value": 2.6592,
8
+ "model_type": "clip",
9
+ "projection_dim": 512,
10
+ "text_config": {
11
+ "heads": 8,
12
+ "layers": 12,
13
+ "model_type": "clip_text_model"
14
+ },
15
+ "transformers_version": "4.45.2",
16
+ "vision_config": {
17
+ "model_type": "clip_vision_model",
18
+ "patch_size": 16
19
+ }
20
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
onnx/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63fb5d5fadf18379e86ec8073044200b40d336668d206323c5eaaed612cef52b
3
+ size 599135855
onnx/model_quantized.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abe60cd9f449585416b090b57aee967cfd5a59bab59b9c342b6458f06af266e3
3
+ size 151988869
onnx/text_model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:350b52b451e87421331d19a24ed3cbe6c4b21b11a0934d623b7f9b8a8c47a017
3
+ size 254059280
onnx/text_model_quantized.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af182f98259dd7418b3d88ac85ed473fb0b9df5675a941b602ba0faaae4df243
3
+ size 64529343
onnx/vision_model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:962f4e8a124bb72311172478ac9a6bca90e92c046ad2f67b03184cc9ea8d849d
3
+ size 345075087
onnx/vision_model_quantized.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:342e70c0f41ef40e45063c38c0e812a48ab29d88db73191c220f0eac28017efd
3
+ size 87500784
preprocessor_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
+ "do_center_crop": true,
7
+ "do_convert_rgb": true,
8
+ "do_normalize": true,
9
+ "do_rescale": true,
10
+ "do_resize": true,
11
+ "image_mean": [
12
+ 0.48145466,
13
+ 0.4578275,
14
+ 0.40821073
15
+ ],
16
+ "image_processor_type": "CLIPFeatureExtractor",
17
+ "image_std": [
18
+ 0.26862954,
19
+ 0.26130258,
20
+ 0.27577711
21
+ ],
22
+ "processor_class": "CLIPProcessor",
23
+ "resample": 3,
24
+ "rescale_factor": 0.00392156862745098,
25
+ "size": {
26
+ "shortest_edge": 224
27
+ }
28
+ }
quantize_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "per_channel": true,
3
+ "reduce_range": true,
4
+ "per_model_config": {
5
+ "vision_model": {
6
+ "op_types": [
7
+ "Add",
8
+ "Slice",
9
+ "Sub",
10
+ "Where",
11
+ "Shape",
12
+ "Reshape",
13
+ "MatMul",
14
+ "ConstantOfShape",
15
+ "Concat",
16
+ "Sqrt",
17
+ "Pow",
18
+ "Gather",
19
+ "Softmax",
20
+ "Div",
21
+ "Constant",
22
+ "Equal",
23
+ "Conv",
24
+ "Expand",
25
+ "Unsqueeze",
26
+ "Mul",
27
+ "Transpose",
28
+ "Cast",
29
+ "Sigmoid",
30
+ "ReduceMean"
31
+ ],
32
+ "weight_type": "QUInt8"
33
+ },
34
+ "text_model": {
35
+ "op_types": [
36
+ "Add",
37
+ "Slice",
38
+ "ArgMax",
39
+ "Sub",
40
+ "Where",
41
+ "Shape",
42
+ "Squeeze",
43
+ "Reshape",
44
+ "Less",
45
+ "MatMul",
46
+ "ConstantOfShape",
47
+ "Concat",
48
+ "Sqrt",
49
+ "Range",
50
+ "Pow",
51
+ "Gather",
52
+ "Softmax",
53
+ "Flatten",
54
+ "Div",
55
+ "Constant",
56
+ "Equal",
57
+ "Expand",
58
+ "Unsqueeze",
59
+ "Mul",
60
+ "Transpose",
61
+ "Cast",
62
+ "Sigmoid",
63
+ "ReduceMean"
64
+ ],
65
+ "weight_type": "QInt8"
66
+ },
67
+ "model": {
68
+ "op_types": [
69
+ "Add",
70
+ "Slice",
71
+ "ArgMax",
72
+ "Sub",
73
+ "Where",
74
+ "Shape",
75
+ "Squeeze",
76
+ "Reshape",
77
+ "Less",
78
+ "MatMul",
79
+ "ConstantOfShape",
80
+ "Concat",
81
+ "Sqrt",
82
+ "Range",
83
+ "Pow",
84
+ "Gather",
85
+ "Exp",
86
+ "Softmax",
87
+ "Flatten",
88
+ "Div",
89
+ "ReduceSum",
90
+ "Constant",
91
+ "Equal",
92
+ "Conv",
93
+ "Expand",
94
+ "Unsqueeze",
95
+ "Mul",
96
+ "Transpose",
97
+ "Cast",
98
+ "Sigmoid",
99
+ "ReduceMean"
100
+ ],
101
+ "weight_type": "QUInt8"
102
+ }
103
+ }
104
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<|endoftext|>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "49406": {
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "49407": {
13
+ "content": "<|endoftext|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ }
20
+ },
21
+ "additional_special_tokens": [],
22
+ "bos_token": "<|startoftext|>",
23
+ "clean_up_tokenization_spaces": true,
24
+ "do_lower_case": true,
25
+ "eos_token": "<|endoftext|>",
26
+ "errors": "replace",
27
+ "model_max_length": 77,
28
+ "pad_token": "<|endoftext|>",
29
+ "processor_class": "CLIPProcessor",
30
+ "tokenizer_class": "CLIPTokenizer",
31
+ "unk_token": "<|endoftext|>"
32
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff