openvino-ci commited on
Commit
3bdd55e
1 Parent(s): 68d603b

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,90 +1,54 @@
1
  ---
2
  license: mit
3
- language:
4
- - en
5
  ---
6
-
7
  # dolly-v2-3b-fp16-ov
8
-
9
- * Model creator: [Databricks](https://huggingface.co/databricks)
10
  * Original model: [dolly-v2-3b](https://huggingface.co/databricks/dolly-v2-3b)
11
 
12
  ## Description
13
 
14
- This is [dolly-v2-3b](https://huggingface.co/databricks/dolly-v2-3b) model converted to the [OpenVINO™ IR](https://docs.openvino.ai/2024/documentation/openvino-ir-format.html) (Intermediate Representation) format.
15
-
16
  ## Compatibility
17
 
18
  The provided OpenVINO™ IR model is compatible with:
19
 
20
- * OpenVINO version 2024.2.0 and higher
21
- * Optimum Intel 1.17.0 and higher
22
 
23
- ## Running Model Inference with [Optimum Intel](https://huggingface.co/docs/optimum/intel/index)
24
 
25
  1. Install packages required for using [Optimum Intel](https://huggingface.co/docs/optimum/intel/index) integration with the OpenVINO backend:
26
 
27
- ```
28
- pip install optimum[openvino]
29
- ```
30
-
31
- 2. Run model inference:
32
-
33
- ```
34
- from transformers import AutoTokenizer
35
- from optimum.intel.openvino import OVModelForCausalLM
36
-
37
- model_id = "OpenVINO/dolly-v2-3b-fp16-ov"
38
- tokenizer = AutoTokenizer.from_pretrained(model_id)
39
- model = OVModelForCausalLM.from_pretrained(model_id)
40
-
41
- inputs = tokenizer("What is OpenVINO?", return_tensors="pt")
42
-
43
- outputs = model.generate(**inputs, max_length=200)
44
- text = tokenizer.batch_decode(outputs)[0]
45
- print(text)
46
- ```
47
-
48
- For more examples and possible optimizations, refer to the [OpenVINO Large Language Model Inference Guide](https://docs.openvino.ai/2024/learn-openvino/llm_inference_guide.html).
49
-
50
- ## Running Model Inference with [OpenVINO GenAI](https://github.com/openvinotoolkit/openvino.genai)
51
-
52
- 1. Install packages required for using OpenVINO GenAI.
53
  ```
54
- pip install openvino-genai huggingface_hub
55
- ```
56
-
57
- 2. Download model from HuggingFace Hub
58
-
59
  ```
60
- import huggingface_hub as hf_hub
61
 
62
- model_id = "OpenVINO/dolly-v2-3b-fp16-ov""
63
- model_path = "dolly-v2-3b-fp16-ov""
64
-
65
- hf_hub.snapshot_download(model_id, local_dir=model_path)
66
 
67
  ```
 
 
68
 
69
- 3. Run model inference:
 
 
70
 
71
- ```
72
- import openvino_genai as ov_genai
73
 
74
- device = "CPU"
75
- pipe = ov_genai.LLMPipeline(model_path, device)
76
- print(pipe.generate("What is OpenVINO?", max_length=200))
77
  ```
78
 
79
- More GenAI usage examples can be found in OpenVINO GenAI library [docs](https://github.com/openvinotoolkit/openvino.genai/blob/master/src/README.md) and [samples](https://github.com/openvinotoolkit/openvino.genai?tab=readme-ov-file#openvino-genai-samples)
80
 
81
  ## Limitations
82
 
83
- Check the original model card for [limitations](https://huggingface.co/databricks/dolly-v2-3b#known-limitations).
84
 
85
  ## Legal information
86
 
87
- The original model is distributed under [MIT](https://choosealicense.com/licenses/mit/) license. More details can be found in [original model card](https://huggingface.co/databricks/dolly-v2-3b).
88
 
89
  ## Disclaimer
90
 
 
1
  ---
2
  license: mit
3
+ license_link: https://choosealicense.com/licenses/mit/
 
4
  ---
 
5
  # dolly-v2-3b-fp16-ov
6
+ * Model creator: [Databricks](https://huggingface.co/databricks)
 
7
  * Original model: [dolly-v2-3b](https://huggingface.co/databricks/dolly-v2-3b)
8
 
9
  ## Description
10
 
 
 
11
  ## Compatibility
12
 
13
  The provided OpenVINO™ IR model is compatible with:
14
 
15
+ * OpenVINO version 2024.4.0 and higher
16
+ * Optimum Intel 1.20.0 and higher
17
 
18
+ ## Running Model Inference
19
 
20
  1. Install packages required for using [Optimum Intel](https://huggingface.co/docs/optimum/intel/index) integration with the OpenVINO backend:
21
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
  ```
23
+ pip install optimum[openvino]
 
 
 
 
24
  ```
 
25
 
26
+ 2. Run model inference:
 
 
 
27
 
28
  ```
29
+ from transformers import AutoTokenizer
30
+ from optimum.intel.openvino import OVModelForCausalLM
31
 
32
+ model_id = "OpenVINO/dolly-v2-3b-fp16-ov"
33
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
34
+ model = OVModelForCausalLM.from_pretrained(model_id)
35
 
36
+ inputs = tokenizer("What is OpenVINO?", return_tensors="pt")
 
37
 
38
+ outputs = model.generate(**inputs, max_length=200)
39
+ text = tokenizer.batch_decode(outputs)[0]
40
+ print(text)
41
  ```
42
 
43
+ For more examples and possible optimizations, refer to the [OpenVINO Large Language Model Inference Guide](https://docs.openvino.ai/2024/learn-openvino/llm_inference_guide.html).
44
 
45
  ## Limitations
46
 
47
+ Check the original model card for [original model card](https://huggingface.co/databricks/dolly-v2-3b) for limitations.
48
 
49
  ## Legal information
50
 
51
+ The original model is distributed under [mit](https://choosealicense.com/licenses/mit/) license. More details can be found in [original model card](https://huggingface.co/databricks/dolly-v2-3b).
52
 
53
  ## Disclaimer
54
 
config.json CHANGED
@@ -25,13 +25,15 @@
25
  "model_type": "gpt_neox",
26
  "num_attention_heads": 32,
27
  "num_hidden_layers": 32,
 
28
  "rope_scaling": null,
 
29
  "rotary_emb_base": 10000,
30
  "rotary_pct": 0.25,
31
  "tie_word_embeddings": false,
32
- "torch_dtype": "float32",
33
- "transformers_version": "4.40.1",
34
  "use_cache": true,
35
  "use_parallel_residual": true,
36
  "vocab_size": 50280
37
- }
 
25
  "model_type": "gpt_neox",
26
  "num_attention_heads": 32,
27
  "num_hidden_layers": 32,
28
+ "partial_rotary_factor": 0.25,
29
  "rope_scaling": null,
30
+ "rope_theta": 10000,
31
  "rotary_emb_base": 10000,
32
  "rotary_pct": 0.25,
33
  "tie_word_embeddings": false,
34
+ "torch_dtype": "float16",
35
+ "transformers_version": "4.45.2",
36
  "use_cache": true,
37
  "use_parallel_residual": true,
38
  "vocab_size": 50280
39
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "eos_token_id": 0,
5
+ "transformers_version": "4.45.2"
6
+ }
openvino_detokenizer.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e3d0218341805b3876fc9c8e95c98d75cc9ddd0fa34fac6df212e790b6f91a08
3
- size 558494
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1e43770f23d5b9dbfc8bf99bbea4fe501870adf36235dff20156f6c0a129a47
3
+ size 514078
openvino_detokenizer.xml CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7675447cb4bf5c86e1f08f505dd3df3616bc5e96c5e67569836fed97d6cac47
3
- size 5981
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:978a0d7f70294bcc62041496c85f3988aa2198f81697b338896a3a12b933b1a4
3
+ size 4498
openvino_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a7e96cb1c20f52ead1cb0280485e2b1588b96f726c6ae9f1c66d9d10b07def2
3
- size 5554525594
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea2579c5c7a6ae1e1bf0afa54c363aa231816b6ca4b88859a004a00ea9e6c604
3
+ size 5550172420
openvino_model.xml CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:57f0715d3d331f91a95996642aedf6774aeea993c3b4a19f8599710de5e0df2b
3
- size 3581222
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:051b8420c5a95133352fc66b1db055d33ce5f6695feed62255572221bfc0f075
3
+ size 2383298
openvino_tokenizer.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf4bb4307e428d32680d2973f34247c51210fad1f4c6408b2087dfdfb053e210
3
- size 1166376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c378d88077ae7c7e13ef61745e1ceef76412338e9a7398445c09632413e52abe
3
+ size 1227935
openvino_tokenizer.xml CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c7cdfc7122104652c01e43777808b3c7c4670445cceaf2976554adc694d9064
3
- size 27473
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a72a76f00a28f33541aa4d9bda2a9559ace3f5fa60cfa6d7556e072c4cc0a4c
3
+ size 22293
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -234,7 +234,7 @@
234
  "### Response:"
235
  ],
236
  "bos_token": "<|endoftext|>",
237
- "clean_up_tokenization_spaces": true,
238
  "eos_token": "<|endoftext|>",
239
  "model_max_length": 1000000000000000019884624838656,
240
  "pad_token": "<|endoftext|>",
 
234
  "### Response:"
235
  ],
236
  "bos_token": "<|endoftext|>",
237
+ "clean_up_tokenization_spaces": false,
238
  "eos_token": "<|endoftext|>",
239
  "model_max_length": 1000000000000000019884624838656,
240
  "pad_token": "<|endoftext|>",