xinhe commited on
Commit
2750b4f
1 Parent(s): 4974aaf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -0
README.md CHANGED
@@ -12,16 +12,22 @@ metrics:
12
  ---
13
 
14
  # INT8 T5 large finetuned on CNN DailyMail
 
15
  ### Post-training dynamic quantization
 
16
  This is an INT8 PyTorch model quantized with [huggingface/optimum-intel](https://github.com/huggingface/optimum-intel) through the usage of [Intel® Neural Compressor](https://github.com/intel/neural-compressor).
17
 
18
  The original fp32 model comes from the fine-tuned model [sysresearch101/t5-large-finetuned-xsum-cnn](https://huggingface.co/sysresearch101/t5-large-finetuned-xsum-cnn).
 
19
  ### Evaluation result
 
20
  | |INT8|FP32|
21
  |---|:---:|:---:|
22
  | **Accuracy (eval-rougeLsum)** | 29.6346 |29.7451|
23
  | **Model size** |879M|3021M|
 
24
  ### Load with optimum:
 
25
  ```python
26
  from optimum.intel.neural_compressor.quantization import IncQuantizedModelForSeq2SeqLM
27
  int8_model = IncQuantizedModelForSeq2SeqLM.from_pretrained(
 
12
  ---
13
 
14
  # INT8 T5 large finetuned on CNN DailyMail
15
+
16
  ### Post-training dynamic quantization
17
+
18
  This is an INT8 PyTorch model quantized with [huggingface/optimum-intel](https://github.com/huggingface/optimum-intel) through the usage of [Intel® Neural Compressor](https://github.com/intel/neural-compressor).
19
 
20
  The original fp32 model comes from the fine-tuned model [sysresearch101/t5-large-finetuned-xsum-cnn](https://huggingface.co/sysresearch101/t5-large-finetuned-xsum-cnn).
21
+
22
  ### Evaluation result
23
+
24
  | |INT8|FP32|
25
  |---|:---:|:---:|
26
  | **Accuracy (eval-rougeLsum)** | 29.6346 |29.7451|
27
  | **Model size** |879M|3021M|
28
+
29
  ### Load with optimum:
30
+
31
  ```python
32
  from optimum.intel.neural_compressor.quantization import IncQuantizedModelForSeq2SeqLM
33
  int8_model = IncQuantizedModelForSeq2SeqLM.from_pretrained(