Update README.md
Browse files
README.md
CHANGED
@@ -707,6 +707,9 @@ model-index:
|
|
707 |
- type: max_f1
|
708 |
value: 78.14740888593184
|
709 |
---
|
|
|
|
|
|
|
710 |
This is the quantized (INT8) ONNX variant of the [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) embeddings model created with [DeepSparse Optimum](https://github.com/neuralmagic/optimum-deepsparse) for ONNX export/inference pipeline and Neural Magic's [Sparsify](https://github.com/neuralmagic/sparsify) for one-shot quantization.
|
711 |
|
712 |
Current list of sparse and quantized bge ONNX models:
|
|
|
707 |
- type: max_f1
|
708 |
value: 78.14740888593184
|
709 |
---
|
710 |
+
|
711 |
+
# bge-base-en-v1.5-quant
|
712 |
+
|
713 |
This is the quantized (INT8) ONNX variant of the [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) embeddings model created with [DeepSparse Optimum](https://github.com/neuralmagic/optimum-deepsparse) for ONNX export/inference pipeline and Neural Magic's [Sparsify](https://github.com/neuralmagic/sparsify) for one-shot quantization.
|
714 |
|
715 |
Current list of sparse and quantized bge ONNX models:
|