Update README.md
Browse files
README.md
CHANGED
|
@@ -707,6 +707,9 @@ model-index:
|
|
| 707 |
- type: max_f1
|
| 708 |
value: 78.14740888593184
|
| 709 |
---
|
|
|
|
|
|
|
|
|
|
| 710 |
This is the quantized (INT8) ONNX variant of the [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) embeddings model created with [DeepSparse Optimum](https://github.com/neuralmagic/optimum-deepsparse) for ONNX export/inference pipeline and Neural Magic's [Sparsify](https://github.com/neuralmagic/sparsify) for one-shot quantization.
|
| 711 |
|
| 712 |
Current list of sparse and quantized bge ONNX models:
|
|
|
|
| 707 |
- type: max_f1
|
| 708 |
value: 78.14740888593184
|
| 709 |
---
|
| 710 |
+
|
| 711 |
+
# bge-base-en-v1.5-quant
|
| 712 |
+
|
| 713 |
This is the quantized (INT8) ONNX variant of the [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) embeddings model created with [DeepSparse Optimum](https://github.com/neuralmagic/optimum-deepsparse) for ONNX export/inference pipeline and Neural Magic's [Sparsify](https://github.com/neuralmagic/sparsify) for one-shot quantization.
|
| 714 |
|
| 715 |
Current list of sparse and quantized bge ONNX models:
|