gte-small-quant
This is the quantized (INT8) ONNX variant of the gte-small embeddings model created with DeepSparse Optimum for ONNX export/inference and Neural Magic's Sparsify for one-shot quantization.
Current list of sparse and quantized gte ONNX models:
Links | Sparsification Method |
---|---|
zeroshot/gte-large-sparse | Quantization (INT8) & 50% Pruning |
zeroshot/gte-large-quant | Quantization (INT8) |
zeroshot/gte-base-sparse | Quantization (INT8) & 50% Pruning |
zeroshot/gte-base-quant | Quantization (INT8) |
zeroshot/gte-small-sparse | Quantization (INT8) & 50% Pruning |
zeroshot/gte-small-quant | Quantization (INT8) |
pip install -U deepsparse-nightly[sentence_transformers]
from deepsparse.sentence_transformers import SentenceTransformer
model = SentenceTransformer('zeroshot/gte-small-quant', export=False)
# Our sentences we like to encode
sentences = ['This framework generates embeddings for each input sentence',
'Sentences are passed as a list of string.',
'The quick brown fox jumps over the lazy dog.']
# Sentences are encoded by calling model.encode()
embeddings = model.encode(sentences)
# Print the embeddings
for sentence, embedding in zip(sentences, embeddings):
print("Sentence:", sentence)
print("Embedding:", embedding.shape)
print("")
For further details regarding DeepSparse & Sentence Transformers integration, refer to the DeepSparse README.
For general questions on these models and sparsification methods, reach out to the engineering team on our community Slack.
- Downloads last month
- 209
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.
Spaces using zeroshot/gte-small-quant 4
Evaluation results
- accuracy on MTEB AmazonCounterfactualClassification (en)test set self-reported72.881
- ap on MTEB AmazonCounterfactualClassification (en)test set self-reported35.742
- f1 on MTEB AmazonCounterfactualClassification (en)test set self-reported66.981
- accuracy on MTEB AmazonPolarityClassificationtest set self-reported91.032
- ap on MTEB AmazonPolarityClassificationtest set self-reported87.607
- f1 on MTEB AmazonPolarityClassificationtest set self-reported91.010
- accuracy on MTEB AmazonReviewsClassification (en)test set self-reported46.944
- f1 on MTEB AmazonReviewsClassification (en)test set self-reported46.333
- map on MTEB AskUbuntuDupQuestionstest set self-reported60.757
- mrr on MTEB AskUbuntuDupQuestionstest set self-reported73.516