diff --git "a/last-checkpoint/README.md" "b/last-checkpoint/README.md"
--- "a/last-checkpoint/README.md"
+++ "b/last-checkpoint/README.md"
@@ -7,12 +7,12 @@ tags:
- sentence-similarity
- feature-extraction
- generated_from_trainer
-- dataset_size:131566
-- loss:MultipleNegativesRankingLoss
-- loss:CoSENTLoss
+- dataset_size:526885
- loss:GISTEmbedLoss
+- loss:CoSENTLoss
- loss:OnlineContrastiveLoss
- loss:MultipleNegativesSymmetricRankingLoss
+- loss:MarginMSELoss
base_model: microsoft/deberta-v3-small
datasets:
- sentence-transformers/all-nli
@@ -25,46 +25,137 @@ datasets:
- allenai/sciq
- allenai/qasc
- allenai/openbookqa
-- sentence-transformers/msmarco-msmarco-distilbert-base-v3
- sentence-transformers/natural-questions
- sentence-transformers/trivia-qa
- sentence-transformers/quora-duplicates
- sentence-transformers/gooaq
+metrics:
+- pearson_cosine
+- spearman_cosine
+- pearson_manhattan
+- spearman_manhattan
+- pearson_euclidean
+- spearman_euclidean
+- pearson_dot
+- spearman_dot
+- pearson_max
+- spearman_max
widget:
-- source_sentence: Centrosome-independent mitotic spindle formation in vertebrates.
+- source_sentence: A man in a Santa Claus costume is sitting on a wooden chair holding
+ a microphone and a stringed instrument.
sentences:
- - Birds pair up with the same bird in mating season.
- - We use voltage to keep track of electric potential energy.
- - A mitotic spindle forms from the centrosomes.
-- source_sentence: A dog carrying a stick in its mouth runs through a snow-covered
- field.
+ - The man is is near the ball.
+ - The man is wearing a costume.
+ - People are having a picnic.
+- source_sentence: A street vendor selling his art.
sentences:
- - The children played on the floor.
- - A pair of people play video games together on a couch.
- - A animal carried a stick through a snow covered field.
-- source_sentence: A guy on a skateboard, jumping off some steps.
+ - A man is selling things on the street.
+ - A woman is walking outside.
+ - A clown is talking into a microphone.
+- source_sentence: A boy looks surly as his father looks at the camera.
sentences:
- - A woman is making music.
- - a guy with a skateboard making a jump
- - A dog holds an object in the water.
-- source_sentence: A photographer with bushy dark hair takes a photo of a skateboarder
- at an indoor park.
+ - a boy looks at his farther
+ - A dark-haired girl in a spotted shirt is pointing at the picture while sitting
+ next to a boy wearing a purple shirt and jeans.
+ - Man and woman stop and chat with each other.
+- source_sentence: Which company provided streetcar connections between downtown and
+ the hospital?
sentences:
- - The person with the camera photographs the person skating.
- - A man starring at a piece of paper.
- - The man is riding a bike in sand.
-- source_sentence: Why did oil start getting priced in terms of gold?
+ - In 1914 developers Billings & Meyering acquired the tract, completed street development,
+ provided the last of the necessary municipal improvements including water service,
+ and began marketing the property with fervor.
+ - The war was fought primarily along the frontiers between New France and the British
+ colonies, from Virginia in the South to Nova Scotia in the North.
+ - 'On the basis of CST, Burnet developed a theory of how an immune response is triggered
+ according to the self/nonself distinction: "self" constituents (constituents of
+ the body) do not trigger destructive immune responses, while "nonself" entities
+ (pathogens, an allograft) trigger a destructive immune response.'
+- source_sentence: What language did Tesla study while in school?
sentences:
- - Because oil was priced in dollars, oil producers' real income decreased.
- - This allows all set top boxes in a household to share recordings and other media.
- - Only the series from 2009 onwards are available on Blu-ray, except for the 1970
- story Spearhead from Space, released in July 2013.
+ - Because of the complexity of medications including specific indications, effectiveness
+ of treatment regimens, safety of medications (i.e., drug interactions) and patient
+ compliance issues (in the hospital and at home) many pharmacists practicing in
+ hospitals gain more education and training after pharmacy school through a pharmacy
+ practice residency and sometimes followed by another residency in a specific area.
+ - Rev. Jimmy Creech was defrocked after a highly publicized church trial in 1999
+ on account of his participation in same-sex union ceremonies.
+ - Tesla was the fourth of five children.
pipeline_tag: sentence-similarity
+model-index:
+- name: SentenceTransformer based on microsoft/deberta-v3-small
+ results:
+ - task:
+ type: semantic-similarity
+ name: Semantic Similarity
+ dataset:
+ name: sts test
+ type: sts-test
+ metrics:
+ - type: pearson_cosine
+ value: 0.2520910673470529
+ name: Pearson Cosine
+ - type: spearman_cosine
+ value: 0.2588662067006675
+ name: Spearman Cosine
+ - type: pearson_manhattan
+ value: 0.30439718484055006
+ name: Pearson Manhattan
+ - type: spearman_manhattan
+ value: 0.3013780326567434
+ name: Spearman Manhattan
+ - type: pearson_euclidean
+ value: 0.25977707672353506
+ name: Pearson Euclidean
+ - type: spearman_euclidean
+ value: 0.26078444276128726
+ name: Spearman Euclidean
+ - type: pearson_dot
+ value: 0.08121075567918108
+ name: Pearson Dot
+ - type: spearman_dot
+ value: 0.0753891417253212
+ name: Spearman Dot
+ - type: pearson_max
+ value: 0.30439718484055006
+ name: Pearson Max
+ - type: spearman_max
+ value: 0.3013780326567434
+ name: Spearman Max
+ - type: pearson_cosine
+ value: 0.2520910673470529
+ name: Pearson Cosine
+ - type: spearman_cosine
+ value: 0.2588662067006675
+ name: Spearman Cosine
+ - type: pearson_manhattan
+ value: 0.30439718484055006
+ name: Pearson Manhattan
+ - type: spearman_manhattan
+ value: 0.3013780326567434
+ name: Spearman Manhattan
+ - type: pearson_euclidean
+ value: 0.25977707672353506
+ name: Pearson Euclidean
+ - type: spearman_euclidean
+ value: 0.26078444276128726
+ name: Spearman Euclidean
+ - type: pearson_dot
+ value: 0.08121075567918108
+ name: Pearson Dot
+ - type: spearman_dot
+ value: 0.0753891417253212
+ name: Spearman Dot
+ - type: pearson_max
+ value: 0.30439718484055006
+ name: Pearson Max
+ - type: spearman_max
+ value: 0.3013780326567434
+ name: Spearman Max
---
# SentenceTransformer based on microsoft/deberta-v3-small
-This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [microsoft/deberta-v3-small](https://huggingface.co/microsoft/deberta-v3-small) on the [nli-pairs](https://huggingface.co/datasets/sentence-transformers/all-nli), [sts-label](https://huggingface.co/datasets/sentence-transformers/stsb), [vitaminc-pairs](https://huggingface.co/datasets/tals/vitaminc), [qnli-contrastive](https://huggingface.co/datasets/nyu-mll/glue), [scitail-pairs-qa](https://huggingface.co/datasets/allenai/scitail), [scitail-pairs-pos](https://huggingface.co/datasets/allenai/scitail), [xsum-pairs](https://huggingface.co/datasets/sentence-transformers/xsum), [compression-pairs](https://huggingface.co/datasets/sentence-transformers/sentence-compression), [sciq_pairs](https://huggingface.co/datasets/allenai/sciq), [qasc_pairs](https://huggingface.co/datasets/allenai/qasc), [openbookqa_pairs](https://huggingface.co/datasets/allenai/openbookqa), [msmarco_pairs](https://huggingface.co/datasets/sentence-transformers/msmarco-msmarco-distilbert-base-v3), [nq_pairs](https://huggingface.co/datasets/sentence-transformers/natural-questions), [trivia_pairs](https://huggingface.co/datasets/sentence-transformers/trivia-qa), [quora_pairs](https://huggingface.co/datasets/sentence-transformers/quora-duplicates) and [gooaq_pairs](https://huggingface.co/datasets/sentence-transformers/gooaq) datasets. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
+This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [microsoft/deberta-v3-small](https://huggingface.co/microsoft/deberta-v3-small) on the [nli-pairs](https://huggingface.co/datasets/sentence-transformers/all-nli), [sts-label](https://huggingface.co/datasets/sentence-transformers/stsb), [vitaminc-pairs](https://huggingface.co/datasets/tals/vitaminc), [qnli-contrastive](https://huggingface.co/datasets/nyu-mll/glue), [scitail-pairs-qa](https://huggingface.co/datasets/allenai/scitail), [scitail-pairs-pos](https://huggingface.co/datasets/allenai/scitail), [xsum-pairs](https://huggingface.co/datasets/sentence-transformers/xsum), [compression-pairs](https://huggingface.co/datasets/sentence-transformers/sentence-compression), [sciq_pairs](https://huggingface.co/datasets/allenai/sciq), [qasc_pairs](https://huggingface.co/datasets/allenai/qasc), [openbookqa_pairs](https://huggingface.co/datasets/allenai/openbookqa), msmarco_pairs, [nq_pairs](https://huggingface.co/datasets/sentence-transformers/natural-questions), [trivia_pairs](https://huggingface.co/datasets/sentence-transformers/trivia-qa), [quora_pairs](https://huggingface.co/datasets/sentence-transformers/quora-duplicates) and [gooaq_pairs](https://huggingface.co/datasets/sentence-transformers/gooaq) datasets. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
## Model Details
@@ -86,7 +177,7 @@ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [m
- [sciq_pairs](https://huggingface.co/datasets/allenai/sciq)
- [qasc_pairs](https://huggingface.co/datasets/allenai/qasc)
- [openbookqa_pairs](https://huggingface.co/datasets/allenai/openbookqa)
- - [msmarco_pairs](https://huggingface.co/datasets/sentence-transformers/msmarco-msmarco-distilbert-base-v3)
+ - msmarco_pairs
- [nq_pairs](https://huggingface.co/datasets/sentence-transformers/natural-questions)
- [trivia_pairs](https://huggingface.co/datasets/sentence-transformers/trivia-qa)
- [quora_pairs](https://huggingface.co/datasets/sentence-transformers/quora-duplicates)
@@ -127,9 +218,9 @@ from sentence_transformers import SentenceTransformer
model = SentenceTransformer("bobox/DeBERTaV3-small-GeneralSentenceTransformer-v2-checkpoints-tmp")
# Run inference
sentences = [
- 'Why did oil start getting priced in terms of gold?',
- "Because oil was priced in dollars, oil producers' real income decreased.",
- 'This allows all set top boxes in a household to share recordings and other media.',
+ 'What language did Tesla study while in school?',
+ 'Tesla was the fourth of five children.',
+ 'Rev. Jimmy Creech was defrocked after a highly publicized church trial in 1999 on account of his participation in same-sex union ceremonies.',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
@@ -165,6 +256,44 @@ You can finetune this model on your own dataset.
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
-->
+## Evaluation
+
+### Metrics
+
+#### Semantic Similarity
+* Dataset: `sts-test`
+* Evaluated with [EmbeddingSimilarityEvaluator
](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator)
+
+| Metric | Value |
+|:--------------------|:-----------|
+| pearson_cosine | 0.2521 |
+| **spearman_cosine** | **0.2589** |
+| pearson_manhattan | 0.3044 |
+| spearman_manhattan | 0.3014 |
+| pearson_euclidean | 0.2598 |
+| spearman_euclidean | 0.2608 |
+| pearson_dot | 0.0812 |
+| spearman_dot | 0.0754 |
+| pearson_max | 0.3044 |
+| spearman_max | 0.3014 |
+
+#### Semantic Similarity
+* Dataset: `sts-test`
+* Evaluated with [EmbeddingSimilarityEvaluator
](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator)
+
+| Metric | Value |
+|:--------------------|:-----------|
+| pearson_cosine | 0.2521 |
+| **spearman_cosine** | **0.2589** |
+| pearson_manhattan | 0.3044 |
+| spearman_manhattan | 0.3014 |
+| pearson_euclidean | 0.2598 |
+| spearman_euclidean | 0.2608 |
+| pearson_dot | 0.0812 |
+| spearman_dot | 0.0754 |
+| pearson_max | 0.3044 |
+| spearman_max | 0.3014 |
+