diff --git "a/README.md" "b/README.md" new file mode 100644--- /dev/null +++ "b/README.md" @@ -0,0 +1,7084 @@ +--- +base_model: sentence-transformers/all-MiniLM-L6-v2 +language: +- en +library_name: sentence-transformers +license: apache-2.0 +metrics: +- cosine_accuracy +pipeline_tag: sentence-similarity +tags: +- sentence-transformers +- sentence-similarity +- feature-extraction +- generated_from_trainer +- dataset_size:1943715 +- loss:MultipleNegativesRankingLoss +widget: +- source_sentence: on the use of the wavelet decomposition for time series prediction + . + sentences: + - Combining Neural Network Forecasts on Wavelet-Transformed Time Series + - Human fertilization fertilization process. Human fertilization Human fertilization + is the union of a human egg and sperm, usually occurring in the ampulla of the + fallopian tube. The result of this union is the production of a zygote cell, or + fertilized egg, initiating prenatal development. Scientists discovered the dynamics + of human fertilization in the nineteenth century. The process of fertilization + involves a sperm fusing with an ovum. The most common sequence begins with ejaculation + during copulation, follows with ovulation, and finishes with fertilization. Various + exceptions to this sequence are possible, including artificial insemination, "in + vitro" fertilization, external ejaculation without copulation, or copulation shortly + - Hydraulic failure and repair are not routine in trees +- source_sentence: city in the middle of the united states + sentences: + - Air India and became the first Asian airline to induct a jet aircraft in its fleet. + In 2000–01, attempts were made to privatise Air India and from 2006 onwards, it + suffered losses after its merger with Indian Airlines. Air India also operates + flights to domestic and Asian destinations through its subsidiaries Alliance Air + and Air India Express. Air India's mascot is "the Maharajah" (Emperor) and the + logo consists of a flying swan with the wheel of Konark inside it. Air India had + its origin as Tata Air Services later renamed to Tata Airlines founded by J. R. + D. Tata of Tata Sons, + - Hiram, Georgia none of those age 65 or over. The city of Hiram has two public + schools named after it, Hiram High School and Hiram Elementary School. The nearest + middle schools are Dobbins Middle School, P.B. Ritch Middle School and East Paulding + Middle School. Hiram, Georgia Hiram is a city in Paulding County, Georgia, United + States. As of 2016, the population is 3,898. The city was named after Hiram Baggett, + the town's first postmaster. Hiram is located at (33.865575, -84.774593). According + to the United States Census Bureau, the city has a total area of , of which is + land and (0.66%) is + - Geographic center of the contiguous United States Geographic center of the contiguous + United States The geographic center of the contiguous United States is the center + of 48 U.S. states. It has been regarded as such by the U.S. National Geodetic + Survey (NGS) since the 1912 additions of New Mexico and Arizona to the United + States. Its position as located in a 1918 survey is located at , about northwest + of the center of Lebanon, Kansas, approximately south of the Kansas–Nebraska border. + While any measurement of the exact center of a land mass will always be imprecise + due to changing shorelines and other factors, the NGS coordinates are +- source_sentence: who sings the chorus in sing for the moment by eminem + sentences: + - Dream On (Aerosmith song) tried to sing a bit lower and sound more like soul artists, + such as James Brown. The song is also famous for its building climax to showcase + Tyler's trademark screams. The main riff and chorus of the song were sampled in + the 2002 song "Sing for the Moment" by rapper Eminem on "The Eminem Show" that + also features a solo from Joe Perry. The original 1973 issue of this 45 (45894) + has a shorter edit of "Dream On" (3:25). Most of the intro has been edited out + and the first chorus is replaced with the second chorus. The second issue + - Mayflower Mayflower The Mayflower was an English ship that transported the first + English Puritans, known today as the Pilgrims, from Plymouth, England, to the + New World in 1620. There were 102 passengers, and the crew is estimated to have + been about 30, but the exact number is unknown. The ship has become a cultural + icon in the history of the United States. The Pilgrims signed the Mayflower Compact + prior to leaving the ship and establishing Plymouth Colony, a document which established + a rudimentary form of democracy with each member contributing to the welfare of + the community. There was a second ship + - Sing for the Moment States, "Sing for the Moment" reached number fourteen on the + "Billboard" Hot 100. The song, along with the original "Dream On", was used in + a trailer for the 2016 animated film "Sing". "Sing for the Moment" contains samples + of the song "Dream On" by the rock band Aerosmith. Joe Perry plays the guitar + solo at the end of the song, and a sample of Steven Tyler singing is used as the + chorus for this song. Eminem chants "sing" when Tyler starts to sing the chorus, + and Eminem also chants "sing with me" and "come on". Eminem says the words in +- source_sentence: 'The concurrent use of anticholinergics and cholinesterase inhibitors: + rare event or common practice?' + sentences: + - Winter Olympic Games well as Summer) Games to proclaim the superiority of their + political systems. The Winter Olympics has been hosted on three continents by + twelve different countries. The Games have been held four times in the United + States (in 1932, 1960, 1980 and 2002); three times in France (in 1924, 1968 and + 1992); and twice each in Austria (1964, 1976), Canada (1988, 2010), Japan (1972, + 1998), Italy (1956, 2006), Norway (1952, 1994), and Switzerland (1928, 1948). + Also, the Games have been held just once each in Germany (1936), Yugoslavia (1984), + Russia (2014) and South Korea (2018). The IOC has selected Beijing, China, + - 'Previous studies have shown that positive peritoneal washings may adversely affect + cancer survival rates and that hysteroscopy is associated with a higher risk of + positive washings in patients with endometrial carcinoma. Our aim was to assess + if diagnostic hysteroscopy increases the risk of positive peritoneal washings + in patients with endometrial cancer and affects the prognosis after surgery.Retrospective + cohort study. The medical records of 50 consecutive patients with endometrial + carcinoma, diagnosed with hysteroscopy and tissue sampling and treated by abdominal + hysterectomy with bilateral salpingo-oophorectomy and peritoneal washings were + reviewed.Of the 43 patients with endometrial carcinoma FIGO stage I, none had + positive peritoneal washings (95%CI: 0-8.2%). The mean interval between hysteroscopy + and surgery was 33.5 days. The 5-year disease-specific survival rate was 91.8%, + the 5-year recurrence-free survival rate was 85.4%.' + - 'To measure the prevalence of anticholinergic use cross-sectionally in patients + receiving cholinesterase inhibitors and to describe change in use of anticholinergics + upon inception of cholinesterase inhibitor treatment.Cross-sectional and inception + cohort studies.State of Iowa.Iowa Medicaid beneficiaries aged 50 and older with + a pharmacy claim for a cholinesterase inhibitor during January 1997 through February + 2000.Anticholinergic use was determined for all patients with a cholinesterase + inhibitor pharmacy claim during January and February of 2000. A frequency distribution + of all anticholinergics was compiled, with emphasis placed on those considered + inappropriate in the elderly. In a separate analysis, anticholinergic use was + determined at two points: 90 days before and after cholinesterase inhibitor inception.Of + 557 patients receiving a cholinesterase inhibitor, 197 (35.4%) received an anticholinergic + concurrently. Of all anticholinergics, 74.5% (178/239) had been identified as + inappropriate for use in the elderly, 22.2% (53/239) under any circumstances. + At the time of cholinesterase inhibitor inception, 30.2% (143/474) and 33.5% (159/474) + of patients received an anticholinergic 90 days before and 90 days after inception, + respectively. Increases in anticholinergic prescribing upon cholinesterase inhibitor + inception exceeded decreases (Wilcoxon signed-rank test, S=529, P=.020).' +- source_sentence: Should I stay or should I go? + sentences: + - 'Several recent studies suggest that acceleration of the head at impact during + sporting activities may have a detrimental effect on cognitive function. Reducing + acceleration of impact in these sports could reduce neurologic sequelae.To measure + the effectiveness of a regulation football helmet to reduce acceleration of impact + for both low- and moderate-force impacts.An experimental paired study design was + used. Male volunteers between 16 and 30 years of age headed soccer balls traveling + approximately 35 miles per hour bareheaded and with a helmet. An intraoral accelerometer + worn inside a plastic mouthpiece measured acceleration of the head. The helmet + also had an accelerometer placed inside the padding. For more forceful impacts, + cadaver heads, both with and without helmets, were instrumented with intraoral + (IO) and intracranial (IC) accelerometers and struck with a pendulum device. Simultaneous + IO and IC accelerations were measured and compared between helmeted and unhelmeted + cadaver heads. The main outcome was mean peak acceleration of the head and/or + brain associated with low- and moderate-force impacts with and without protective + headgear.Mean peak Gs, measured by the mouthpiece accelerometer, were significantly + reduced when the participants heading soccer balls were wearing a helmet (7.7 + Gs with vs 19.2 Gs without, p = 0.01). Wearing a helmet also significantly lowered + the peak Gs measured intraorally and intracranially in cadavers subjected to moderate-force + pendulum impacts: 28.7 Gs with vs 62.6 Gs without, p<0.001; and 56.4 Gs with vs + 81.6 Gs without, p<0.001, respectively.' + - The aim of this study was to examine the experiences of parents encountering the + critical deterioration and resuscitative care of other children in the pediatric + intensive care unit where their own child was admitted.Grounded theory qualitative + methodology.Pediatric intensive care unit of a pediatric tertiary care center + in Montreal, Canada.Ten parents of critically ill children who witnessed resuscitative + measures on another child.None.Semistructured interviews were conducted. While + witnessing resuscitation, parents struggled with "Should I stay or should I go?" + Their decision depended on specific contributing factors that were intrinsic to + parents (curiosity or apprehension, the child's sake, trust or distrust) or extrinsic + (limited space). These parents were not "spectators." Despite using coping strategies, + the experiences were distressing in the majority of cases, although sometimes + comforting. The impact on witnessing critical events had divergent effects on + parental trust with healthcare professionals. + - Shut Up and Dance (Walk the Moon song) Tour on July 24 in Foxborough, Massachusetts. + Walk the Moon performed the song at the 2015 MTV Video Music Awards pre-show on + August 30, 2015, where the band had played on a circular, multicolored stage prior + to the show's start. The band also performed the song at the 2016 NBA All-Star + Game on February 13, 2016 in Toronto, Canada. The music video, a 1980s club-themed + movie-style music video, was released on YouTube on October 23, 2014. It stars + professional dancer Lauren Taft alongside Petricca. Credits adapted from the liner + notes of "Talking Is Hard". Locations Personnel Shut Up and Dance +model-index: +- name: all-MiniLM-L6-v2 trained on MEDI-MTEB triplets + results: + - task: + type: triplet + name: Triplet + dataset: + name: medi mteb dev + type: medi-mteb-dev + metrics: + - type: cosine_accuracy + value: 0.9144999726312333 + name: Cosine Accuracy +--- + +# all-MiniLM-L6-v2 trained on MEDI-MTEB triplets + +This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) on the NQ, pubmed, specter_train_triples, S2ORC_citations_abstracts, fever, gooaq_pairs, codesearchnet, wikihow, WikiAnswers, eli5_question_answer, amazon-qa, medmcqa, zeroshot, TriviaQA_pairs, PAQ_pairs, stackexchange_duplicate_questions_title-body_title-body, trex, flickr30k_captions, hotpotqa, task671_ambigqa_text_generation, task061_ropes_answer_generation, task285_imdb_answer_generation, task905_hate_speech_offensive_classification, task566_circa_classification, task184_snli_entailment_to_neutral_text_modification, task280_stereoset_classification_stereotype_type, task1599_smcalflow_classification, task1384_deal_or_no_dialog_classification, task591_sciq_answer_generation, task823_peixian-rtgender_sentiment_analysis, task023_cosmosqa_question_generation, task900_freebase_qa_category_classification, task924_event2mind_word_generation, task152_tomqa_find_location_easy_noise, task1368_healthfact_sentence_generation, task1661_super_glue_classification, task1187_politifact_classification, task1728_web_nlg_data_to_text, task112_asset_simple_sentence_identification, task1340_msr_text_compression_compression, task072_abductivenli_answer_generation, task1504_hatexplain_answer_generation, task684_online_privacy_policy_text_information_type_generation, task1290_xsum_summarization, task075_squad1.1_answer_generation, task1587_scifact_classification, task384_socialiqa_question_classification, task1555_scitail_answer_generation, task1532_daily_dialog_emotion_classification, task239_tweetqa_answer_generation, task596_mocha_question_generation, task1411_dart_subject_identification, task1359_numer_sense_answer_generation, task329_gap_classification, task220_rocstories_title_classification, task316_crows-pairs_classification_stereotype, task495_semeval_headline_classification, task1168_brown_coarse_pos_tagging, task348_squad2.0_unanswerable_question_generation, task049_multirc_questions_needed_to_answer, task1534_daily_dialog_question_classification, task322_jigsaw_classification_threat, task295_semeval_2020_task4_commonsense_reasoning, task186_snli_contradiction_to_entailment_text_modification, task034_winogrande_question_modification_object, task160_replace_letter_in_a_sentence, task469_mrqa_answer_generation, task105_story_cloze-rocstories_sentence_generation, task649_race_blank_question_generation, task1536_daily_dialog_happiness_classification, task683_online_privacy_policy_text_purpose_answer_generation, task024_cosmosqa_answer_generation, task584_udeps_eng_fine_pos_tagging, task066_timetravel_binary_consistency_classification, task413_mickey_en_sentence_perturbation_generation, task182_duorc_question_generation, task028_drop_answer_generation, task1601_webquestions_answer_generation, task1295_adversarial_qa_question_answering, task201_mnli_neutral_classification, task038_qasc_combined_fact, task293_storycommonsense_emotion_text_generation, task572_recipe_nlg_text_generation, task517_emo_classify_emotion_of_dialogue, task382_hybridqa_answer_generation, task176_break_decompose_questions, task1291_multi_news_summarization, task155_count_nouns_verbs, task031_winogrande_question_generation_object, task279_stereoset_classification_stereotype, task1336_peixian_equity_evaluation_corpus_gender_classifier, task508_scruples_dilemmas_more_ethical_isidentifiable, task518_emo_different_dialogue_emotions, task077_splash_explanation_to_sql, task923_event2mind_classifier, task470_mrqa_question_generation, task638_multi_woz_classification, task1412_web_questions_question_answering, task847_pubmedqa_question_generation, task678_ollie_actual_relationship_answer_generation, task290_tellmewhy_question_answerability, task575_air_dialogue_classification, task189_snli_neutral_to_contradiction_text_modification, task026_drop_question_generation, task162_count_words_starting_with_letter, task079_conala_concat_strings, task610_conllpp_ner, task046_miscellaneous_question_typing, task197_mnli_domain_answer_generation, task1325_qa_zre_question_generation_on_subject_relation, task430_senteval_subject_count, task672_nummersense, task402_grailqa_paraphrase_generation, task904_hate_speech_offensive_classification, task192_hotpotqa_sentence_generation, task069_abductivenli_classification, task574_air_dialogue_sentence_generation, task187_snli_entailment_to_contradiction_text_modification, task749_glucose_reverse_cause_emotion_detection, task1552_scitail_question_generation, task750_aqua_multiple_choice_answering, task327_jigsaw_classification_toxic, task1502_hatexplain_classification, task328_jigsaw_classification_insult, task304_numeric_fused_head_resolution, task1293_kilt_tasks_hotpotqa_question_answering, task216_rocstories_correct_answer_generation, task1326_qa_zre_question_generation_from_answer, task1338_peixian_equity_evaluation_corpus_sentiment_classifier, task1729_personachat_generate_next, task1202_atomic_classification_xneed, task400_paws_paraphrase_classification, task502_scruples_anecdotes_whoiswrong_verification, task088_identify_typo_verification, task221_rocstories_two_choice_classification, task200_mnli_entailment_classification, task074_squad1.1_question_generation, task581_socialiqa_question_generation, task1186_nne_hrngo_classification, task898_freebase_qa_answer_generation, task1408_dart_similarity_classification, task168_strategyqa_question_decomposition, task1357_xlsum_summary_generation, task390_torque_text_span_selection, task165_mcscript_question_answering_commonsense, task1533_daily_dialog_formal_classification, task002_quoref_answer_generation, task1297_qasc_question_answering, task305_jeopardy_answer_generation_normal, task029_winogrande_full_object, task1327_qa_zre_answer_generation_from_question, task326_jigsaw_classification_obscene, task1542_every_ith_element_from_starting, task570_recipe_nlg_ner_generation, task1409_dart_text_generation, task401_numeric_fused_head_reference, task846_pubmedqa_classification, task1712_poki_classification, task344_hybridqa_answer_generation, task875_emotion_classification, task1214_atomic_classification_xwant, task106_scruples_ethical_judgment, task238_iirc_answer_from_passage_answer_generation, task1391_winogrande_easy_answer_generation, task195_sentiment140_classification, task163_count_words_ending_with_letter, task579_socialiqa_classification, task569_recipe_nlg_text_generation, task1602_webquestion_question_genreation, task747_glucose_cause_emotion_detection, task219_rocstories_title_answer_generation, task178_quartz_question_answering, task103_facts2story_long_text_generation, task301_record_question_generation, task1369_healthfact_sentence_generation, task515_senteval_odd_word_out, task496_semeval_answer_generation, task1658_billsum_summarization, task1204_atomic_classification_hinderedby, task1392_superglue_multirc_answer_verification, task306_jeopardy_answer_generation_double, task1286_openbookqa_question_answering, task159_check_frequency_of_words_in_sentence_pair, task151_tomqa_find_location_easy_clean, task323_jigsaw_classification_sexually_explicit, task037_qasc_generate_related_fact, task027_drop_answer_type_generation, task1596_event2mind_text_generation_2, task141_odd-man-out_classification_category, task194_duorc_answer_generation, task679_hope_edi_english_text_classification, task246_dream_question_generation, task1195_disflqa_disfluent_to_fluent_conversion, task065_timetravel_consistent_sentence_classification, task351_winomt_classification_gender_identifiability_anti, task580_socialiqa_answer_generation, task583_udeps_eng_coarse_pos_tagging, task202_mnli_contradiction_classification, task222_rocstories_two_chioce_slotting_classification, task498_scruples_anecdotes_whoiswrong_classification, task067_abductivenli_answer_generation, task616_cola_classification, task286_olid_offense_judgment, task188_snli_neutral_to_entailment_text_modification, task223_quartz_explanation_generation, task820_protoqa_answer_generation, task196_sentiment140_answer_generation, task1678_mathqa_answer_selection, task349_squad2.0_answerable_unanswerable_question_classification, task154_tomqa_find_location_hard_noise, task333_hateeval_classification_hate_en, task235_iirc_question_from_subtext_answer_generation, task1554_scitail_classification, task210_logic2text_structured_text_generation, task035_winogrande_question_modification_person, task230_iirc_passage_classification, task1356_xlsum_title_generation, task1726_mathqa_correct_answer_generation, task302_record_classification, task380_boolq_yes_no_question, task212_logic2text_classification, task748_glucose_reverse_cause_event_detection, task834_mathdataset_classification, task350_winomt_classification_gender_identifiability_pro, task191_hotpotqa_question_generation, task236_iirc_question_from_passage_answer_generation, task217_rocstories_ordering_answer_generation, task568_circa_question_generation, task614_glucose_cause_event_detection, task361_spolin_yesand_prompt_response_classification, task421_persent_sentence_sentiment_classification, task203_mnli_sentence_generation, task420_persent_document_sentiment_classification, task153_tomqa_find_location_hard_clean, task346_hybridqa_classification, task1211_atomic_classification_hassubevent, task360_spolin_yesand_response_generation, task510_reddit_tifu_title_summarization, task511_reddit_tifu_long_text_summarization, task345_hybridqa_answer_generation, task270_csrg_counterfactual_context_generation, task307_jeopardy_answer_generation_final, task001_quoref_question_generation, task089_swap_words_verification, task1196_atomic_classification_oeffect, task080_piqa_answer_generation, task1598_nyc_long_text_generation, task240_tweetqa_question_generation, task615_moviesqa_answer_generation, task1347_glue_sts-b_similarity_classification, task114_is_the_given_word_longest, task292_storycommonsense_character_text_generation, task115_help_advice_classification, task431_senteval_object_count, task1360_numer_sense_multiple_choice_qa_generation, task177_para-nmt_paraphrasing, task132_dais_text_modification, task269_csrg_counterfactual_story_generation, task233_iirc_link_exists_classification, task161_count_words_containing_letter, task1205_atomic_classification_isafter, task571_recipe_nlg_ner_generation, task1292_yelp_review_full_text_categorization, task428_senteval_inversion, task311_race_question_generation, task429_senteval_tense, task403_creak_commonsense_inference, task929_products_reviews_classification, task582_naturalquestion_answer_generation, task237_iirc_answer_from_subtext_answer_generation, task050_multirc_answerability, task184_break_generate_question, task669_ambigqa_answer_generation, task169_strategyqa_sentence_generation, task500_scruples_anecdotes_title_generation, task241_tweetqa_classification, task1345_glue_qqp_question_paraprashing, task218_rocstories_swap_order_answer_generation, task613_politifact_text_generation, task1167_penn_treebank_coarse_pos_tagging, task1422_mathqa_physics, task247_dream_answer_generation, task199_mnli_classification, task164_mcscript_question_answering_text, task1541_agnews_classification, task516_senteval_conjoints_inversion, task294_storycommonsense_motiv_text_generation, task501_scruples_anecdotes_post_type_verification, task213_rocstories_correct_ending_classification, task821_protoqa_question_generation, task493_review_polarity_classification, task308_jeopardy_answer_generation_all, task1595_event2mind_text_generation_1, task040_qasc_question_generation, task231_iirc_link_classification, task1727_wiqa_what_is_the_effect, task578_curiosity_dialogs_answer_generation, task310_race_classification, task309_race_answer_generation, task379_agnews_topic_classification, task030_winogrande_full_person, task1540_parsed_pdfs_summarization, task039_qasc_find_overlapping_words, task1206_atomic_classification_isbefore, task157_count_vowels_and_consonants, task339_record_answer_generation, task453_swag_answer_generation, task848_pubmedqa_classification, task673_google_wellformed_query_classification, task676_ollie_relationship_answer_generation, task268_casehold_legal_answer_generation, task844_financial_phrasebank_classification, task330_gap_answer_generation, task595_mocha_answer_generation, task1285_kpa_keypoint_matching, task234_iirc_passage_line_answer_generation, task494_review_polarity_answer_generation, task670_ambigqa_question_generation, task289_gigaword_summarization, npr, nli, SimpleWiki, amazon_review_2018, ccnews_title_text, agnews, xsum, msmarco, yahoo_answers_title_answer, squad_pairs, wow, mteb-amazon_counterfactual-avs_triplets, mteb-amazon_massive_intent-avs_triplets, mteb-amazon_massive_scenario-avs_triplets, mteb-amazon_reviews_multi-avs_triplets, mteb-banking77-avs_triplets, mteb-emotion-avs_triplets, mteb-imdb-avs_triplets, mteb-mtop_domain-avs_triplets, mteb-mtop_intent-avs_triplets, mteb-toxic_conversations_50k-avs_triplets, mteb-tweet_sentiment_extraction-avs_triplets and covid-bing-query-gpt4-avs_triplets datasets. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. + +## Model Details + +### Model Description +- **Model Type:** Sentence Transformer +- **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) +- **Maximum Sequence Length:** 256 tokens +- **Output Dimensionality:** 768 dimensions +- **Similarity Function:** Cosine Similarity +- **Training Datasets:** + - NQ + - pubmed + - specter_train_triples + - S2ORC_citations_abstracts + - fever + - gooaq_pairs + - codesearchnet + - wikihow + - WikiAnswers + - eli5_question_answer + - amazon-qa + - medmcqa + - zeroshot + - TriviaQA_pairs + - PAQ_pairs + - stackexchange_duplicate_questions_title-body_title-body + - trex + - flickr30k_captions + - hotpotqa + - task671_ambigqa_text_generation + - task061_ropes_answer_generation + - task285_imdb_answer_generation + - task905_hate_speech_offensive_classification + - task566_circa_classification + - task184_snli_entailment_to_neutral_text_modification + - task280_stereoset_classification_stereotype_type + - task1599_smcalflow_classification + - task1384_deal_or_no_dialog_classification + - task591_sciq_answer_generation + - task823_peixian-rtgender_sentiment_analysis + - task023_cosmosqa_question_generation + - task900_freebase_qa_category_classification + - task924_event2mind_word_generation + - task152_tomqa_find_location_easy_noise + - task1368_healthfact_sentence_generation + - task1661_super_glue_classification + - task1187_politifact_classification + - task1728_web_nlg_data_to_text + - task112_asset_simple_sentence_identification + - task1340_msr_text_compression_compression + - task072_abductivenli_answer_generation + - task1504_hatexplain_answer_generation + - task684_online_privacy_policy_text_information_type_generation + - task1290_xsum_summarization + - task075_squad1.1_answer_generation + - task1587_scifact_classification + - task384_socialiqa_question_classification + - task1555_scitail_answer_generation + - task1532_daily_dialog_emotion_classification + - task239_tweetqa_answer_generation + - task596_mocha_question_generation + - task1411_dart_subject_identification + - task1359_numer_sense_answer_generation + - task329_gap_classification + - task220_rocstories_title_classification + - task316_crows-pairs_classification_stereotype + - task495_semeval_headline_classification + - task1168_brown_coarse_pos_tagging + - task348_squad2.0_unanswerable_question_generation + - task049_multirc_questions_needed_to_answer + - task1534_daily_dialog_question_classification + - task322_jigsaw_classification_threat + - task295_semeval_2020_task4_commonsense_reasoning + - task186_snli_contradiction_to_entailment_text_modification + - task034_winogrande_question_modification_object + - task160_replace_letter_in_a_sentence + - task469_mrqa_answer_generation + - task105_story_cloze-rocstories_sentence_generation + - task649_race_blank_question_generation + - task1536_daily_dialog_happiness_classification + - task683_online_privacy_policy_text_purpose_answer_generation + - task024_cosmosqa_answer_generation + - task584_udeps_eng_fine_pos_tagging + - task066_timetravel_binary_consistency_classification + - task413_mickey_en_sentence_perturbation_generation + - task182_duorc_question_generation + - task028_drop_answer_generation + - task1601_webquestions_answer_generation + - task1295_adversarial_qa_question_answering + - task201_mnli_neutral_classification + - task038_qasc_combined_fact + - task293_storycommonsense_emotion_text_generation + - task572_recipe_nlg_text_generation + - task517_emo_classify_emotion_of_dialogue + - task382_hybridqa_answer_generation + - task176_break_decompose_questions + - task1291_multi_news_summarization + - task155_count_nouns_verbs + - task031_winogrande_question_generation_object + - task279_stereoset_classification_stereotype + - task1336_peixian_equity_evaluation_corpus_gender_classifier + - task508_scruples_dilemmas_more_ethical_isidentifiable + - task518_emo_different_dialogue_emotions + - task077_splash_explanation_to_sql + - task923_event2mind_classifier + - task470_mrqa_question_generation + - task638_multi_woz_classification + - task1412_web_questions_question_answering + - task847_pubmedqa_question_generation + - task678_ollie_actual_relationship_answer_generation + - task290_tellmewhy_question_answerability + - task575_air_dialogue_classification + - task189_snli_neutral_to_contradiction_text_modification + - task026_drop_question_generation + - task162_count_words_starting_with_letter + - task079_conala_concat_strings + - task610_conllpp_ner + - task046_miscellaneous_question_typing + - task197_mnli_domain_answer_generation + - task1325_qa_zre_question_generation_on_subject_relation + - task430_senteval_subject_count + - task672_nummersense + - task402_grailqa_paraphrase_generation + - task904_hate_speech_offensive_classification + - task192_hotpotqa_sentence_generation + - task069_abductivenli_classification + - task574_air_dialogue_sentence_generation + - task187_snli_entailment_to_contradiction_text_modification + - task749_glucose_reverse_cause_emotion_detection + - task1552_scitail_question_generation + - task750_aqua_multiple_choice_answering + - task327_jigsaw_classification_toxic + - task1502_hatexplain_classification + - task328_jigsaw_classification_insult + - task304_numeric_fused_head_resolution + - task1293_kilt_tasks_hotpotqa_question_answering + - task216_rocstories_correct_answer_generation + - task1326_qa_zre_question_generation_from_answer + - task1338_peixian_equity_evaluation_corpus_sentiment_classifier + - task1729_personachat_generate_next + - task1202_atomic_classification_xneed + - task400_paws_paraphrase_classification + - task502_scruples_anecdotes_whoiswrong_verification + - task088_identify_typo_verification + - task221_rocstories_two_choice_classification + - task200_mnli_entailment_classification + - task074_squad1.1_question_generation + - task581_socialiqa_question_generation + - task1186_nne_hrngo_classification + - task898_freebase_qa_answer_generation + - task1408_dart_similarity_classification + - task168_strategyqa_question_decomposition + - task1357_xlsum_summary_generation + - task390_torque_text_span_selection + - task165_mcscript_question_answering_commonsense + - task1533_daily_dialog_formal_classification + - task002_quoref_answer_generation + - task1297_qasc_question_answering + - task305_jeopardy_answer_generation_normal + - task029_winogrande_full_object + - task1327_qa_zre_answer_generation_from_question + - task326_jigsaw_classification_obscene + - task1542_every_ith_element_from_starting + - task570_recipe_nlg_ner_generation + - task1409_dart_text_generation + - task401_numeric_fused_head_reference + - task846_pubmedqa_classification + - task1712_poki_classification + - task344_hybridqa_answer_generation + - task875_emotion_classification + - task1214_atomic_classification_xwant + - task106_scruples_ethical_judgment + - task238_iirc_answer_from_passage_answer_generation + - task1391_winogrande_easy_answer_generation + - task195_sentiment140_classification + - task163_count_words_ending_with_letter + - task579_socialiqa_classification + - task569_recipe_nlg_text_generation + - task1602_webquestion_question_genreation + - task747_glucose_cause_emotion_detection + - task219_rocstories_title_answer_generation + - task178_quartz_question_answering + - task103_facts2story_long_text_generation + - task301_record_question_generation + - task1369_healthfact_sentence_generation + - task515_senteval_odd_word_out + - task496_semeval_answer_generation + - task1658_billsum_summarization + - task1204_atomic_classification_hinderedby + - task1392_superglue_multirc_answer_verification + - task306_jeopardy_answer_generation_double + - task1286_openbookqa_question_answering + - task159_check_frequency_of_words_in_sentence_pair + - task151_tomqa_find_location_easy_clean + - task323_jigsaw_classification_sexually_explicit + - task037_qasc_generate_related_fact + - task027_drop_answer_type_generation + - task1596_event2mind_text_generation_2 + - task141_odd-man-out_classification_category + - task194_duorc_answer_generation + - task679_hope_edi_english_text_classification + - task246_dream_question_generation + - task1195_disflqa_disfluent_to_fluent_conversion + - task065_timetravel_consistent_sentence_classification + - task351_winomt_classification_gender_identifiability_anti + - task580_socialiqa_answer_generation + - task583_udeps_eng_coarse_pos_tagging + - task202_mnli_contradiction_classification + - task222_rocstories_two_chioce_slotting_classification + - task498_scruples_anecdotes_whoiswrong_classification + - task067_abductivenli_answer_generation + - task616_cola_classification + - task286_olid_offense_judgment + - task188_snli_neutral_to_entailment_text_modification + - task223_quartz_explanation_generation + - task820_protoqa_answer_generation + - task196_sentiment140_answer_generation + - task1678_mathqa_answer_selection + - task349_squad2.0_answerable_unanswerable_question_classification + - task154_tomqa_find_location_hard_noise + - task333_hateeval_classification_hate_en + - task235_iirc_question_from_subtext_answer_generation + - task1554_scitail_classification + - task210_logic2text_structured_text_generation + - task035_winogrande_question_modification_person + - task230_iirc_passage_classification + - task1356_xlsum_title_generation + - task1726_mathqa_correct_answer_generation + - task302_record_classification + - task380_boolq_yes_no_question + - task212_logic2text_classification + - task748_glucose_reverse_cause_event_detection + - task834_mathdataset_classification + - task350_winomt_classification_gender_identifiability_pro + - task191_hotpotqa_question_generation + - task236_iirc_question_from_passage_answer_generation + - task217_rocstories_ordering_answer_generation + - task568_circa_question_generation + - task614_glucose_cause_event_detection + - task361_spolin_yesand_prompt_response_classification + - task421_persent_sentence_sentiment_classification + - task203_mnli_sentence_generation + - task420_persent_document_sentiment_classification + - task153_tomqa_find_location_hard_clean + - task346_hybridqa_classification + - task1211_atomic_classification_hassubevent + - task360_spolin_yesand_response_generation + - task510_reddit_tifu_title_summarization + - task511_reddit_tifu_long_text_summarization + - task345_hybridqa_answer_generation + - task270_csrg_counterfactual_context_generation + - task307_jeopardy_answer_generation_final + - task001_quoref_question_generation + - task089_swap_words_verification + - task1196_atomic_classification_oeffect + - task080_piqa_answer_generation + - task1598_nyc_long_text_generation + - task240_tweetqa_question_generation + - task615_moviesqa_answer_generation + - task1347_glue_sts-b_similarity_classification + - task114_is_the_given_word_longest + - task292_storycommonsense_character_text_generation + - task115_help_advice_classification + - task431_senteval_object_count + - task1360_numer_sense_multiple_choice_qa_generation + - task177_para-nmt_paraphrasing + - task132_dais_text_modification + - task269_csrg_counterfactual_story_generation + - task233_iirc_link_exists_classification + - task161_count_words_containing_letter + - task1205_atomic_classification_isafter + - task571_recipe_nlg_ner_generation + - task1292_yelp_review_full_text_categorization + - task428_senteval_inversion + - task311_race_question_generation + - task429_senteval_tense + - task403_creak_commonsense_inference + - task929_products_reviews_classification + - task582_naturalquestion_answer_generation + - task237_iirc_answer_from_subtext_answer_generation + - task050_multirc_answerability + - task184_break_generate_question + - task669_ambigqa_answer_generation + - task169_strategyqa_sentence_generation + - task500_scruples_anecdotes_title_generation + - task241_tweetqa_classification + - task1345_glue_qqp_question_paraprashing + - task218_rocstories_swap_order_answer_generation + - task613_politifact_text_generation + - task1167_penn_treebank_coarse_pos_tagging + - task1422_mathqa_physics + - task247_dream_answer_generation + - task199_mnli_classification + - task164_mcscript_question_answering_text + - task1541_agnews_classification + - task516_senteval_conjoints_inversion + - task294_storycommonsense_motiv_text_generation + - task501_scruples_anecdotes_post_type_verification + - task213_rocstories_correct_ending_classification + - task821_protoqa_question_generation + - task493_review_polarity_classification + - task308_jeopardy_answer_generation_all + - task1595_event2mind_text_generation_1 + - task040_qasc_question_generation + - task231_iirc_link_classification + - task1727_wiqa_what_is_the_effect + - task578_curiosity_dialogs_answer_generation + - task310_race_classification + - task309_race_answer_generation + - task379_agnews_topic_classification + - task030_winogrande_full_person + - task1540_parsed_pdfs_summarization + - task039_qasc_find_overlapping_words + - task1206_atomic_classification_isbefore + - task157_count_vowels_and_consonants + - task339_record_answer_generation + - task453_swag_answer_generation + - task848_pubmedqa_classification + - task673_google_wellformed_query_classification + - task676_ollie_relationship_answer_generation + - task268_casehold_legal_answer_generation + - task844_financial_phrasebank_classification + - task330_gap_answer_generation + - task595_mocha_answer_generation + - task1285_kpa_keypoint_matching + - task234_iirc_passage_line_answer_generation + - task494_review_polarity_answer_generation + - task670_ambigqa_question_generation + - task289_gigaword_summarization + - npr + - nli + - SimpleWiki + - amazon_review_2018 + - ccnews_title_text + - agnews + - xsum + - msmarco + - yahoo_answers_title_answer + - squad_pairs + - wow + - mteb-amazon_counterfactual-avs_triplets + - mteb-amazon_massive_intent-avs_triplets + - mteb-amazon_massive_scenario-avs_triplets + - mteb-amazon_reviews_multi-avs_triplets + - mteb-banking77-avs_triplets + - mteb-emotion-avs_triplets + - mteb-imdb-avs_triplets + - mteb-mtop_domain-avs_triplets + - mteb-mtop_intent-avs_triplets + - mteb-toxic_conversations_50k-avs_triplets + - mteb-tweet_sentiment_extraction-avs_triplets + - covid-bing-query-gpt4-avs_triplets +- **Language:** en +- **License:** apache-2.0 + +### Model Sources + +- **Documentation:** [Sentence Transformers Documentation](https://sbert.net) +- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) +- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) + +### Full Model Architecture + +``` +SentenceTransformer( + (0): Transformer({'max_seq_length': 256, 'do_lower_case': False}) with Transformer model: BertModel + (1): RandomProjection({'in_features': 384, 'out_features': 768, 'seed': 42, 'requires_grad': True}) +) +``` + +## Usage + +### Direct Usage (Sentence Transformers) + +First install the Sentence Transformers library: + +```bash +pip install -U sentence-transformers +``` + +Then you can load this model and run inference. +```python +from sentence_transformers import SentenceTransformer + +# Download from the 🤗 Hub +model = SentenceTransformer("avsolatorio/all-MiniLM-L6-v2-MEDI-MTEB-triplet-randproj-trainable-512-final") +# Run inference +sentences = [ + 'Should I stay or should I go?', + 'The aim of this study was to examine the experiences of parents encountering the critical deterioration and resuscitative care of other children in the pediatric intensive care unit where their own child was admitted.Grounded theory qualitative methodology.Pediatric intensive care unit of a pediatric tertiary care center in Montreal, Canada.Ten parents of critically ill children who witnessed resuscitative measures on another child.None.Semistructured interviews were conducted. While witnessing resuscitation, parents struggled with "Should I stay or should I go?" Their decision depended on specific contributing factors that were intrinsic to parents (curiosity or apprehension, the child\'s sake, trust or distrust) or extrinsic (limited space). These parents were not "spectators." Despite using coping strategies, the experiences were distressing in the majority of cases, although sometimes comforting. The impact on witnessing critical events had divergent effects on parental trust with healthcare professionals.', + 'Several recent studies suggest that acceleration of the head at impact during sporting activities may have a detrimental effect on cognitive function. Reducing acceleration of impact in these sports could reduce neurologic sequelae.To measure the effectiveness of a regulation football helmet to reduce acceleration of impact for both low- and moderate-force impacts.An experimental paired study design was used. Male volunteers between 16 and 30 years of age headed soccer balls traveling approximately 35 miles per hour bareheaded and with a helmet. An intraoral accelerometer worn inside a plastic mouthpiece measured acceleration of the head. The helmet also had an accelerometer placed inside the padding. For more forceful impacts, cadaver heads, both with and without helmets, were instrumented with intraoral (IO) and intracranial (IC) accelerometers and struck with a pendulum device. Simultaneous IO and IC accelerations were measured and compared between helmeted and unhelmeted cadaver heads. The main outcome was mean peak acceleration of the head and/or brain associated with low- and moderate-force impacts with and without protective headgear.Mean peak Gs, measured by the mouthpiece accelerometer, were significantly reduced when the participants heading soccer balls were wearing a helmet (7.7 Gs with vs 19.2 Gs without, p = 0.01). Wearing a helmet also significantly lowered the peak Gs measured intraorally and intracranially in cadavers subjected to moderate-force pendulum impacts: 28.7 Gs with vs 62.6 Gs without, p<0.001; and 56.4 Gs with vs 81.6 Gs without, p<0.001, respectively.', +] +embeddings = model.encode(sentences) +print(embeddings.shape) +# [3, 768] + +# Get the similarity scores for the embeddings +similarities = model.similarity(embeddings, embeddings) +print(similarities.shape) +# [3, 3] +``` + + + + + + + +## Evaluation + +### Metrics + +#### Triplet + +* Dataset: `medi-mteb-dev` +* Evaluated with [TripletEvaluator](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.TripletEvaluator) + +| Metric | Value | +|:--------------------|:-----------| +| **cosine_accuracy** | **0.9145** | + + + + + +## Training Details + +### Training Datasets + +#### NQ + +* Dataset: NQ +* Size: 49,676 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### pubmed + +* Dataset: pubmed +* Size: 29,908 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### specter_train_triples + +* Dataset: specter_train_triples +* Size: 49,676 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### S2ORC_citations_abstracts + +* Dataset: S2ORC_citations_abstracts +* Size: 99,352 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### fever + +* Dataset: fever +* Size: 74,514 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### gooaq_pairs + +* Dataset: gooaq_pairs +* Size: 24,838 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### codesearchnet + +* Dataset: codesearchnet +* Size: 15,210 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### wikihow + +* Dataset: wikihow +* Size: 5,070 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### WikiAnswers + +* Dataset: WikiAnswers +* Size: 24,838 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### eli5_question_answer + +* Dataset: eli5_question_answer +* Size: 24,838 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### amazon-qa + +* Dataset: amazon-qa +* Size: 99,352 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### medmcqa + +* Dataset: medmcqa +* Size: 29,908 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### zeroshot + +* Dataset: zeroshot +* Size: 15,210 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### TriviaQA_pairs + +* Dataset: TriviaQA_pairs +* Size: 49,676 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### PAQ_pairs + +* Dataset: PAQ_pairs +* Size: 24,838 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### stackexchange_duplicate_questions_title-body_title-body + +* Dataset: stackexchange_duplicate_questions_title-body_title-body +* Size: 24,838 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### trex + +* Dataset: trex +* Size: 29,908 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### flickr30k_captions + +* Dataset: flickr30k_captions +* Size: 24,838 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### hotpotqa + +* Dataset: hotpotqa +* Size: 40,048 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task671_ambigqa_text_generation + +* Dataset: task671_ambigqa_text_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task061_ropes_answer_generation + +* Dataset: task061_ropes_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task285_imdb_answer_generation + +* Dataset: task285_imdb_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task905_hate_speech_offensive_classification + +* Dataset: task905_hate_speech_offensive_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task566_circa_classification + +* Dataset: task566_circa_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task184_snli_entailment_to_neutral_text_modification + +* Dataset: task184_snli_entailment_to_neutral_text_modification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task280_stereoset_classification_stereotype_type + +* Dataset: task280_stereoset_classification_stereotype_type +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1599_smcalflow_classification + +* Dataset: task1599_smcalflow_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1384_deal_or_no_dialog_classification + +* Dataset: task1384_deal_or_no_dialog_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task591_sciq_answer_generation + +* Dataset: task591_sciq_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task823_peixian-rtgender_sentiment_analysis + +* Dataset: task823_peixian-rtgender_sentiment_analysis +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task023_cosmosqa_question_generation + +* Dataset: task023_cosmosqa_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task900_freebase_qa_category_classification + +* Dataset: task900_freebase_qa_category_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task924_event2mind_word_generation + +* Dataset: task924_event2mind_word_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task152_tomqa_find_location_easy_noise + +* Dataset: task152_tomqa_find_location_easy_noise +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1368_healthfact_sentence_generation + +* Dataset: task1368_healthfact_sentence_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1661_super_glue_classification + +* Dataset: task1661_super_glue_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1187_politifact_classification + +* Dataset: task1187_politifact_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1728_web_nlg_data_to_text + +* Dataset: task1728_web_nlg_data_to_text +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task112_asset_simple_sentence_identification + +* Dataset: task112_asset_simple_sentence_identification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1340_msr_text_compression_compression + +* Dataset: task1340_msr_text_compression_compression +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task072_abductivenli_answer_generation + +* Dataset: task072_abductivenli_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1504_hatexplain_answer_generation + +* Dataset: task1504_hatexplain_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task684_online_privacy_policy_text_information_type_generation + +* Dataset: task684_online_privacy_policy_text_information_type_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1290_xsum_summarization + +* Dataset: task1290_xsum_summarization +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task075_squad1.1_answer_generation + +* Dataset: task075_squad1.1_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1587_scifact_classification + +* Dataset: task1587_scifact_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task384_socialiqa_question_classification + +* Dataset: task384_socialiqa_question_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1555_scitail_answer_generation + +* Dataset: task1555_scitail_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1532_daily_dialog_emotion_classification + +* Dataset: task1532_daily_dialog_emotion_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task239_tweetqa_answer_generation + +* Dataset: task239_tweetqa_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task596_mocha_question_generation + +* Dataset: task596_mocha_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1411_dart_subject_identification + +* Dataset: task1411_dart_subject_identification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1359_numer_sense_answer_generation + +* Dataset: task1359_numer_sense_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task329_gap_classification + +* Dataset: task329_gap_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task220_rocstories_title_classification + +* Dataset: task220_rocstories_title_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task316_crows-pairs_classification_stereotype + +* Dataset: task316_crows-pairs_classification_stereotype +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task495_semeval_headline_classification + +* Dataset: task495_semeval_headline_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1168_brown_coarse_pos_tagging + +* Dataset: task1168_brown_coarse_pos_tagging +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task348_squad2.0_unanswerable_question_generation + +* Dataset: task348_squad2.0_unanswerable_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task049_multirc_questions_needed_to_answer + +* Dataset: task049_multirc_questions_needed_to_answer +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1534_daily_dialog_question_classification + +* Dataset: task1534_daily_dialog_question_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task322_jigsaw_classification_threat + +* Dataset: task322_jigsaw_classification_threat +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task295_semeval_2020_task4_commonsense_reasoning + +* Dataset: task295_semeval_2020_task4_commonsense_reasoning +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task186_snli_contradiction_to_entailment_text_modification + +* Dataset: task186_snli_contradiction_to_entailment_text_modification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task034_winogrande_question_modification_object + +* Dataset: task034_winogrande_question_modification_object +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task160_replace_letter_in_a_sentence + +* Dataset: task160_replace_letter_in_a_sentence +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task469_mrqa_answer_generation + +* Dataset: task469_mrqa_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task105_story_cloze-rocstories_sentence_generation + +* Dataset: task105_story_cloze-rocstories_sentence_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task649_race_blank_question_generation + +* Dataset: task649_race_blank_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1536_daily_dialog_happiness_classification + +* Dataset: task1536_daily_dialog_happiness_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task683_online_privacy_policy_text_purpose_answer_generation + +* Dataset: task683_online_privacy_policy_text_purpose_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task024_cosmosqa_answer_generation + +* Dataset: task024_cosmosqa_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task584_udeps_eng_fine_pos_tagging + +* Dataset: task584_udeps_eng_fine_pos_tagging +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task066_timetravel_binary_consistency_classification + +* Dataset: task066_timetravel_binary_consistency_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task413_mickey_en_sentence_perturbation_generation + +* Dataset: task413_mickey_en_sentence_perturbation_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task182_duorc_question_generation + +* Dataset: task182_duorc_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task028_drop_answer_generation + +* Dataset: task028_drop_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1601_webquestions_answer_generation + +* Dataset: task1601_webquestions_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1295_adversarial_qa_question_answering + +* Dataset: task1295_adversarial_qa_question_answering +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task201_mnli_neutral_classification + +* Dataset: task201_mnli_neutral_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task038_qasc_combined_fact + +* Dataset: task038_qasc_combined_fact +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task293_storycommonsense_emotion_text_generation + +* Dataset: task293_storycommonsense_emotion_text_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task572_recipe_nlg_text_generation + +* Dataset: task572_recipe_nlg_text_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task517_emo_classify_emotion_of_dialogue + +* Dataset: task517_emo_classify_emotion_of_dialogue +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task382_hybridqa_answer_generation + +* Dataset: task382_hybridqa_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task176_break_decompose_questions + +* Dataset: task176_break_decompose_questions +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1291_multi_news_summarization + +* Dataset: task1291_multi_news_summarization +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task155_count_nouns_verbs + +* Dataset: task155_count_nouns_verbs +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task031_winogrande_question_generation_object + +* Dataset: task031_winogrande_question_generation_object +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:--------------------------------------------------------------------------------|:---------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task279_stereoset_classification_stereotype + +* Dataset: task279_stereoset_classification_stereotype +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1336_peixian_equity_evaluation_corpus_gender_classifier + +* Dataset: task1336_peixian_equity_evaluation_corpus_gender_classifier +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task508_scruples_dilemmas_more_ethical_isidentifiable + +* Dataset: task508_scruples_dilemmas_more_ethical_isidentifiable +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task518_emo_different_dialogue_emotions + +* Dataset: task518_emo_different_dialogue_emotions +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task077_splash_explanation_to_sql + +* Dataset: task077_splash_explanation_to_sql +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task923_event2mind_classifier + +* Dataset: task923_event2mind_classifier +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task470_mrqa_question_generation + +* Dataset: task470_mrqa_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task638_multi_woz_classification + +* Dataset: task638_multi_woz_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1412_web_questions_question_answering + +* Dataset: task1412_web_questions_question_answering +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task847_pubmedqa_question_generation + +* Dataset: task847_pubmedqa_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task678_ollie_actual_relationship_answer_generation + +* Dataset: task678_ollie_actual_relationship_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task290_tellmewhy_question_answerability + +* Dataset: task290_tellmewhy_question_answerability +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task575_air_dialogue_classification + +* Dataset: task575_air_dialogue_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task189_snli_neutral_to_contradiction_text_modification + +* Dataset: task189_snli_neutral_to_contradiction_text_modification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task026_drop_question_generation + +* Dataset: task026_drop_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task162_count_words_starting_with_letter + +* Dataset: task162_count_words_starting_with_letter +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task079_conala_concat_strings + +* Dataset: task079_conala_concat_strings +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task610_conllpp_ner + +* Dataset: task610_conllpp_ner +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task046_miscellaneous_question_typing + +* Dataset: task046_miscellaneous_question_typing +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task197_mnli_domain_answer_generation + +* Dataset: task197_mnli_domain_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1325_qa_zre_question_generation_on_subject_relation + +* Dataset: task1325_qa_zre_question_generation_on_subject_relation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task430_senteval_subject_count + +* Dataset: task430_senteval_subject_count +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task672_nummersense + +* Dataset: task672_nummersense +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task402_grailqa_paraphrase_generation + +* Dataset: task402_grailqa_paraphrase_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task904_hate_speech_offensive_classification + +* Dataset: task904_hate_speech_offensive_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task192_hotpotqa_sentence_generation + +* Dataset: task192_hotpotqa_sentence_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task069_abductivenli_classification + +* Dataset: task069_abductivenli_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task574_air_dialogue_sentence_generation + +* Dataset: task574_air_dialogue_sentence_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task187_snli_entailment_to_contradiction_text_modification + +* Dataset: task187_snli_entailment_to_contradiction_text_modification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task749_glucose_reverse_cause_emotion_detection + +* Dataset: task749_glucose_reverse_cause_emotion_detection +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1552_scitail_question_generation + +* Dataset: task1552_scitail_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task750_aqua_multiple_choice_answering + +* Dataset: task750_aqua_multiple_choice_answering +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task327_jigsaw_classification_toxic + +* Dataset: task327_jigsaw_classification_toxic +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1502_hatexplain_classification + +* Dataset: task1502_hatexplain_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task328_jigsaw_classification_insult + +* Dataset: task328_jigsaw_classification_insult +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task304_numeric_fused_head_resolution + +* Dataset: task304_numeric_fused_head_resolution +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1293_kilt_tasks_hotpotqa_question_answering + +* Dataset: task1293_kilt_tasks_hotpotqa_question_answering +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task216_rocstories_correct_answer_generation + +* Dataset: task216_rocstories_correct_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1326_qa_zre_question_generation_from_answer + +* Dataset: task1326_qa_zre_question_generation_from_answer +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1338_peixian_equity_evaluation_corpus_sentiment_classifier + +* Dataset: task1338_peixian_equity_evaluation_corpus_sentiment_classifier +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:--------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1729_personachat_generate_next + +* Dataset: task1729_personachat_generate_next +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1202_atomic_classification_xneed + +* Dataset: task1202_atomic_classification_xneed +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task400_paws_paraphrase_classification + +* Dataset: task400_paws_paraphrase_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task502_scruples_anecdotes_whoiswrong_verification + +* Dataset: task502_scruples_anecdotes_whoiswrong_verification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task088_identify_typo_verification + +* Dataset: task088_identify_typo_verification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task221_rocstories_two_choice_classification + +* Dataset: task221_rocstories_two_choice_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task200_mnli_entailment_classification + +* Dataset: task200_mnli_entailment_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task074_squad1.1_question_generation + +* Dataset: task074_squad1.1_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task581_socialiqa_question_generation + +* Dataset: task581_socialiqa_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1186_nne_hrngo_classification + +* Dataset: task1186_nne_hrngo_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task898_freebase_qa_answer_generation + +* Dataset: task898_freebase_qa_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1408_dart_similarity_classification + +* Dataset: task1408_dart_similarity_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task168_strategyqa_question_decomposition + +* Dataset: task168_strategyqa_question_decomposition +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1357_xlsum_summary_generation + +* Dataset: task1357_xlsum_summary_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task390_torque_text_span_selection + +* Dataset: task390_torque_text_span_selection +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task165_mcscript_question_answering_commonsense + +* Dataset: task165_mcscript_question_answering_commonsense +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1533_daily_dialog_formal_classification + +* Dataset: task1533_daily_dialog_formal_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task002_quoref_answer_generation + +* Dataset: task002_quoref_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1297_qasc_question_answering + +* Dataset: task1297_qasc_question_answering +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task305_jeopardy_answer_generation_normal + +* Dataset: task305_jeopardy_answer_generation_normal +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task029_winogrande_full_object + +* Dataset: task029_winogrande_full_object +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1327_qa_zre_answer_generation_from_question + +* Dataset: task1327_qa_zre_answer_generation_from_question +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task326_jigsaw_classification_obscene + +* Dataset: task326_jigsaw_classification_obscene +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1542_every_ith_element_from_starting + +* Dataset: task1542_every_ith_element_from_starting +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task570_recipe_nlg_ner_generation + +* Dataset: task570_recipe_nlg_ner_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1409_dart_text_generation + +* Dataset: task1409_dart_text_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task401_numeric_fused_head_reference + +* Dataset: task401_numeric_fused_head_reference +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task846_pubmedqa_classification + +* Dataset: task846_pubmedqa_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1712_poki_classification + +* Dataset: task1712_poki_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task344_hybridqa_answer_generation + +* Dataset: task344_hybridqa_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task875_emotion_classification + +* Dataset: task875_emotion_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1214_atomic_classification_xwant + +* Dataset: task1214_atomic_classification_xwant +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task106_scruples_ethical_judgment + +* Dataset: task106_scruples_ethical_judgment +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task238_iirc_answer_from_passage_answer_generation + +* Dataset: task238_iirc_answer_from_passage_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1391_winogrande_easy_answer_generation + +* Dataset: task1391_winogrande_easy_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task195_sentiment140_classification + +* Dataset: task195_sentiment140_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task163_count_words_ending_with_letter + +* Dataset: task163_count_words_ending_with_letter +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task579_socialiqa_classification + +* Dataset: task579_socialiqa_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task569_recipe_nlg_text_generation + +* Dataset: task569_recipe_nlg_text_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1602_webquestion_question_genreation + +* Dataset: task1602_webquestion_question_genreation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task747_glucose_cause_emotion_detection + +* Dataset: task747_glucose_cause_emotion_detection +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task219_rocstories_title_answer_generation + +* Dataset: task219_rocstories_title_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task178_quartz_question_answering + +* Dataset: task178_quartz_question_answering +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task103_facts2story_long_text_generation + +* Dataset: task103_facts2story_long_text_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task301_record_question_generation + +* Dataset: task301_record_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1369_healthfact_sentence_generation + +* Dataset: task1369_healthfact_sentence_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task515_senteval_odd_word_out + +* Dataset: task515_senteval_odd_word_out +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task496_semeval_answer_generation + +* Dataset: task496_semeval_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1658_billsum_summarization + +* Dataset: task1658_billsum_summarization +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1204_atomic_classification_hinderedby + +* Dataset: task1204_atomic_classification_hinderedby +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1392_superglue_multirc_answer_verification + +* Dataset: task1392_superglue_multirc_answer_verification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task306_jeopardy_answer_generation_double + +* Dataset: task306_jeopardy_answer_generation_double +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1286_openbookqa_question_answering + +* Dataset: task1286_openbookqa_question_answering +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task159_check_frequency_of_words_in_sentence_pair + +* Dataset: task159_check_frequency_of_words_in_sentence_pair +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task151_tomqa_find_location_easy_clean + +* Dataset: task151_tomqa_find_location_easy_clean +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task323_jigsaw_classification_sexually_explicit + +* Dataset: task323_jigsaw_classification_sexually_explicit +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task037_qasc_generate_related_fact + +* Dataset: task037_qasc_generate_related_fact +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task027_drop_answer_type_generation + +* Dataset: task027_drop_answer_type_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1596_event2mind_text_generation_2 + +* Dataset: task1596_event2mind_text_generation_2 +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task141_odd-man-out_classification_category + +* Dataset: task141_odd-man-out_classification_category +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task194_duorc_answer_generation + +* Dataset: task194_duorc_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task679_hope_edi_english_text_classification + +* Dataset: task679_hope_edi_english_text_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task246_dream_question_generation + +* Dataset: task246_dream_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1195_disflqa_disfluent_to_fluent_conversion + +* Dataset: task1195_disflqa_disfluent_to_fluent_conversion +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task065_timetravel_consistent_sentence_classification + +* Dataset: task065_timetravel_consistent_sentence_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task351_winomt_classification_gender_identifiability_anti + +* Dataset: task351_winomt_classification_gender_identifiability_anti +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task580_socialiqa_answer_generation + +* Dataset: task580_socialiqa_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task583_udeps_eng_coarse_pos_tagging + +* Dataset: task583_udeps_eng_coarse_pos_tagging +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task202_mnli_contradiction_classification + +* Dataset: task202_mnli_contradiction_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task222_rocstories_two_chioce_slotting_classification + +* Dataset: task222_rocstories_two_chioce_slotting_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task498_scruples_anecdotes_whoiswrong_classification + +* Dataset: task498_scruples_anecdotes_whoiswrong_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task067_abductivenli_answer_generation + +* Dataset: task067_abductivenli_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task616_cola_classification + +* Dataset: task616_cola_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task286_olid_offense_judgment + +* Dataset: task286_olid_offense_judgment +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task188_snli_neutral_to_entailment_text_modification + +* Dataset: task188_snli_neutral_to_entailment_text_modification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task223_quartz_explanation_generation + +* Dataset: task223_quartz_explanation_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task820_protoqa_answer_generation + +* Dataset: task820_protoqa_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task196_sentiment140_answer_generation + +* Dataset: task196_sentiment140_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1678_mathqa_answer_selection + +* Dataset: task1678_mathqa_answer_selection +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task349_squad2.0_answerable_unanswerable_question_classification + +* Dataset: task349_squad2.0_answerable_unanswerable_question_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task154_tomqa_find_location_hard_noise + +* Dataset: task154_tomqa_find_location_hard_noise +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task333_hateeval_classification_hate_en + +* Dataset: task333_hateeval_classification_hate_en +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task235_iirc_question_from_subtext_answer_generation + +* Dataset: task235_iirc_question_from_subtext_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1554_scitail_classification + +* Dataset: task1554_scitail_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task210_logic2text_structured_text_generation + +* Dataset: task210_logic2text_structured_text_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task035_winogrande_question_modification_person + +* Dataset: task035_winogrande_question_modification_person +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task230_iirc_passage_classification + +* Dataset: task230_iirc_passage_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1356_xlsum_title_generation + +* Dataset: task1356_xlsum_title_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1726_mathqa_correct_answer_generation + +* Dataset: task1726_mathqa_correct_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task302_record_classification + +* Dataset: task302_record_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task380_boolq_yes_no_question + +* Dataset: task380_boolq_yes_no_question +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task212_logic2text_classification + +* Dataset: task212_logic2text_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task748_glucose_reverse_cause_event_detection + +* Dataset: task748_glucose_reverse_cause_event_detection +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task834_mathdataset_classification + +* Dataset: task834_mathdataset_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task350_winomt_classification_gender_identifiability_pro + +* Dataset: task350_winomt_classification_gender_identifiability_pro +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task191_hotpotqa_question_generation + +* Dataset: task191_hotpotqa_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task236_iirc_question_from_passage_answer_generation + +* Dataset: task236_iirc_question_from_passage_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task217_rocstories_ordering_answer_generation + +* Dataset: task217_rocstories_ordering_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task568_circa_question_generation + +* Dataset: task568_circa_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task614_glucose_cause_event_detection + +* Dataset: task614_glucose_cause_event_detection +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task361_spolin_yesand_prompt_response_classification + +* Dataset: task361_spolin_yesand_prompt_response_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task421_persent_sentence_sentiment_classification + +* Dataset: task421_persent_sentence_sentiment_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task203_mnli_sentence_generation + +* Dataset: task203_mnli_sentence_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task420_persent_document_sentiment_classification + +* Dataset: task420_persent_document_sentiment_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task153_tomqa_find_location_hard_clean + +* Dataset: task153_tomqa_find_location_hard_clean +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task346_hybridqa_classification + +* Dataset: task346_hybridqa_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1211_atomic_classification_hassubevent + +* Dataset: task1211_atomic_classification_hassubevent +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task360_spolin_yesand_response_generation + +* Dataset: task360_spolin_yesand_response_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task510_reddit_tifu_title_summarization + +* Dataset: task510_reddit_tifu_title_summarization +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task511_reddit_tifu_long_text_summarization + +* Dataset: task511_reddit_tifu_long_text_summarization +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task345_hybridqa_answer_generation + +* Dataset: task345_hybridqa_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task270_csrg_counterfactual_context_generation + +* Dataset: task270_csrg_counterfactual_context_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task307_jeopardy_answer_generation_final + +* Dataset: task307_jeopardy_answer_generation_final +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task001_quoref_question_generation + +* Dataset: task001_quoref_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task089_swap_words_verification + +* Dataset: task089_swap_words_verification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1196_atomic_classification_oeffect + +* Dataset: task1196_atomic_classification_oeffect +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task080_piqa_answer_generation + +* Dataset: task080_piqa_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1598_nyc_long_text_generation + +* Dataset: task1598_nyc_long_text_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task240_tweetqa_question_generation + +* Dataset: task240_tweetqa_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task615_moviesqa_answer_generation + +* Dataset: task615_moviesqa_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1347_glue_sts-b_similarity_classification + +* Dataset: task1347_glue_sts-b_similarity_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task114_is_the_given_word_longest + +* Dataset: task114_is_the_given_word_longest +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task292_storycommonsense_character_text_generation + +* Dataset: task292_storycommonsense_character_text_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task115_help_advice_classification + +* Dataset: task115_help_advice_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task431_senteval_object_count + +* Dataset: task431_senteval_object_count +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1360_numer_sense_multiple_choice_qa_generation + +* Dataset: task1360_numer_sense_multiple_choice_qa_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task177_para-nmt_paraphrasing + +* Dataset: task177_para-nmt_paraphrasing +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task132_dais_text_modification + +* Dataset: task132_dais_text_modification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:--------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task269_csrg_counterfactual_story_generation + +* Dataset: task269_csrg_counterfactual_story_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task233_iirc_link_exists_classification + +* Dataset: task233_iirc_link_exists_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task161_count_words_containing_letter + +* Dataset: task161_count_words_containing_letter +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1205_atomic_classification_isafter + +* Dataset: task1205_atomic_classification_isafter +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task571_recipe_nlg_ner_generation + +* Dataset: task571_recipe_nlg_ner_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1292_yelp_review_full_text_categorization + +* Dataset: task1292_yelp_review_full_text_categorization +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task428_senteval_inversion + +* Dataset: task428_senteval_inversion +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task311_race_question_generation + +* Dataset: task311_race_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task429_senteval_tense + +* Dataset: task429_senteval_tense +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task403_creak_commonsense_inference + +* Dataset: task403_creak_commonsense_inference +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task929_products_reviews_classification + +* Dataset: task929_products_reviews_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task582_naturalquestion_answer_generation + +* Dataset: task582_naturalquestion_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task237_iirc_answer_from_subtext_answer_generation + +* Dataset: task237_iirc_answer_from_subtext_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task050_multirc_answerability + +* Dataset: task050_multirc_answerability +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task184_break_generate_question + +* Dataset: task184_break_generate_question +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task669_ambigqa_answer_generation + +* Dataset: task669_ambigqa_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task169_strategyqa_sentence_generation + +* Dataset: task169_strategyqa_sentence_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task500_scruples_anecdotes_title_generation + +* Dataset: task500_scruples_anecdotes_title_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task241_tweetqa_classification + +* Dataset: task241_tweetqa_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1345_glue_qqp_question_paraprashing + +* Dataset: task1345_glue_qqp_question_paraprashing +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task218_rocstories_swap_order_answer_generation + +* Dataset: task218_rocstories_swap_order_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task613_politifact_text_generation + +* Dataset: task613_politifact_text_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1167_penn_treebank_coarse_pos_tagging + +* Dataset: task1167_penn_treebank_coarse_pos_tagging +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1422_mathqa_physics + +* Dataset: task1422_mathqa_physics +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task247_dream_answer_generation + +* Dataset: task247_dream_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task199_mnli_classification + +* Dataset: task199_mnli_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task164_mcscript_question_answering_text + +* Dataset: task164_mcscript_question_answering_text +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1541_agnews_classification + +* Dataset: task1541_agnews_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task516_senteval_conjoints_inversion + +* Dataset: task516_senteval_conjoints_inversion +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task294_storycommonsense_motiv_text_generation + +* Dataset: task294_storycommonsense_motiv_text_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task501_scruples_anecdotes_post_type_verification + +* Dataset: task501_scruples_anecdotes_post_type_verification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task213_rocstories_correct_ending_classification + +* Dataset: task213_rocstories_correct_ending_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task821_protoqa_question_generation + +* Dataset: task821_protoqa_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task493_review_polarity_classification + +* Dataset: task493_review_polarity_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task308_jeopardy_answer_generation_all + +* Dataset: task308_jeopardy_answer_generation_all +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1595_event2mind_text_generation_1 + +* Dataset: task1595_event2mind_text_generation_1 +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task040_qasc_question_generation + +* Dataset: task040_qasc_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task231_iirc_link_classification + +* Dataset: task231_iirc_link_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1727_wiqa_what_is_the_effect + +* Dataset: task1727_wiqa_what_is_the_effect +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task578_curiosity_dialogs_answer_generation + +* Dataset: task578_curiosity_dialogs_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task310_race_classification + +* Dataset: task310_race_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task309_race_answer_generation + +* Dataset: task309_race_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task379_agnews_topic_classification + +* Dataset: task379_agnews_topic_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task030_winogrande_full_person + +* Dataset: task030_winogrande_full_person +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1540_parsed_pdfs_summarization + +* Dataset: task1540_parsed_pdfs_summarization +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task039_qasc_find_overlapping_words + +* Dataset: task039_qasc_find_overlapping_words +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1206_atomic_classification_isbefore + +* Dataset: task1206_atomic_classification_isbefore +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task157_count_vowels_and_consonants + +* Dataset: task157_count_vowels_and_consonants +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task339_record_answer_generation + +* Dataset: task339_record_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task453_swag_answer_generation + +* Dataset: task453_swag_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task848_pubmedqa_classification + +* Dataset: task848_pubmedqa_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task673_google_wellformed_query_classification + +* Dataset: task673_google_wellformed_query_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task676_ollie_relationship_answer_generation + +* Dataset: task676_ollie_relationship_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task268_casehold_legal_answer_generation + +* Dataset: task268_casehold_legal_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task844_financial_phrasebank_classification + +* Dataset: task844_financial_phrasebank_classification +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task330_gap_answer_generation + +* Dataset: task330_gap_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task595_mocha_answer_generation + +* Dataset: task595_mocha_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task1285_kpa_keypoint_matching + +* Dataset: task1285_kpa_keypoint_matching +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task234_iirc_passage_line_answer_generation + +* Dataset: task234_iirc_passage_line_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task494_review_polarity_answer_generation + +* Dataset: task494_review_polarity_answer_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task670_ambigqa_question_generation + +* Dataset: task670_ambigqa_question_generation +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### task289_gigaword_summarization + +* Dataset: task289_gigaword_summarization +* Size: 1,018 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### npr + +* Dataset: npr +* Size: 24,838 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### nli + +* Dataset: nli +* Size: 49,676 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### SimpleWiki + +* Dataset: SimpleWiki +* Size: 5,070 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### amazon_review_2018 + +* Dataset: amazon_review_2018 +* Size: 99,352 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### ccnews_title_text + +* Dataset: ccnews_title_text +* Size: 24,838 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### agnews + +* Dataset: agnews +* Size: 44,606 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### xsum + +* Dataset: xsum +* Size: 10,140 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### msmarco + +* Dataset: msmarco +* Size: 173,354 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### yahoo_answers_title_answer + +* Dataset: yahoo_answers_title_answer +* Size: 24,838 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### squad_pairs + +* Dataset: squad_pairs +* Size: 24,838 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### wow + +* Dataset: wow +* Size: 29,908 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### mteb-amazon_counterfactual-avs_triplets + +* Dataset: mteb-amazon_counterfactual-avs_triplets +* Size: 4,055 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### mteb-amazon_massive_intent-avs_triplets + +* Dataset: mteb-amazon_massive_intent-avs_triplets +* Size: 11,661 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### mteb-amazon_massive_scenario-avs_triplets + +* Dataset: mteb-amazon_massive_scenario-avs_triplets +* Size: 11,661 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### mteb-amazon_reviews_multi-avs_triplets + +* Dataset: mteb-amazon_reviews_multi-avs_triplets +* Size: 198,192 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### mteb-banking77-avs_triplets + +* Dataset: mteb-banking77-avs_triplets +* Size: 10,139 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### mteb-emotion-avs_triplets + +* Dataset: mteb-emotion-avs_triplets +* Size: 16,224 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### mteb-imdb-avs_triplets + +* Dataset: mteb-imdb-avs_triplets +* Size: 24,839 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### mteb-mtop_domain-avs_triplets + +* Dataset: mteb-mtop_domain-avs_triplets +* Size: 15,715 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### mteb-mtop_intent-avs_triplets + +* Dataset: mteb-mtop_intent-avs_triplets +* Size: 15,715 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### mteb-toxic_conversations_50k-avs_triplets + +* Dataset: mteb-toxic_conversations_50k-avs_triplets +* Size: 49,677 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### mteb-tweet_sentiment_extraction-avs_triplets + +* Dataset: mteb-tweet_sentiment_extraction-avs_triplets +* Size: 27,373 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +#### covid-bing-query-gpt4-avs_triplets + +* Dataset: covid-bing-query-gpt4-avs_triplets +* Size: 5,070 training samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:---------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +### Evaluation Dataset + +#### Unnamed Dataset + + +* Size: 18,269 evaluation samples +* Columns: anchor, positive, and negative +* Approximate statistics based on the first 1000 samples: + | | anchor | positive | negative | + |:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| + | type | string | string | string | + | details | | | | +* Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: + ```json + { + "scale": 20.0, + "similarity_fct": "cos_sim" + } + ``` + +### Training Hyperparameters +#### Non-Default Hyperparameters + +- `eval_strategy`: steps +- `per_device_train_batch_size`: 512 +- `per_device_eval_batch_size`: 512 +- `learning_rate`: 5.656854249492381e-05 +- `num_train_epochs`: 10 +- `warmup_ratio`: 0.1 +- `fp16`: True +- `gradient_checkpointing`: True +- `batch_sampler`: no_duplicates + +#### All Hyperparameters +
Click to expand + +- `overwrite_output_dir`: False +- `do_predict`: False +- `eval_strategy`: steps +- `prediction_loss_only`: True +- `per_device_train_batch_size`: 512 +- `per_device_eval_batch_size`: 512 +- `per_gpu_train_batch_size`: None +- `per_gpu_eval_batch_size`: None +- `gradient_accumulation_steps`: 1 +- `eval_accumulation_steps`: None +- `torch_empty_cache_steps`: None +- `learning_rate`: 5.656854249492381e-05 +- `weight_decay`: 0.0 +- `adam_beta1`: 0.9 +- `adam_beta2`: 0.999 +- `adam_epsilon`: 1e-08 +- `max_grad_norm`: 1.0 +- `num_train_epochs`: 10 +- `max_steps`: -1 +- `lr_scheduler_type`: linear +- `lr_scheduler_kwargs`: {} +- `warmup_ratio`: 0.1 +- `warmup_steps`: 0 +- `log_level`: passive +- `log_level_replica`: warning +- `log_on_each_node`: True +- `logging_nan_inf_filter`: True +- `save_safetensors`: True +- `save_on_each_node`: False +- `save_only_model`: False +- `restore_callback_states_from_checkpoint`: False +- `no_cuda`: False +- `use_cpu`: False +- `use_mps_device`: False +- `seed`: 42 +- `data_seed`: None +- `jit_mode_eval`: False +- `use_ipex`: False +- `bf16`: False +- `fp16`: True +- `fp16_opt_level`: O1 +- `half_precision_backend`: auto +- `bf16_full_eval`: False +- `fp16_full_eval`: False +- `tf32`: None +- `local_rank`: 0 +- `ddp_backend`: None +- `tpu_num_cores`: None +- `tpu_metrics_debug`: False +- `debug`: [] +- `dataloader_drop_last`: False +- `dataloader_num_workers`: 0 +- `dataloader_prefetch_factor`: None +- `past_index`: -1 +- `disable_tqdm`: False +- `remove_unused_columns`: True +- `label_names`: None +- `load_best_model_at_end`: False +- `ignore_data_skip`: False +- `fsdp`: [] +- `fsdp_min_num_params`: 0 +- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} +- `fsdp_transformer_layer_cls_to_wrap`: None +- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} +- `deepspeed`: None +- `label_smoothing_factor`: 0.0 +- `optim`: adamw_torch +- `optim_args`: None +- `adafactor`: False +- `group_by_length`: False +- `length_column_name`: length +- `ddp_find_unused_parameters`: None +- `ddp_bucket_cap_mb`: None +- `ddp_broadcast_buffers`: False +- `dataloader_pin_memory`: True +- `dataloader_persistent_workers`: False +- `skip_memory_metrics`: True +- `use_legacy_prediction_loop`: False +- `push_to_hub`: False +- `resume_from_checkpoint`: None +- `hub_model_id`: None +- `hub_strategy`: every_save +- `hub_private_repo`: False +- `hub_always_push`: False +- `gradient_checkpointing`: True +- `gradient_checkpointing_kwargs`: None +- `include_inputs_for_metrics`: False +- `include_for_metrics`: [] +- `eval_do_concat_batches`: True +- `fp16_backend`: auto +- `push_to_hub_model_id`: None +- `push_to_hub_organization`: None +- `mp_parameters`: +- `auto_find_batch_size`: False +- `full_determinism`: False +- `torchdynamo`: None +- `ray_scope`: last +- `ddp_timeout`: 1800 +- `torch_compile`: False +- `torch_compile_backend`: None +- `torch_compile_mode`: None +- `dispatch_batches`: None +- `split_batches`: None +- `include_tokens_per_second`: False +- `include_num_input_tokens_seen`: False +- `neftune_noise_alpha`: None +- `optim_target_modules`: None +- `batch_eval_metrics`: False +- `eval_on_start`: False +- `use_liger_kernel`: False +- `eval_use_gather_object`: False +- `average_tokens_across_devices`: False +- `prompts`: None +- `batch_sampler`: no_duplicates +- `multi_dataset_batch_sampler`: proportional + +
+ +### Training Logs +| Epoch | Step | Training Loss | Validation Loss | medi-mteb-dev_cosine_accuracy | +|:------:|:-----:|:-------------:|:---------------:|:-----------------------------:| +| 0 | 0 | - | - | 0.8394 | +| 0.1308 | 500 | 2.671 | 1.1796 | 0.8794 | +| 0.2616 | 1000 | 1.9941 | 1.1051 | 0.8880 | +| 0.3925 | 1500 | 2.0147 | 1.0550 | 0.8926 | +| 0.5233 | 2000 | 1.7696 | 1.0167 | 0.8948 | +| 0.6541 | 2500 | 1.892 | 0.9942 | 0.8973 | +| 0.7849 | 3000 | 1.7924 | 1.0039 | 0.9000 | +| 0.9158 | 3500 | 1.8434 | 1.0105 | 0.8958 | +| 1.0466 | 4000 | 1.7597 | 0.9599 | 0.9011 | +| 1.1774 | 4500 | 1.8684 | 1.0748 | 0.9027 | +| 1.3082 | 5000 | 1.692 | 0.9666 | 0.9032 | +| 1.4390 | 5500 | 1.7115 | 1.0497 | 0.9031 | +| 1.5699 | 6000 | 1.6607 | 1.0262 | 0.9040 | +| 1.7007 | 6500 | 1.6804 | 0.9984 | 0.9052 | +| 1.8315 | 7000 | 1.6108 | 0.9315 | 0.9048 | +| 1.9623 | 7500 | 1.5806 | 1.0537 | 0.9062 | +| 2.0931 | 8000 | 1.6489 | 1.0271 | 0.9075 | +| 2.2240 | 8500 | 1.5841 | 1.1238 | 0.9078 | +| 2.3548 | 9000 | 1.6315 | 1.0886 | 0.9069 | +| 2.4856 | 9500 | 1.4484 | 1.0287 | 0.9079 | +| 2.6164 | 10000 | 1.5661 | 1.1722 | 0.9095 | +| 2.7473 | 10500 | 1.4791 | 1.0988 | 0.9090 | +| 2.8781 | 11000 | 1.5247 | 1.0828 | 0.9100 | +| 3.0089 | 11500 | 1.4124 | 1.0981 | 0.9096 | +| 3.1397 | 12000 | 1.569 | 1.0372 | 0.9111 | +| 3.2705 | 12500 | 1.4468 | 0.9301 | 0.9106 | +| 3.4014 | 13000 | 1.5556 | 1.0313 | 0.9118 | +| 3.5322 | 13500 | 1.346 | 1.0433 | 0.9078 | +| 3.6630 | 14000 | 1.4514 | 0.9846 | 0.9101 | +| 3.7938 | 14500 | 1.3815 | 1.1034 | 0.9131 | +| 3.9246 | 15000 | 1.4323 | 1.0120 | 0.9103 | +| 4.0555 | 15500 | 1.3485 | 0.9873 | 0.9117 | +| 4.1863 | 16000 | 1.4595 | 1.0307 | 0.9103 | +| 4.3171 | 16500 | 1.3718 | 1.1036 | 0.9134 | +| 4.4479 | 17000 | 1.3685 | 1.0405 | 0.9102 | +| 4.5788 | 17500 | 1.3662 | 1.0109 | 0.9112 | +| 4.7096 | 18000 | 1.3363 | 1.0407 | 0.9130 | +| 4.8404 | 18500 | 1.3321 | 1.0848 | 0.9123 | +| 4.9712 | 19000 | 1.3313 | 1.0468 | 0.9130 | +| 5.1020 | 19500 | 1.3656 | 0.9708 | 0.9121 | +| 5.2329 | 20000 | 1.3311 | 1.0208 | 0.9148 | +| 5.3637 | 20500 | 1.403 | 1.0025 | 0.9115 | +| 5.4945 | 21000 | 1.2109 | 1.0739 | 0.9131 | +| 5.6253 | 21500 | 1.3038 | 1.1280 | 0.9120 | +| 5.7561 | 22000 | 1.2577 | 1.0245 | 0.9131 | +| 5.8870 | 22500 | 1.3112 | 0.9378 | 0.9149 | +| 6.0178 | 23000 | 1.2141 | 1.0292 | 0.9126 | +| 6.1486 | 23500 | 1.3696 | 1.1213 | 0.9141 | +| 6.2794 | 24000 | 1.2436 | 0.9875 | 0.9141 | +| 6.4103 | 24500 | 1.3514 | 1.0064 | 0.9146 | +| 6.5411 | 25000 | 1.1827 | 1.0174 | 0.9117 | +| 6.6719 | 25500 | 1.2619 | 1.0304 | 0.9120 | +| 6.8027 | 26000 | 1.1997 | 1.0499 | 0.9149 | +| 6.9335 | 26500 | 1.2609 | 1.0160 | 0.9141 | +| 7.0644 | 27000 | 1.2065 | 1.0216 | 0.9140 | +| 7.1952 | 27500 | 1.2802 | 1.0620 | 0.9135 | +| 7.3260 | 28000 | 1.2501 | 1.0798 | 0.9155 | +| 7.4568 | 28500 | 1.201 | 1.0196 | 0.9142 | +| 7.5877 | 29000 | 1.2249 | 1.0325 | 0.9143 | +| 7.7185 | 29500 | 1.1867 | 1.0195 | 0.9130 | +| 7.8493 | 30000 | 1.1917 | 1.0016 | 0.9137 | +| 7.9801 | 30500 | 1.194 | 1.0858 | 0.9156 | +| 8.1109 | 31000 | 1.2351 | 0.9960 | 0.9144 | +| 8.2418 | 31500 | 1.1834 | 1.0464 | 0.9161 | +| 8.3726 | 32000 | 1.3046 | 1.0395 | 0.9145 | +| 8.5034 | 32500 | 1.106 | 1.0235 | 0.9140 | +| 8.6342 | 33000 | 1.1845 | 1.0615 | 0.9134 | +| 8.7650 | 33500 | 1.1372 | 1.0205 | 0.9146 | +| 8.8959 | 34000 | 1.2218 | 0.9796 | 0.9148 | +| 9.0267 | 34500 | 1.0983 | 1.0065 | 0.9147 | +| 9.1575 | 35000 | 1.2656 | 1.0339 | 0.9154 | +| 9.2883 | 35500 | 1.1522 | 1.0168 | 0.9154 | +| 9.4192 | 36000 | 1.2407 | 1.0145 | 0.9150 | +| 9.5500 | 36500 | 1.1091 | 1.0321 | 0.9150 | +| 9.6808 | 37000 | 1.1689 | 1.0270 | 0.9145 | +| 9.8116 | 37500 | 1.1116 | 1.0237 | 0.9148 | +| 9.9424 | 38000 | 1.1824 | 1.0135 | 0.9145 | + + +### Framework Versions +- Python: 3.10.10 +- Sentence Transformers: 3.4.0.dev0 +- Transformers: 4.46.3 +- PyTorch: 2.5.1+cu124 +- Accelerate: 0.34.2 +- Datasets: 2.21.0 +- Tokenizers: 0.20.3 + +## Citation + +### BibTeX + +#### Sentence Transformers +```bibtex +@inproceedings{reimers-2019-sentence-bert, + title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", + author = "Reimers, Nils and Gurevych, Iryna", + booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", + month = "11", + year = "2019", + publisher = "Association for Computational Linguistics", + url = "https://arxiv.org/abs/1908.10084", +} +``` + +#### MultipleNegativesRankingLoss +```bibtex +@misc{henderson2017efficient, + title={Efficient Natural Language Response Suggestion for Smart Reply}, + author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil}, + year={2017}, + eprint={1705.00652}, + archivePrefix={arXiv}, + primaryClass={cs.CL} +} +``` + + + + + + \ No newline at end of file