Upload 14 files
Browse files- results/Amanda/bge_portuguese_v6/FloresBitextMining.json +193 -0
- results/Amanda/bge_portuguese_v6/MassiveIntentClassification.json +25 -0
- results/Amanda/bge_portuguese_v6/MassiveScenarioClassification.json +25 -0
- results/Amanda/bge_portuguese_v6/MintakaRetrieval.json +45 -0
- results/Amanda/bge_portuguese_v6/MultiHateClassification.json +17 -0
- results/Amanda/bge_portuguese_v6/MultiLongDocRetrieval.json +85 -0
- results/Amanda/bge_portuguese_v6/NTREXBitextMining.json +92 -0
- results/Amanda/bge_portuguese_v6/PTT_AskUbuntuDupQuestions.json +10 -0
- results/Amanda/bge_portuguese_v6/PTT_CQADupstackStatsRetrieval.json +43 -0
- results/Amanda/bge_portuguese_v6/PTT_MedrxivClusteringS2S.json +23 -0
- results/Amanda/bge_portuguese_v6/PTT_SprintDuplicateQuestions.json +93 -0
- results/Amanda/bge_portuguese_v6/PTT_StackOverflowDupQuestions.json +10 -0
- results/Amanda/bge_portuguese_v6/PTT_TwentyNewsgroupsClustering.json +23 -0
- results/Amanda/bge_portuguese_v6/PTT_TwitterSemEval2015.json +49 -0
results/Amanda/bge_portuguese_v6/FloresBitextMining.json
ADDED
@@ -0,0 +1,193 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "80dc3040d19756742c9a18267ab30f54fb8e226b",
|
3 |
+
"dev": {
|
4 |
+
"arb_Arab-por_Latn": {
|
5 |
+
"accuracy": 0.9979939819458375,
|
6 |
+
"f1": 0.9973253092611166,
|
7 |
+
"main_score": 0.9973253092611166,
|
8 |
+
"precision": 0.9969909729187563,
|
9 |
+
"recall": 0.9979939819458375
|
10 |
+
},
|
11 |
+
"deu_Latn-por_Latn": {
|
12 |
+
"accuracy": 0.9969909729187563,
|
13 |
+
"f1": 0.995987963891675,
|
14 |
+
"main_score": 0.995987963891675,
|
15 |
+
"precision": 0.9954864593781344,
|
16 |
+
"recall": 0.9969909729187563
|
17 |
+
},
|
18 |
+
"eng_Latn-por_Latn": {
|
19 |
+
"accuracy": 0.9989969909729187,
|
20 |
+
"f1": 0.9986626546305583,
|
21 |
+
"main_score": 0.9986626546305583,
|
22 |
+
"precision": 0.9984954864593781,
|
23 |
+
"recall": 0.9989969909729187
|
24 |
+
},
|
25 |
+
"evaluation_time": 180.36,
|
26 |
+
"fra_Latn-por_Latn": {
|
27 |
+
"accuracy": 0.9979939819458375,
|
28 |
+
"f1": 0.9973253092611166,
|
29 |
+
"main_score": 0.9973253092611166,
|
30 |
+
"precision": 0.9969909729187563,
|
31 |
+
"recall": 0.9979939819458375
|
32 |
+
},
|
33 |
+
"ita_Latn-por_Latn": {
|
34 |
+
"accuracy": 0.9979939819458375,
|
35 |
+
"f1": 0.9973253092611166,
|
36 |
+
"main_score": 0.9973253092611166,
|
37 |
+
"precision": 0.9969909729187563,
|
38 |
+
"recall": 0.9979939819458375
|
39 |
+
},
|
40 |
+
"jpn_Jpan-por_Latn": {
|
41 |
+
"accuracy": 0.9979939819458375,
|
42 |
+
"f1": 0.9973253092611166,
|
43 |
+
"main_score": 0.9973253092611166,
|
44 |
+
"precision": 0.9969909729187563,
|
45 |
+
"recall": 0.9979939819458375
|
46 |
+
},
|
47 |
+
"kor_Hang-por_Latn": {
|
48 |
+
"accuracy": 0.9929789368104313,
|
49 |
+
"f1": 0.9906385824139085,
|
50 |
+
"main_score": 0.9906385824139085,
|
51 |
+
"precision": 0.9894684052156469,
|
52 |
+
"recall": 0.9929789368104313
|
53 |
+
},
|
54 |
+
"pol_Latn-por_Latn": {
|
55 |
+
"accuracy": 0.995987963891675,
|
56 |
+
"f1": 0.9946506185222335,
|
57 |
+
"main_score": 0.9946506185222335,
|
58 |
+
"precision": 0.9939819458375125,
|
59 |
+
"recall": 0.995987963891675
|
60 |
+
},
|
61 |
+
"rus_Cyrl-por_Latn": {
|
62 |
+
"accuracy": 0.995987963891675,
|
63 |
+
"f1": 0.9946506185222334,
|
64 |
+
"main_score": 0.9946506185222334,
|
65 |
+
"precision": 0.9939819458375125,
|
66 |
+
"recall": 0.995987963891675
|
67 |
+
},
|
68 |
+
"spa_Latn-por_Latn": {
|
69 |
+
"accuracy": 0.9989969909729187,
|
70 |
+
"f1": 0.9986626546305583,
|
71 |
+
"main_score": 0.9986626546305583,
|
72 |
+
"precision": 0.9984954864593781,
|
73 |
+
"recall": 0.9989969909729187
|
74 |
+
},
|
75 |
+
"swe_Latn-por_Latn": {
|
76 |
+
"accuracy": 1.0,
|
77 |
+
"f1": 1.0,
|
78 |
+
"main_score": 1.0,
|
79 |
+
"precision": 1.0,
|
80 |
+
"recall": 1.0
|
81 |
+
},
|
82 |
+
"zho_Hans-por_Latn": {
|
83 |
+
"accuracy": 0.9969909729187563,
|
84 |
+
"f1": 0.995987963891675,
|
85 |
+
"main_score": 0.995987963891675,
|
86 |
+
"precision": 0.9954864593781344,
|
87 |
+
"recall": 0.9969909729187563
|
88 |
+
},
|
89 |
+
"zho_Hant-por_Latn": {
|
90 |
+
"accuracy": 0.9939819458375125,
|
91 |
+
"f1": 0.9919759277833501,
|
92 |
+
"main_score": 0.9919759277833501,
|
93 |
+
"precision": 0.9909729187562688,
|
94 |
+
"recall": 0.9939819458375125
|
95 |
+
}
|
96 |
+
},
|
97 |
+
"devtest": {
|
98 |
+
"arb_Arab-por_Latn": {
|
99 |
+
"accuracy": 0.9970355731225297,
|
100 |
+
"f1": 0.9960474308300395,
|
101 |
+
"main_score": 0.9960474308300395,
|
102 |
+
"precision": 0.9955533596837944,
|
103 |
+
"recall": 0.9970355731225297
|
104 |
+
},
|
105 |
+
"deu_Latn-por_Latn": {
|
106 |
+
"accuracy": 0.9970355731225297,
|
107 |
+
"f1": 0.9960474308300395,
|
108 |
+
"main_score": 0.9960474308300395,
|
109 |
+
"precision": 0.9955533596837944,
|
110 |
+
"recall": 0.9970355731225297
|
111 |
+
},
|
112 |
+
"eng_Latn-por_Latn": {
|
113 |
+
"accuracy": 0.9990118577075099,
|
114 |
+
"f1": 0.9986824769433464,
|
115 |
+
"main_score": 0.9986824769433464,
|
116 |
+
"precision": 0.9985177865612648,
|
117 |
+
"recall": 0.9990118577075099
|
118 |
+
},
|
119 |
+
"evaluation_time": 196.89,
|
120 |
+
"fra_Latn-por_Latn": {
|
121 |
+
"accuracy": 0.9980237154150198,
|
122 |
+
"f1": 0.997364953886693,
|
123 |
+
"main_score": 0.997364953886693,
|
124 |
+
"precision": 0.9970355731225297,
|
125 |
+
"recall": 0.9980237154150198
|
126 |
+
},
|
127 |
+
"ita_Latn-por_Latn": {
|
128 |
+
"accuracy": 0.9980237154150198,
|
129 |
+
"f1": 0.997364953886693,
|
130 |
+
"main_score": 0.997364953886693,
|
131 |
+
"precision": 0.9970355731225297,
|
132 |
+
"recall": 0.9980237154150198
|
133 |
+
},
|
134 |
+
"jpn_Jpan-por_Latn": {
|
135 |
+
"accuracy": 0.9960474308300395,
|
136 |
+
"f1": 0.994729907773386,
|
137 |
+
"main_score": 0.994729907773386,
|
138 |
+
"precision": 0.9940711462450593,
|
139 |
+
"recall": 0.9960474308300395
|
140 |
+
},
|
141 |
+
"kor_Hang-por_Latn": {
|
142 |
+
"accuracy": 0.9940711462450593,
|
143 |
+
"f1": 0.9920948616600791,
|
144 |
+
"main_score": 0.9920948616600791,
|
145 |
+
"precision": 0.991106719367589,
|
146 |
+
"recall": 0.9940711462450593
|
147 |
+
},
|
148 |
+
"pol_Latn-por_Latn": {
|
149 |
+
"accuracy": 0.9960474308300395,
|
150 |
+
"f1": 0.994729907773386,
|
151 |
+
"main_score": 0.994729907773386,
|
152 |
+
"precision": 0.9940711462450593,
|
153 |
+
"recall": 0.9960474308300395
|
154 |
+
},
|
155 |
+
"rus_Cyrl-por_Latn": {
|
156 |
+
"accuracy": 0.9980237154150198,
|
157 |
+
"f1": 0.997364953886693,
|
158 |
+
"main_score": 0.997364953886693,
|
159 |
+
"precision": 0.9970355731225297,
|
160 |
+
"recall": 0.9980237154150198
|
161 |
+
},
|
162 |
+
"spa_Latn-por_Latn": {
|
163 |
+
"accuracy": 0.9980237154150198,
|
164 |
+
"f1": 0.997364953886693,
|
165 |
+
"main_score": 0.997364953886693,
|
166 |
+
"precision": 0.9970355731225297,
|
167 |
+
"recall": 0.9980237154150198
|
168 |
+
},
|
169 |
+
"swe_Latn-por_Latn": {
|
170 |
+
"accuracy": 1.0,
|
171 |
+
"f1": 1.0,
|
172 |
+
"main_score": 1.0,
|
173 |
+
"precision": 1.0,
|
174 |
+
"recall": 1.0
|
175 |
+
},
|
176 |
+
"zho_Hans-por_Latn": {
|
177 |
+
"accuracy": 0.9960474308300395,
|
178 |
+
"f1": 0.994729907773386,
|
179 |
+
"main_score": 0.994729907773386,
|
180 |
+
"precision": 0.9940711462450593,
|
181 |
+
"recall": 0.9960474308300395
|
182 |
+
},
|
183 |
+
"zho_Hant-por_Latn": {
|
184 |
+
"accuracy": 0.9960474308300395,
|
185 |
+
"f1": 0.994729907773386,
|
186 |
+
"main_score": 0.994729907773386,
|
187 |
+
"precision": 0.9940711462450593,
|
188 |
+
"recall": 0.9960474308300395
|
189 |
+
}
|
190 |
+
},
|
191 |
+
"mteb_dataset_name": "FloresBitextMining",
|
192 |
+
"mteb_version": 2
|
193 |
+
}
|
results/Amanda/bge_portuguese_v6/MassiveIntentClassification.json
ADDED
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7",
|
3 |
+
"mteb_dataset_name": "MassiveIntentClassification",
|
4 |
+
"mteb_version": 2,
|
5 |
+
"test": {
|
6 |
+
"evaluation_time": 96.89,
|
7 |
+
"pt": {
|
8 |
+
"accuracy": 0.6429388029589779,
|
9 |
+
"accuracy_stderr": 0.01574159007730203,
|
10 |
+
"f1": 0.624149449605407,
|
11 |
+
"f1_stderr": 0.011597207566151069,
|
12 |
+
"main_score": 0.6429388029589779
|
13 |
+
}
|
14 |
+
},
|
15 |
+
"validation": {
|
16 |
+
"evaluation_time": 105.97,
|
17 |
+
"pt": {
|
18 |
+
"accuracy": 0.6440727988194788,
|
19 |
+
"accuracy_stderr": 0.018382512054495048,
|
20 |
+
"f1": 0.6093616344832975,
|
21 |
+
"f1_stderr": 0.01076436730695683,
|
22 |
+
"main_score": 0.6440727988194788
|
23 |
+
}
|
24 |
+
}
|
25 |
+
}
|
results/Amanda/bge_portuguese_v6/MassiveScenarioClassification.json
ADDED
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634",
|
3 |
+
"mteb_dataset_name": "MassiveScenarioClassification",
|
4 |
+
"mteb_version": 2,
|
5 |
+
"test": {
|
6 |
+
"evaluation_time": 62.45,
|
7 |
+
"pt": {
|
8 |
+
"accuracy": 0.6876933422999327,
|
9 |
+
"accuracy_stderr": 0.014407566848919572,
|
10 |
+
"f1": 0.6829025521195558,
|
11 |
+
"f1_stderr": 0.013884840252469367,
|
12 |
+
"main_score": 0.6876933422999327
|
13 |
+
}
|
14 |
+
},
|
15 |
+
"validation": {
|
16 |
+
"evaluation_time": 47.48,
|
17 |
+
"pt": {
|
18 |
+
"accuracy": 0.683915395966552,
|
19 |
+
"accuracy_stderr": 0.017521288958123687,
|
20 |
+
"f1": 0.6775990739298102,
|
21 |
+
"f1_stderr": 0.017684324125881695,
|
22 |
+
"main_score": 0.683915395966552
|
23 |
+
}
|
24 |
+
}
|
25 |
+
}
|
results/Amanda/bge_portuguese_v6/MintakaRetrieval.json
ADDED
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "efa78cc2f74bbcd21eff2261f9e13aebe40b814e",
|
3 |
+
"mteb_dataset_name": "MintakaRetrieval",
|
4 |
+
"mteb_version": 2,
|
5 |
+
"test": {
|
6 |
+
"evaluation_time": 53.13,
|
7 |
+
"pt": {
|
8 |
+
"map_at_1": 0.15081,
|
9 |
+
"map_at_10": 0.21799,
|
10 |
+
"map_at_100": 0.2268,
|
11 |
+
"map_at_1000": 0.22794,
|
12 |
+
"map_at_20": 0.22234,
|
13 |
+
"map_at_3": 0.19867,
|
14 |
+
"map_at_5": 0.20946,
|
15 |
+
"mrr_at_1": 0.15081,
|
16 |
+
"mrr_at_10": 0.21803,
|
17 |
+
"mrr_at_100": 0.22683,
|
18 |
+
"mrr_at_1000": 0.22797,
|
19 |
+
"mrr_at_20": 0.22237,
|
20 |
+
"mrr_at_3": 0.19867,
|
21 |
+
"mrr_at_5": 0.20946,
|
22 |
+
"ndcg_at_1": 0.15081,
|
23 |
+
"ndcg_at_10": 0.25454,
|
24 |
+
"ndcg_at_100": 0.30478,
|
25 |
+
"ndcg_at_1000": 0.34213,
|
26 |
+
"ndcg_at_20": 0.27041,
|
27 |
+
"ndcg_at_3": 0.214,
|
28 |
+
"ndcg_at_5": 0.23352,
|
29 |
+
"precision_at_1": 0.15081,
|
30 |
+
"precision_at_10": 0.03717,
|
31 |
+
"precision_at_100": 0.00624,
|
32 |
+
"precision_at_1000": 0.00093,
|
33 |
+
"precision_at_20": 0.02173,
|
34 |
+
"precision_at_3": 0.08609,
|
35 |
+
"precision_at_5": 0.06117,
|
36 |
+
"recall_at_1": 0.15081,
|
37 |
+
"recall_at_10": 0.3715,
|
38 |
+
"recall_at_100": 0.62383,
|
39 |
+
"recall_at_1000": 0.93373,
|
40 |
+
"recall_at_20": 0.43437,
|
41 |
+
"recall_at_3": 0.25828,
|
42 |
+
"recall_at_5": 0.30586
|
43 |
+
}
|
44 |
+
}
|
45 |
+
}
|
results/Amanda/bge_portuguese_v6/MultiHateClassification.json
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "ef137ea2b7c719183f8f60edf536b50f56d1365b",
|
3 |
+
"mteb_dataset_name": "MultiHateClassification",
|
4 |
+
"mteb_version": 2,
|
5 |
+
"test": {
|
6 |
+
"evaluation_time": 21.52,
|
7 |
+
"por": {
|
8 |
+
"accuracy": 0.6023,
|
9 |
+
"accuracy_stderr": 0.040573513527916216,
|
10 |
+
"ap": 0.3465812754549832,
|
11 |
+
"ap_stderr": 0.024717781473017272,
|
12 |
+
"f1": 0.5641933623962817,
|
13 |
+
"f1_stderr": 0.037581938438066446,
|
14 |
+
"main_score": 0.6023
|
15 |
+
}
|
16 |
+
}
|
17 |
+
}
|
results/Amanda/bge_portuguese_v6/MultiLongDocRetrieval.json
ADDED
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "d67138e705d963e346253a80e59676ddb418810a",
|
3 |
+
"dev": {
|
4 |
+
"evaluation_time": 9073.93,
|
5 |
+
"pt": {
|
6 |
+
"map_at_1": 0.55,
|
7 |
+
"map_at_10": 0.5978,
|
8 |
+
"map_at_100": 0.60393,
|
9 |
+
"map_at_1000": 0.60447,
|
10 |
+
"map_at_20": 0.60064,
|
11 |
+
"map_at_3": 0.585,
|
12 |
+
"map_at_5": 0.59275,
|
13 |
+
"mrr_at_1": 0.55,
|
14 |
+
"mrr_at_10": 0.5978,
|
15 |
+
"mrr_at_100": 0.60393,
|
16 |
+
"mrr_at_1000": 0.60447,
|
17 |
+
"mrr_at_20": 0.60064,
|
18 |
+
"mrr_at_3": 0.585,
|
19 |
+
"mrr_at_5": 0.59275,
|
20 |
+
"ndcg_at_1": 0.55,
|
21 |
+
"ndcg_at_10": 0.62244,
|
22 |
+
"ndcg_at_100": 0.65668,
|
23 |
+
"ndcg_at_1000": 0.66971,
|
24 |
+
"ndcg_at_20": 0.63264,
|
25 |
+
"ndcg_at_3": 0.59655,
|
26 |
+
"ndcg_at_5": 0.61074,
|
27 |
+
"precision_at_1": 0.55,
|
28 |
+
"precision_at_10": 0.07,
|
29 |
+
"precision_at_100": 0.0087,
|
30 |
+
"precision_at_1000": 0.00097,
|
31 |
+
"precision_at_20": 0.037,
|
32 |
+
"precision_at_3": 0.21,
|
33 |
+
"precision_at_5": 0.133,
|
34 |
+
"recall_at_1": 0.55,
|
35 |
+
"recall_at_10": 0.7,
|
36 |
+
"recall_at_100": 0.87,
|
37 |
+
"recall_at_1000": 0.97,
|
38 |
+
"recall_at_20": 0.74,
|
39 |
+
"recall_at_3": 0.63,
|
40 |
+
"recall_at_5": 0.665
|
41 |
+
}
|
42 |
+
},
|
43 |
+
"mteb_dataset_name": "MultiLongDocRetrieval",
|
44 |
+
"mteb_version": 2,
|
45 |
+
"test": {
|
46 |
+
"evaluation_time": 9078.26,
|
47 |
+
"pt": {
|
48 |
+
"map_at_1": 0.525,
|
49 |
+
"map_at_10": 0.59302,
|
50 |
+
"map_at_100": 0.59871,
|
51 |
+
"map_at_1000": 0.59907,
|
52 |
+
"map_at_20": 0.59621,
|
53 |
+
"map_at_3": 0.58083,
|
54 |
+
"map_at_5": 0.58958,
|
55 |
+
"mrr_at_1": 0.525,
|
56 |
+
"mrr_at_10": 0.59302,
|
57 |
+
"mrr_at_100": 0.59871,
|
58 |
+
"mrr_at_1000": 0.59907,
|
59 |
+
"mrr_at_20": 0.59621,
|
60 |
+
"mrr_at_3": 0.58083,
|
61 |
+
"mrr_at_5": 0.58958,
|
62 |
+
"ndcg_at_1": 0.525,
|
63 |
+
"ndcg_at_10": 0.62175,
|
64 |
+
"ndcg_at_100": 0.65232,
|
65 |
+
"ndcg_at_1000": 0.66342,
|
66 |
+
"ndcg_at_20": 0.63321,
|
67 |
+
"ndcg_at_3": 0.59744,
|
68 |
+
"ndcg_at_5": 0.61357,
|
69 |
+
"precision_at_1": 0.525,
|
70 |
+
"precision_at_10": 0.071,
|
71 |
+
"precision_at_100": 0.0086,
|
72 |
+
"precision_at_1000": 0.00095,
|
73 |
+
"precision_at_20": 0.03775,
|
74 |
+
"precision_at_3": 0.215,
|
75 |
+
"precision_at_5": 0.137,
|
76 |
+
"recall_at_1": 0.525,
|
77 |
+
"recall_at_10": 0.71,
|
78 |
+
"recall_at_100": 0.86,
|
79 |
+
"recall_at_1000": 0.95,
|
80 |
+
"recall_at_20": 0.755,
|
81 |
+
"recall_at_3": 0.645,
|
82 |
+
"recall_at_5": 0.685
|
83 |
+
}
|
84 |
+
}
|
85 |
+
}
|
results/Amanda/bge_portuguese_v6/NTREXBitextMining.json
ADDED
@@ -0,0 +1,92 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "fd20d54141b6da952d5c68a2989472892489da0f",
|
3 |
+
"mteb_dataset_name": "NTREXBitextMining",
|
4 |
+
"mteb_version": 2,
|
5 |
+
"test": {
|
6 |
+
"arb_Arab-por_Latn": {
|
7 |
+
"accuracy": 0.9694541812719079,
|
8 |
+
"f1": 0.9597729928225672,
|
9 |
+
"main_score": 0.9597729928225672,
|
10 |
+
"precision": 0.9550158571190119,
|
11 |
+
"recall": 0.9694541812719079
|
12 |
+
},
|
13 |
+
"deu_Latn-por_Latn": {
|
14 |
+
"accuracy": 0.9844767150726089,
|
15 |
+
"f1": 0.9798030378901686,
|
16 |
+
"main_score": 0.9798030378901686,
|
17 |
+
"precision": 0.9775496578200633,
|
18 |
+
"recall": 0.9844767150726089
|
19 |
+
},
|
20 |
+
"eng_Latn-por_Latn": {
|
21 |
+
"accuracy": 0.9904857285928893,
|
22 |
+
"f1": 0.9873977633116341,
|
23 |
+
"main_score": 0.9873977633116341,
|
24 |
+
"precision": 0.9858955099315639,
|
25 |
+
"recall": 0.9904857285928893
|
26 |
+
},
|
27 |
+
"evaluation_time": 315.58,
|
28 |
+
"fra_Latn-por_Latn": {
|
29 |
+
"accuracy": 0.9779669504256384,
|
30 |
+
"f1": 0.971123351694208,
|
31 |
+
"main_score": 0.971123351694208,
|
32 |
+
"precision": 0.9677850108496078,
|
33 |
+
"recall": 0.9779669504256384
|
34 |
+
},
|
35 |
+
"ita_Latn-por_Latn": {
|
36 |
+
"accuracy": 0.9869804707060591,
|
37 |
+
"f1": 0.9829744616925388,
|
38 |
+
"main_score": 0.9829744616925388,
|
39 |
+
"precision": 0.9810549157068936,
|
40 |
+
"recall": 0.9869804707060591
|
41 |
+
},
|
42 |
+
"jpn_Jpan-por_Latn": {
|
43 |
+
"accuracy": 0.9679519278918377,
|
44 |
+
"f1": 0.957686529794692,
|
45 |
+
"main_score": 0.957686529794692,
|
46 |
+
"precision": 0.9525955600066767,
|
47 |
+
"recall": 0.9679519278918377
|
48 |
+
},
|
49 |
+
"kor_Hang-por_Latn": {
|
50 |
+
"accuracy": 0.9689534301452178,
|
51 |
+
"f1": 0.9593557002169921,
|
52 |
+
"main_score": 0.9593557002169921,
|
53 |
+
"precision": 0.9546820230345519,
|
54 |
+
"recall": 0.9689534301452178
|
55 |
+
},
|
56 |
+
"pol_Latn-por_Latn": {
|
57 |
+
"accuracy": 0.9799699549323986,
|
58 |
+
"f1": 0.9741612418627943,
|
59 |
+
"main_score": 0.9741612418627943,
|
60 |
+
"precision": 0.9713319979969955,
|
61 |
+
"recall": 0.9799699549323986
|
62 |
+
},
|
63 |
+
"rus_Cyrl-por_Latn": {
|
64 |
+
"accuracy": 0.9669504256384577,
|
65 |
+
"f1": 0.956685027541312,
|
66 |
+
"main_score": 0.956685027541312,
|
67 |
+
"precision": 0.9516775162744117,
|
68 |
+
"recall": 0.9669504256384577
|
69 |
+
},
|
70 |
+
"spa_Latn-por_Latn": {
|
71 |
+
"accuracy": 0.9889834752128193,
|
72 |
+
"f1": 0.985478217325989,
|
73 |
+
"main_score": 0.985478217325989,
|
74 |
+
"precision": 0.9837255883825738,
|
75 |
+
"recall": 0.9889834752128193
|
76 |
+
},
|
77 |
+
"swe_Latn-por_Latn": {
|
78 |
+
"accuracy": 0.9884827240861291,
|
79 |
+
"f1": 0.984727090635954,
|
80 |
+
"main_score": 0.984727090635954,
|
81 |
+
"precision": 0.9828910031714237,
|
82 |
+
"recall": 0.9884827240861291
|
83 |
+
},
|
84 |
+
"zho_Hant-por_Latn": {
|
85 |
+
"accuracy": 0.9684526790185278,
|
86 |
+
"f1": 0.959522617259222,
|
87 |
+
"main_score": 0.959522617259222,
|
88 |
+
"precision": 0.9552662326823568,
|
89 |
+
"recall": 0.9684526790185278
|
90 |
+
}
|
91 |
+
}
|
92 |
+
}
|
results/Amanda/bge_portuguese_v6/PTT_AskUbuntuDupQuestions.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "main",
|
3 |
+
"mteb_dataset_name": "PTT_AskUbuntuDupQuestions",
|
4 |
+
"mteb_version": 2,
|
5 |
+
"test": {
|
6 |
+
"evaluation_time": 12.01,
|
7 |
+
"map": 0.5871499931945899,
|
8 |
+
"mrr": 0.7170887745679989
|
9 |
+
}
|
10 |
+
}
|
results/Amanda/bge_portuguese_v6/PTT_CQADupstackStatsRetrieval.json
ADDED
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "main",
|
3 |
+
"mteb_dataset_name": "PTT_CQADupstackStatsRetrieval",
|
4 |
+
"mteb_version": 2,
|
5 |
+
"test": {
|
6 |
+
"evaluation_time": 1154.06,
|
7 |
+
"map_at_1": 0.1641,
|
8 |
+
"map_at_10": 0.21003,
|
9 |
+
"map_at_100": 0.2176,
|
10 |
+
"map_at_1000": 0.21862,
|
11 |
+
"map_at_20": 0.21333,
|
12 |
+
"map_at_3": 0.19249,
|
13 |
+
"map_at_5": 0.20368,
|
14 |
+
"mrr_at_1": 0.18712,
|
15 |
+
"mrr_at_10": 0.23483,
|
16 |
+
"mrr_at_100": 0.24201,
|
17 |
+
"mrr_at_1000": 0.24281,
|
18 |
+
"mrr_at_20": 0.2382,
|
19 |
+
"mrr_at_3": 0.21677,
|
20 |
+
"mrr_at_5": 0.22866,
|
21 |
+
"ndcg_at_1": 0.18712,
|
22 |
+
"ndcg_at_10": 0.24199,
|
23 |
+
"ndcg_at_100": 0.28338,
|
24 |
+
"ndcg_at_1000": 0.31107,
|
25 |
+
"ndcg_at_20": 0.25378,
|
26 |
+
"ndcg_at_3": 0.20857,
|
27 |
+
"ndcg_at_5": 0.22757,
|
28 |
+
"precision_at_1": 0.18712,
|
29 |
+
"precision_at_10": 0.03834,
|
30 |
+
"precision_at_100": 0.00658,
|
31 |
+
"precision_at_1000": 0.00097,
|
32 |
+
"precision_at_20": 0.02201,
|
33 |
+
"precision_at_3": 0.08947,
|
34 |
+
"precision_at_5": 0.06564,
|
35 |
+
"recall_at_1": 0.1641,
|
36 |
+
"recall_at_10": 0.3173,
|
37 |
+
"recall_at_100": 0.50974,
|
38 |
+
"recall_at_1000": 0.7187,
|
39 |
+
"recall_at_20": 0.36146,
|
40 |
+
"recall_at_3": 0.22689,
|
41 |
+
"recall_at_5": 0.27367
|
42 |
+
}
|
43 |
+
}
|
results/Amanda/bge_portuguese_v6/PTT_MedrxivClusteringS2S.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "main",
|
3 |
+
"mteb_dataset_name": "PTT_MedrxivClusteringS2S",
|
4 |
+
"mteb_version": 2,
|
5 |
+
"test": {
|
6 |
+
"evaluation_time": 133.69,
|
7 |
+
"main_score": 0.24574152331605165,
|
8 |
+
"v_measure": 0.24574152331605165,
|
9 |
+
"v_measure_std": 0.010843354122673225,
|
10 |
+
"v_measures": [
|
11 |
+
0.23696843618653413,
|
12 |
+
0.23483748484109662,
|
13 |
+
0.2360348323630235,
|
14 |
+
0.235950205875761,
|
15 |
+
0.2353593295033998,
|
16 |
+
0.2553759269570152,
|
17 |
+
0.2527774690663667,
|
18 |
+
0.2662060045385906,
|
19 |
+
0.2566302713792394,
|
20 |
+
0.2472752724494892
|
21 |
+
]
|
22 |
+
}
|
23 |
+
}
|
results/Amanda/bge_portuguese_v6/PTT_SprintDuplicateQuestions.json
ADDED
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "main",
|
3 |
+
"mteb_dataset_name": "PTT_SprintDuplicateQuestions",
|
4 |
+
"mteb_version": 2,
|
5 |
+
"test": {
|
6 |
+
"cos_sim": {
|
7 |
+
"accuracy": 0.9975544554455446,
|
8 |
+
"accuracy_threshold": 0.8014061450958252,
|
9 |
+
"ap": 0.9340558806075424,
|
10 |
+
"f1": 0.8720870015535991,
|
11 |
+
"f1_threshold": 0.7990845441818237,
|
12 |
+
"precision": 0.9044038668098818,
|
13 |
+
"recall": 0.842
|
14 |
+
},
|
15 |
+
"dot": {
|
16 |
+
"accuracy": 0.9975544554455446,
|
17 |
+
"accuracy_threshold": 0.8014061450958252,
|
18 |
+
"ap": 0.9340558806075424,
|
19 |
+
"f1": 0.8720870015535991,
|
20 |
+
"f1_threshold": 0.7990845441818237,
|
21 |
+
"precision": 0.9044038668098818,
|
22 |
+
"recall": 0.842
|
23 |
+
},
|
24 |
+
"euclidean": {
|
25 |
+
"accuracy": 0.9975544554455446,
|
26 |
+
"accuracy_threshold": 0.6302282810211182,
|
27 |
+
"ap": 0.9340558806075424,
|
28 |
+
"f1": 0.8720870015535991,
|
29 |
+
"f1_threshold": 0.6339012384414673,
|
30 |
+
"precision": 0.9044038668098818,
|
31 |
+
"recall": 0.842
|
32 |
+
},
|
33 |
+
"evaluation_time": 22.38,
|
34 |
+
"manhattan": {
|
35 |
+
"accuracy": 0.9975148514851485,
|
36 |
+
"accuracy_threshold": 15.930070877075195,
|
37 |
+
"ap": 0.9344473861455143,
|
38 |
+
"f1": 0.8712160082093381,
|
39 |
+
"f1_threshold": 16.173168182373047,
|
40 |
+
"precision": 0.8946259220231823,
|
41 |
+
"recall": 0.849
|
42 |
+
},
|
43 |
+
"max": {
|
44 |
+
"accuracy": 0.9975544554455446,
|
45 |
+
"ap": 0.9344473861455143,
|
46 |
+
"f1": 0.8720870015535991
|
47 |
+
}
|
48 |
+
},
|
49 |
+
"validation": {
|
50 |
+
"cos_sim": {
|
51 |
+
"accuracy": 0.9975148514851485,
|
52 |
+
"accuracy_threshold": 0.7869253158569336,
|
53 |
+
"ap": 0.93019369369862,
|
54 |
+
"f1": 0.8700155359917141,
|
55 |
+
"f1_threshold": 0.7869253158569336,
|
56 |
+
"precision": 0.9022556390977443,
|
57 |
+
"recall": 0.84
|
58 |
+
},
|
59 |
+
"dot": {
|
60 |
+
"accuracy": 0.9975148514851485,
|
61 |
+
"accuracy_threshold": 0.7869253158569336,
|
62 |
+
"ap": 0.9301936945331359,
|
63 |
+
"f1": 0.8700155359917141,
|
64 |
+
"f1_threshold": 0.7869253158569336,
|
65 |
+
"precision": 0.9022556390977443,
|
66 |
+
"recall": 0.84
|
67 |
+
},
|
68 |
+
"euclidean": {
|
69 |
+
"accuracy": 0.9975148514851485,
|
70 |
+
"accuracy_threshold": 0.6528010368347168,
|
71 |
+
"ap": 0.9301936936986199,
|
72 |
+
"f1": 0.8700155359917141,
|
73 |
+
"f1_threshold": 0.6528010368347168,
|
74 |
+
"precision": 0.9022556390977443,
|
75 |
+
"recall": 0.84
|
76 |
+
},
|
77 |
+
"evaluation_time": 23.04,
|
78 |
+
"manhattan": {
|
79 |
+
"accuracy": 0.9975247524752475,
|
80 |
+
"accuracy_threshold": 16.37934112548828,
|
81 |
+
"ap": 0.9315169831294592,
|
82 |
+
"f1": 0.8722772277227723,
|
83 |
+
"f1_threshold": 17.05251693725586,
|
84 |
+
"precision": 0.8637254901960785,
|
85 |
+
"recall": 0.881
|
86 |
+
},
|
87 |
+
"max": {
|
88 |
+
"accuracy": 0.9975247524752475,
|
89 |
+
"ap": 0.9315169831294592,
|
90 |
+
"f1": 0.8722772277227723
|
91 |
+
}
|
92 |
+
}
|
93 |
+
}
|
results/Amanda/bge_portuguese_v6/PTT_StackOverflowDupQuestions.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "main",
|
3 |
+
"mteb_dataset_name": "PTT_StackOverflowDupQuestions",
|
4 |
+
"mteb_version": 2,
|
5 |
+
"test": {
|
6 |
+
"evaluation_time": 192.59,
|
7 |
+
"map": 0.45365620296066006,
|
8 |
+
"mrr": 0.4591718126644597
|
9 |
+
}
|
10 |
+
}
|
results/Amanda/bge_portuguese_v6/PTT_TwentyNewsgroupsClustering.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "main",
|
3 |
+
"mteb_dataset_name": "PTT_TwentyNewsgroupsClustering",
|
4 |
+
"mteb_version": 2,
|
5 |
+
"test": {
|
6 |
+
"evaluation_time": 48.28,
|
7 |
+
"main_score": 0.3203252082759369,
|
8 |
+
"v_measure": 0.3203252082759369,
|
9 |
+
"v_measure_std": 0.013038394244103977,
|
10 |
+
"v_measures": [
|
11 |
+
0.34464477221611034,
|
12 |
+
0.3008189919274136,
|
13 |
+
0.30303205031385755,
|
14 |
+
0.3348185647197237,
|
15 |
+
0.32259668239864503,
|
16 |
+
0.3115956277457805,
|
17 |
+
0.33027729830180125,
|
18 |
+
0.32003375415418994,
|
19 |
+
0.3212424329192006,
|
20 |
+
0.3141919080626463
|
21 |
+
]
|
22 |
+
}
|
23 |
+
}
|
results/Amanda/bge_portuguese_v6/PTT_TwitterSemEval2015.json
ADDED
@@ -0,0 +1,49 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dataset_revision": "main",
|
3 |
+
"mteb_dataset_name": "PTT_TwitterSemEval2015",
|
4 |
+
"mteb_version": 2,
|
5 |
+
"test": {
|
6 |
+
"cos_sim": {
|
7 |
+
"accuracy": 0.8337670580768011,
|
8 |
+
"accuracy_threshold": 0.7939179539680481,
|
9 |
+
"ap": 0.6487893788087082,
|
10 |
+
"f1": 0.6073446327683616,
|
11 |
+
"f1_threshold": 0.7287165522575378,
|
12 |
+
"precision": 0.5526992287917738,
|
13 |
+
"recall": 0.6739811912225705
|
14 |
+
},
|
15 |
+
"dot": {
|
16 |
+
"accuracy": 0.8337670580768011,
|
17 |
+
"accuracy_threshold": 0.7939180135726929,
|
18 |
+
"ap": 0.6487893418943692,
|
19 |
+
"f1": 0.6073446327683616,
|
20 |
+
"f1_threshold": 0.7287165522575378,
|
21 |
+
"precision": 0.5526992287917738,
|
22 |
+
"recall": 0.6739811912225705
|
23 |
+
},
|
24 |
+
"euclidean": {
|
25 |
+
"accuracy": 0.8337670580768011,
|
26 |
+
"accuracy_threshold": 0.6420000791549683,
|
27 |
+
"ap": 0.6487893162168841,
|
28 |
+
"f1": 0.6073446327683616,
|
29 |
+
"f1_threshold": 0.7365913391113281,
|
30 |
+
"precision": 0.5526992287917738,
|
31 |
+
"recall": 0.6739811912225705
|
32 |
+
},
|
33 |
+
"evaluation_time": 23.59,
|
34 |
+
"manhattan": {
|
35 |
+
"accuracy": 0.8335131704220883,
|
36 |
+
"accuracy_threshold": 16.44284439086914,
|
37 |
+
"ap": 0.6470422250419247,
|
38 |
+
"f1": 0.6064184852374839,
|
39 |
+
"f1_threshold": 18.61992835998535,
|
40 |
+
"precision": 0.5517402476056996,
|
41 |
+
"recall": 0.6731262467939584
|
42 |
+
},
|
43 |
+
"max": {
|
44 |
+
"accuracy": 0.8337670580768011,
|
45 |
+
"ap": 0.6487893788087082,
|
46 |
+
"f1": 0.6073446327683616
|
47 |
+
}
|
48 |
+
}
|
49 |
+
}
|