parquet-converter
commited on
Commit
•
733ae70
1
Parent(s):
bdaf2bf
Update parquet files
Browse files- .gitattributes +1 -0
- .gitignore +0 -2
- README.md +0 -79
- data/test/test.jsonl.zst → default/test/0000.parquet +2 -2
.gitattributes
CHANGED
@@ -1,2 +1,3 @@
|
|
1 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
2 |
*.gz filter=lfs diff=lfs merge=lfs -text
|
|
|
|
1 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
2 |
*.gz filter=lfs diff=lfs merge=lfs -text
|
3 |
+
default/test/0000.parquet filter=lfs diff=lfs merge=lfs -text
|
.gitignore
DELETED
@@ -1,2 +0,0 @@
|
|
1 |
-
.venv
|
2 |
-
.mypy_cache
|
|
|
|
|
|
README.md
DELETED
@@ -1,79 +0,0 @@
|
|
1 |
-
---
|
2 |
-
language:
|
3 |
-
- en
|
4 |
-
license: apache-2.0
|
5 |
-
tags:
|
6 |
-
- text
|
7 |
-
pretty_name: MTEB/BEIR eval hard negatives
|
8 |
-
size_categories:
|
9 |
-
- "100K<n<1M"
|
10 |
-
source_datasets:
|
11 |
-
- "BeIR"
|
12 |
-
task_categories:
|
13 |
-
- sentence-similarity
|
14 |
-
dataset_info:
|
15 |
-
config_name: default
|
16 |
-
features:
|
17 |
-
- name: query
|
18 |
-
dtype: string
|
19 |
-
- name: positive
|
20 |
-
sequence: string
|
21 |
-
- name: negative
|
22 |
-
sequence: string
|
23 |
-
splits:
|
24 |
-
- name: test
|
25 |
-
num_bytes: 89609915
|
26 |
-
num_examples: 502939
|
27 |
-
train-eval-index:
|
28 |
-
- config: default
|
29 |
-
task: sentence-similarity
|
30 |
-
splits:
|
31 |
-
eval_split: test
|
32 |
-
configs:
|
33 |
-
- config_name: default
|
34 |
-
data_files:
|
35 |
-
- split: test
|
36 |
-
path: "data/test/*"
|
37 |
-
---
|
38 |
-
|
39 |
-
# BEIR/MTEB hard negatives dataset
|
40 |
-
|
41 |
-
A dataset for quick evaluation of embedding models during their training.
|
42 |
-
|
43 |
-
The problem: running a full MTEB evaluation on a single GPU may take 10-20 hours. Most of this time is spent on embedding all 30M docs in all 10+ corpora. This dataset solves this problem by unwrapping a "retrieval" style benchmark into the "reranking" style:
|
44 |
-
|
45 |
-
* We compute embeddings for all documents in the corpora with the [intfloat/e5-base-v2](todo) model.
|
46 |
-
* For each corpus in BEIR/MTEB benchmark we build a Lucene index with text documents and their embeddings.
|
47 |
-
* For each eval query we do a hybrid [RRF](todo)-based retrieval for top-32 negatives
|
48 |
-
|
49 |
-
As BEIR testset is size-unbalanced (TREC-COVID is 42 queries, and MS MARCO is ~4000) we sample top-300 random queries from each dataset.
|
50 |
-
|
51 |
-
It takes around 30-60 seconds to perform eval using Nixietune on a single RTX 4090.
|
52 |
-
|
53 |
-
A dataset in a [nixietune](https://github.com/nixiesearch/nixietune) compatible format:
|
54 |
-
|
55 |
-
```json
|
56 |
-
{
|
57 |
-
"query": ")what was the immediate impact of the success of the manhattan project?",
|
58 |
-
"pos": [
|
59 |
-
"The presence of communication amid scientific minds was equally important to the success of the Manhattan Project as scientific intellect was. The only cloud hanging over the impressive achievement of the atomic researchers and engineers is what their success truly meant; hundreds of thousands of innocent lives obliterated."
|
60 |
-
],
|
61 |
-
"neg": [
|
62 |
-
"Abstract. The pivotal engineering and scientific success of the Twentieth century was the Manhattan Project. The Manhattan Project assimilated concepts and leaders from all scientific fields and engineering disciplines to construct the first two atomic bombs.",
|
63 |
-
"The pivotal engineering and scientific success of the Twentieth century was the Manhattan Project. The Manhattan Project assimilated concepts and leaders from all scientific fields and engineering disciplines to construct the first two atomic bombs."
|
64 |
-
]
|
65 |
-
}
|
66 |
-
```
|
67 |
-
|
68 |
-
## Usage
|
69 |
-
|
70 |
-
```python
|
71 |
-
from datasets import load_dataset
|
72 |
-
|
73 |
-
data = load_dataset('nixiesearch/beir-eval-hardneg')
|
74 |
-
print(data["test"].features)
|
75 |
-
```
|
76 |
-
|
77 |
-
## License
|
78 |
-
|
79 |
-
Apache 2.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
data/test/test.jsonl.zst → default/test/0000.parquet
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c9529a46578e26a1fcdef1b53c9e0ab7ef46e09e42e5a46db015a13db5c37b5
|
3 |
+
size 122190557
|