Add license and pipeline tag (#3)
Browse files- Add license and pipeline tag (00134611fde156839d0275ede355a388832ec45c)
Co-authored-by: Niels Rogge <[email protected]>
README.md
CHANGED
@@ -1,21 +1,21 @@
|
|
1 |
---
|
2 |
-
|
|
|
3 |
datasets:
|
4 |
- microsoft/ms_marco
|
5 |
language:
|
6 |
- en
|
7 |
-
|
8 |
-
|
|
|
9 |
---
|
10 |
|
11 |
# Model Card
|
12 |
This is the official model from the paper [Hypencoder: Hypernetworks for Information Retrieval](https://arxiv.org/abs/2502.05364).
|
13 |
|
14 |
-
|
15 |
## Model Details
|
16 |
This is a Hypencoder Dual Enocder. It contains two trunks the text encoder and Hypencoder. The text encoder converts items into 768 dimension vectors while the Hypencoder converts text into a small neural network which takes the 768 dimension vector from the text encoder as input. This small network is then used to output a relevance score. To use this model please take a look at the [Github](https://github.com/jfkback/hypencoder-paper) page which contains the required code and details on how to run the model.
|
17 |
|
18 |
-
|
19 |
### Model Variants
|
20 |
We released the four models used in the paper. Each model is identical except the small neural networks, which we refer to as q-nets, have different numbers of hidden layers.
|
21 |
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- google-bert/bert-base-uncased
|
4 |
datasets:
|
5 |
- microsoft/ms_marco
|
6 |
language:
|
7 |
- en
|
8 |
+
library_name: transformers
|
9 |
+
pipeline_tag: feature-extraction
|
10 |
+
license: apache-2.0
|
11 |
---
|
12 |
|
13 |
# Model Card
|
14 |
This is the official model from the paper [Hypencoder: Hypernetworks for Information Retrieval](https://arxiv.org/abs/2502.05364).
|
15 |
|
|
|
16 |
## Model Details
|
17 |
This is a Hypencoder Dual Enocder. It contains two trunks the text encoder and Hypencoder. The text encoder converts items into 768 dimension vectors while the Hypencoder converts text into a small neural network which takes the 768 dimension vector from the text encoder as input. This small network is then used to output a relevance score. To use this model please take a look at the [Github](https://github.com/jfkback/hypencoder-paper) page which contains the required code and details on how to run the model.
|
18 |
|
|
|
19 |
### Model Variants
|
20 |
We released the four models used in the paper. Each model is identical except the small neural networks, which we refer to as q-nets, have different numbers of hidden layers.
|
21 |
|