jfkback nielsr HF staff commited on
Commit
68d42e8
·
verified ·
1 Parent(s): 4e1b038

Add license and pipeline tag (#3)

Browse files

- Add license and pipeline tag (00134611fde156839d0275ede355a388832ec45c)


Co-authored-by: Niels Rogge <[email protected]>

Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -1,21 +1,21 @@
1
  ---
2
- library_name: transformers
 
3
  datasets:
4
  - microsoft/ms_marco
5
  language:
6
  - en
7
- base_model:
8
- - google-bert/bert-base-uncased
 
9
  ---
10
 
11
  # Model Card
12
  This is the official model from the paper [Hypencoder: Hypernetworks for Information Retrieval](https://arxiv.org/abs/2502.05364).
13
 
14
-
15
  ## Model Details
16
  This is a Hypencoder Dual Enocder. It contains two trunks the text encoder and Hypencoder. The text encoder converts items into 768 dimension vectors while the Hypencoder converts text into a small neural network which takes the 768 dimension vector from the text encoder as input. This small network is then used to output a relevance score. To use this model please take a look at the [Github](https://github.com/jfkback/hypencoder-paper) page which contains the required code and details on how to run the model.
17
 
18
-
19
  ### Model Variants
20
  We released the four models used in the paper. Each model is identical except the small neural networks, which we refer to as q-nets, have different numbers of hidden layers.
21
 
 
1
  ---
2
+ base_model:
3
+ - google-bert/bert-base-uncased
4
  datasets:
5
  - microsoft/ms_marco
6
  language:
7
  - en
8
+ library_name: transformers
9
+ pipeline_tag: feature-extraction
10
+ license: apache-2.0
11
  ---
12
 
13
  # Model Card
14
  This is the official model from the paper [Hypencoder: Hypernetworks for Information Retrieval](https://arxiv.org/abs/2502.05364).
15
 
 
16
  ## Model Details
17
  This is a Hypencoder Dual Enocder. It contains two trunks the text encoder and Hypencoder. The text encoder converts items into 768 dimension vectors while the Hypencoder converts text into a small neural network which takes the 768 dimension vector from the text encoder as input. This small network is then used to output a relevance score. To use this model please take a look at the [Github](https://github.com/jfkback/hypencoder-paper) page which contains the required code and details on how to run the model.
18
 
 
19
  ### Model Variants
20
  We released the four models used in the paper. Each model is identical except the small neural networks, which we refer to as q-nets, have different numbers of hidden layers.
21