Update README.md
Browse files
README.md
CHANGED
@@ -11,7 +11,7 @@ license: mit
|
|
11 |
|
12 |
# E5-RoPE-Base
|
13 |
|
14 |
-
[LongEmbed: Extending Embedding Models for Long Context Retrieval](). Dawei Zhu, Liang Wang, Nan Yang, Yifan Song, Wenhao Wu, Furu Wei, Sujian Li, arxiv 2024. Github Repo for LongEmbed: https://github.com/dwzhu-pku/LongEmbed.
|
15 |
|
16 |
This model has 12 layers and the embedding size is 768.
|
17 |
|
@@ -47,7 +47,7 @@ print(scores.tolist())
|
|
47 |
|
48 |
## Training Details
|
49 |
|
50 |
-
Please refer to our paper at [https://arxiv.org/
|
51 |
|
52 |
## Benchmark Evaluation
|
53 |
|
@@ -55,7 +55,7 @@ Check out [unilm/e5](https://github.com/microsoft/unilm/tree/master/e5) to repro
|
|
55 |
on the [BEIR](https://arxiv.org/abs/2104.08663) and [MTEB benchmark](https://arxiv.org/abs/2210.07316).
|
56 |
|
57 |
|
58 |
-
Please note that E5-RoPE-Base is not specifically trained for optimized performance. Its purpose is to enable performance comparisons between embedding models that utilize absolute position embeddings (APE) and rotary position embeddings (RoPE). By comparing E5-Base and E5-RoPE-Base, we demonstrate the superiority of RoPE-based embedding models in effectively managing longer context. See our paper [LongEmbed: Extending Embedding Models for Long Context Retrieval]() for more details.
|
59 |
|
60 |
|
61 |
## Citation
|
@@ -63,10 +63,10 @@ Please note that E5-RoPE-Base is not specifically trained for optimized performa
|
|
63 |
If you find our paper or models helpful, please consider cite as follows:
|
64 |
|
65 |
```
|
66 |
-
@article{
|
67 |
-
title={
|
68 |
-
author={
|
69 |
-
journal={arXiv preprint arXiv:
|
70 |
-
year={
|
71 |
}
|
72 |
```
|
|
|
11 |
|
12 |
# E5-RoPE-Base
|
13 |
|
14 |
+
[LongEmbed: Extending Embedding Models for Long Context Retrieval](https://arxiv.org/abs/2404.12096). Dawei Zhu, Liang Wang, Nan Yang, Yifan Song, Wenhao Wu, Furu Wei, Sujian Li, arxiv 2024. Github Repo for LongEmbed: https://github.com/dwzhu-pku/LongEmbed.
|
15 |
|
16 |
This model has 12 layers and the embedding size is 768.
|
17 |
|
|
|
47 |
|
48 |
## Training Details
|
49 |
|
50 |
+
Please refer to our paper at [https://arxiv.org/abs/2404.12096.pdf](https://arxiv.org/abs/2404.12096.pdf).
|
51 |
|
52 |
## Benchmark Evaluation
|
53 |
|
|
|
55 |
on the [BEIR](https://arxiv.org/abs/2104.08663) and [MTEB benchmark](https://arxiv.org/abs/2210.07316).
|
56 |
|
57 |
|
58 |
+
Please note that E5-RoPE-Base is not specifically trained for optimized performance. Its purpose is to enable performance comparisons between embedding models that utilize absolute position embeddings (APE) and rotary position embeddings (RoPE). By comparing E5-Base and E5-RoPE-Base, we demonstrate the superiority of RoPE-based embedding models in effectively managing longer context. See our paper [LongEmbed: Extending Embedding Models for Long Context Retrieval](https://arxiv.org/abs/2404.12096) for more details.
|
59 |
|
60 |
|
61 |
## Citation
|
|
|
63 |
If you find our paper or models helpful, please consider cite as follows:
|
64 |
|
65 |
```
|
66 |
+
@article{zhu2024longembed,
|
67 |
+
title={LongEmbed: Extending Embedding Models for Long Context Retrieval},
|
68 |
+
author={Zhu, Dawei and Wang, Liang and Yang, Nan and Song, Yifan and Wu, Wenhao and Wei, Furu and Li, Sujian},
|
69 |
+
journal={arXiv preprint arXiv:2404.12096},
|
70 |
+
year={2024}
|
71 |
}
|
72 |
```
|