dwzhu commited on
Commit
5fc89ad
1 Parent(s): a520fc8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -8
README.md CHANGED
@@ -11,7 +11,7 @@ license: mit
11
 
12
  # E5-RoPE-Base
13
 
14
- [LongEmbed: Extending Embedding Models for Long Context Retrieval](). Dawei Zhu, Liang Wang, Nan Yang, Yifan Song, Wenhao Wu, Furu Wei, Sujian Li, arxiv 2024. Github Repo for LongEmbed: https://github.com/dwzhu-pku/LongEmbed.
15
 
16
  This model has 12 layers and the embedding size is 768.
17
 
@@ -47,7 +47,7 @@ print(scores.tolist())
47
 
48
  ## Training Details
49
 
50
- Please refer to our paper at [https://arxiv.org/pdf/2212.03533.pdf](https://arxiv.org/pdf/2212.03533.pdf).
51
 
52
  ## Benchmark Evaluation
53
 
@@ -55,7 +55,7 @@ Check out [unilm/e5](https://github.com/microsoft/unilm/tree/master/e5) to repro
55
  on the [BEIR](https://arxiv.org/abs/2104.08663) and [MTEB benchmark](https://arxiv.org/abs/2210.07316).
56
 
57
 
58
- Please note that E5-RoPE-Base is not specifically trained for optimized performance. Its purpose is to enable performance comparisons between embedding models that utilize absolute position embeddings (APE) and rotary position embeddings (RoPE). By comparing E5-Base and E5-RoPE-Base, we demonstrate the superiority of RoPE-based embedding models in effectively managing longer context. See our paper [LongEmbed: Extending Embedding Models for Long Context Retrieval]() for more details.
59
 
60
 
61
  ## Citation
@@ -63,10 +63,10 @@ Please note that E5-RoPE-Base is not specifically trained for optimized performa
63
  If you find our paper or models helpful, please consider cite as follows:
64
 
65
  ```
66
- @article{wang2022text,
67
- title={Text Embeddings by Weakly-Supervised Contrastive Pre-training},
68
- author={Wang, Liang and Yang, Nan and Huang, Xiaolong and Jiao, Binxing and Yang, Linjun and Jiang, Daxin and Majumder, Rangan and Wei, Furu},
69
- journal={arXiv preprint arXiv:2212.03533},
70
- year={2022}
71
  }
72
  ```
 
11
 
12
  # E5-RoPE-Base
13
 
14
+ [LongEmbed: Extending Embedding Models for Long Context Retrieval](https://arxiv.org/abs/2404.12096). Dawei Zhu, Liang Wang, Nan Yang, Yifan Song, Wenhao Wu, Furu Wei, Sujian Li, arxiv 2024. Github Repo for LongEmbed: https://github.com/dwzhu-pku/LongEmbed.
15
 
16
  This model has 12 layers and the embedding size is 768.
17
 
 
47
 
48
  ## Training Details
49
 
50
+ Please refer to our paper at [https://arxiv.org/abs/2404.12096.pdf](https://arxiv.org/abs/2404.12096.pdf).
51
 
52
  ## Benchmark Evaluation
53
 
 
55
  on the [BEIR](https://arxiv.org/abs/2104.08663) and [MTEB benchmark](https://arxiv.org/abs/2210.07316).
56
 
57
 
58
+ Please note that E5-RoPE-Base is not specifically trained for optimized performance. Its purpose is to enable performance comparisons between embedding models that utilize absolute position embeddings (APE) and rotary position embeddings (RoPE). By comparing E5-Base and E5-RoPE-Base, we demonstrate the superiority of RoPE-based embedding models in effectively managing longer context. See our paper [LongEmbed: Extending Embedding Models for Long Context Retrieval](https://arxiv.org/abs/2404.12096) for more details.
59
 
60
 
61
  ## Citation
 
63
  If you find our paper or models helpful, please consider cite as follows:
64
 
65
  ```
66
+ @article{zhu2024longembed,
67
+ title={LongEmbed: Extending Embedding Models for Long Context Retrieval},
68
+ author={Zhu, Dawei and Wang, Liang and Yang, Nan and Song, Yifan and Wu, Wenhao and Wei, Furu and Li, Sujian},
69
+ journal={arXiv preprint arXiv:2404.12096},
70
+ year={2024}
71
  }
72
  ```