peterizsak commited on
Commit
8569aca
1 Parent(s): 2e43ad1

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -8
README.md CHANGED
@@ -6,11 +6,9 @@ language:
6
 
7
  # BGE-large-en-v1.5-rag-int8-static
8
 
9
- A quantized version of [BAAI/BGE-large-en-v1.5](https://huggingface.co/BAAI/bge-large-en-v1.5) embedder compatible with [Optimum-Intel](https://github.com/huggingface/optimum-intel) and [Intel® Neural Compressor](https://github.com/huggingface/optimum-intel).
10
 
11
- The model can be used with [Optimum-Intel](https://github.com/huggingface/optimum-intel) API and as an embedder/ranker model as part of [fastRAG](https://github.com/IntelLabs/fastRAG).
12
-
13
- See [model page](https://huggingface.co/BAAI/bge-large-en-v1.5) for full details on model architecture and training details.
14
 
15
  ## Technical details
16
 
@@ -27,9 +25,12 @@ Instructions how to reproduce the quantized model can be found [here](https://gi
27
 
28
  ## Evaluation - MTEB
29
 
 
 
30
  | | `INT8` | `FP32` | % diff |
31
  |---|:---:|:---:|:---:|
32
- | Reranking | 0.5997 | 0.6003 | -0.108% |
 
33
 
34
  ## Usage
35
 
@@ -38,15 +39,15 @@ Instructions how to reproduce the quantized model can be found [here](https://gi
38
  See [Optimum-intel](https://github.com/huggingface/optimum-intel) installation page for instructions how to install. Or run:
39
 
40
  ``` sh
41
- pip install -U optimum[neural-compressor] intel-extension-for-transformers
42
  ```
43
 
44
  Loading a model:
45
 
46
  ``` python
47
- from optimum.intel import INCModel
48
 
49
- model = INCModel.from_pretrained("Intel/bge-large-en-v1.5-rag-int8-static")
50
  ```
51
 
52
  Running inference:
 
6
 
7
  # BGE-large-en-v1.5-rag-int8-static
8
 
9
+ A quantized version of [BAAI/BGE-large-en-v1.5](https://huggingface.co/BAAI/bge-large-en-v1.5) quantized with [Intel® Neural Compressor](https://github.com/huggingface/optimum-intel) and compatible with [Optimum-Intel](https://github.com/huggingface/optimum-intel).
10
 
11
+ The model can be used with [Optimum-Intel](https://github.com/huggingface/optimum-intel) API and as a standalone model or as an embedder or ranker module as part of [fastRAG](https://github.com/IntelLabs/fastRAG) RAG pipeline.
 
 
12
 
13
  ## Technical details
14
 
 
25
 
26
  ## Evaluation - MTEB
27
 
28
+ Model performance on the [Massive Text Embedding Benchmark (MTEB)](https://huggingface.co/spaces/mteb/leaderboard) *retrieval* and *reranking* tasks.
29
+
30
  | | `INT8` | `FP32` | % diff |
31
  |---|:---:|:---:|:---:|
32
+ | Reranking | 0.5997 | 0.6003 | -0.108% |
33
+ | Retrieval | 0.5346 | 0.5429 | -1.53% |
34
 
35
  ## Usage
36
 
 
39
  See [Optimum-intel](https://github.com/huggingface/optimum-intel) installation page for instructions how to install. Or run:
40
 
41
  ``` sh
42
+ pip install -U optimum[neural-compressor, ipex] intel-extension-for-transformers
43
  ```
44
 
45
  Loading a model:
46
 
47
  ``` python
48
+ from optimum.intel import IPEXModel
49
 
50
+ model = IPEXModel.from_pretrained("Intel/bge-large-en-v1.5-rag-int8-static")
51
  ```
52
 
53
  Running inference: