peralp24 commited on
Commit
1076870
·
verified ·
1 Parent(s): 5f23e11

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -11
README.md CHANGED
@@ -186,17 +186,19 @@ Compare the similarity
186
  We leave the embeddings of the documents untouched and now obtain the following cosine similarities:
187
  Query vs. German TV show: ~0.632
188
  Query vs. Italian polymath: ~0.512
189
- These new cosine similarities imply that the ranking has indeed changed and the paragraph about the German TV show is now more relevant. This shows that instructions can help the model understand nuances in the data better and ultimately lead to embeddings that are more useful for your use-case.
190
- Tips on using the model
191
-
192
-
193
- First try and ideally evaluate the model on your data without instructions to see whether performance aligns with your expectations out-of-the-box
194
- If you decide to use an instruction with the aim of further boosting performance we suggest using this template as a guideline
195
- Template: Represent the [X] to find a [Y] that [describe how the X and Y relate]
196
- Examples
197
- Represent the newspaper paragraph to find a newspaper paragraph with the same topic
198
- Represent the sentence to find another sentence with the same meaning
199
- In cases where the two texts to compare are different in nature (e.g. query and document) – also called “asymmetric” – we suggest to first add an instruction to query texts only. Again, try and ideally evaluate the model in this setting. Then, if your aim is to further boost performance, we suggest that you add instructions to document texts as well where [X] and [Y] are flipped accordingly.
 
 
200
 
201
 
202
 
 
186
  We leave the embeddings of the documents untouched and now obtain the following cosine similarities:
187
  Query vs. German TV show: ~0.632
188
  Query vs. Italian polymath: ~0.512
189
+ These new cosine similarities imply that the ranking has indeed changed and the paragraph about the German TV show is
190
+ **now more relevant**. This shows that instructions can help the model understand nuances in the data better
191
+ and ultimately lead to embeddings that are more useful for your use-case.
192
+
193
+ #### Tips on using the model
194
+
195
+ - First try and ideally evaluate the model on your data without instructions to see whether performance aligns with your expectations out-of-the-box
196
+ - If you decide to use an instruction with the aim of further boosting performance we suggest using this template as a guideline
197
+ * ```Template: Represent the [X] to find a [Y] that [describe how the X and Y relate]```
198
+ * Examples
199
+ 1. Represent the newspaper paragraph to find a newspaper paragraph with the same topic
200
+ 2. Represent the sentence to find another sentence with the same meaning
201
+ - In cases where the two texts to compare are different in nature (e.g. query and document) – also called “asymmetric” – we suggest to first add an instruction to query texts only. Again, try and ideally evaluate the model in this setting. Then, if your aim is to further boost performance, we suggest that you add instructions to document texts as well where [X] and [Y] are flipped accordingly.
202
 
203
 
204