Update README.md
Browse files
README.md
CHANGED
@@ -513,4 +513,6 @@ We took chunks of size 250 tokens, 500 tokens, and 1000 tokens randomly for each
|
|
513 |
|
514 |
We then used these chunks to generate questions and answers based on this text using a state-of-the-art LLM.
|
515 |
|
516 |
-
Finally, we selected negatives for each chunk using the similarity from the dense embeddings of the [BAAI/bge-m3](https://huggingface.co/BAAI/bge-m3) model.
|
|
|
|
|
|
513 |
|
514 |
We then used these chunks to generate questions and answers based on this text using a state-of-the-art LLM.
|
515 |
|
516 |
+
Finally, we selected negatives for each chunk using the similarity from the dense embeddings of the [BAAI/bge-m3](https://huggingface.co/BAAI/bge-m3) model.
|
517 |
+
|
518 |
+
The training data for this model can be found at [lightblue/kurage_training_data](https://huggingface.co/datasets/lightblue/kurage_training_data)
|