Update README.md
Browse files
README.md
CHANGED
@@ -32,8 +32,8 @@ model-index:
|
|
32 |
* **Base Model:** Qwen/Qwen3-Embedding-0.6B
|
33 |
* **Model Type:** Embedding Model
|
34 |
* **Fine-tuning Method:** Low-Rank Adaptation (LoRA)
|
35 |
-
* **Developer:**
|
36 |
-
* **Contact:**
|
37 |
* **Date:** July 13, 2025
|
38 |
|
39 |
---
|
@@ -63,30 +63,28 @@ This model is intended for:
|
|
63 |
* **Domain Specificity:** While fine-tuned, the model's performance may degrade on data significantly different from its training distribution.
|
64 |
* **Inherited Biases:** As it is based on a pre-trained model, it may inherit biases present in the original training data. Users should be aware of potential biases related to gender, race, religion, etc., and test for them in their specific applications.
|
65 |
* **Computational Resources:** While LoRA reduces resource demands for fine-tuning, inference still requires appropriate computational resources.
|
66 |
-
* **Language:** Primarily designed for
|
67 |
|
68 |
---
|
69 |
|
70 |
## Training Details
|
71 |
|
72 |
* **Fine-tuning Method:** LoRA (Low-Rank Adaptation)
|
73 |
-
* **LoRA Rank (r):**
|
74 |
* **LoRA Alpha (alpha):** [Specify LoRA alpha used, e.g., 16, 32, 64]
|
75 |
* **Target Modules for LoRA:** [Specify which modules were targeted, e.g., `q_proj`, `k_proj`, `v_proj`, `out_proj`]
|
76 |
-
* **Training Dataset:**
|
77 |
-
* **Dataset Size:**
|
78 |
* **Dataset Characteristics:** [Any relevant details about the dataset, e.g., "Contains highly technical language," "Focuses on conversational text," "Balanced across different topics."]
|
79 |
* **Training Hardware:** [e.g., NVIDIA A100 GPU, Google Cloud TPU v3]
|
80 |
* **Training Time:** [e.g., 4 hours, 2 days]
|
81 |
-
* **Optimization Strategy:**
|
82 |
-
* **Software Frameworks:**
|
83 |
|
84 |
---
|
85 |
|
86 |
## Performance Metrics
|
87 |
|
88 |
-
*(Note: Provide actual metrics from your evaluation. Examples below are placeholders.)*
|
89 |
-
|
90 |
* **Evaluation Dataset:** [Brief description of your evaluation dataset]
|
91 |
* **Metric 1 (e.g., Average Precision @ K):** [Value] (e.g., 0.85)
|
92 |
* **Metric 2 (e.g., Recall @ K):** [Value] (e.g., 0.92)
|
|
|
32 |
* **Base Model:** Qwen/Qwen3-Embedding-0.6B
|
33 |
* **Model Type:** Embedding Model
|
34 |
* **Fine-tuning Method:** Low-Rank Adaptation (LoRA)
|
35 |
+
* **Developer:** https://github.com/gauravprasadgp
|
36 |
+
* **Contact:** https://github.com/gauravprasadgp
|
37 |
* **Date:** July 13, 2025
|
38 |
|
39 |
---
|
|
|
63 |
* **Domain Specificity:** While fine-tuned, the model's performance may degrade on data significantly different from its training distribution.
|
64 |
* **Inherited Biases:** As it is based on a pre-trained model, it may inherit biases present in the original training data. Users should be aware of potential biases related to gender, race, religion, etc., and test for them in their specific applications.
|
65 |
* **Computational Resources:** While LoRA reduces resource demands for fine-tuning, inference still requires appropriate computational resources.
|
66 |
+
* **Language:** Primarily designed for English text. Performance on other languages may vary.
|
67 |
|
68 |
---
|
69 |
|
70 |
## Training Details
|
71 |
|
72 |
* **Fine-tuning Method:** LoRA (Low-Rank Adaptation)
|
73 |
+
* **LoRA Rank (r):** 16
|
74 |
* **LoRA Alpha (alpha):** [Specify LoRA alpha used, e.g., 16, 32, 64]
|
75 |
* **Target Modules for LoRA:** [Specify which modules were targeted, e.g., `q_proj`, `k_proj`, `v_proj`, `out_proj`]
|
76 |
+
* **Training Dataset:** Semantic Similar Sentences
|
77 |
+
* **Dataset Size:** 100
|
78 |
* **Dataset Characteristics:** [Any relevant details about the dataset, e.g., "Contains highly technical language," "Focuses on conversational text," "Balanced across different topics."]
|
79 |
* **Training Hardware:** [e.g., NVIDIA A100 GPU, Google Cloud TPU v3]
|
80 |
* **Training Time:** [e.g., 4 hours, 2 days]
|
81 |
+
* **Optimization Strategy:** AdamW
|
82 |
+
* **Software Frameworks:** PyTorch, Hugging Face Transformers, PEFT library
|
83 |
|
84 |
---
|
85 |
|
86 |
## Performance Metrics
|
87 |
|
|
|
|
|
88 |
* **Evaluation Dataset:** [Brief description of your evaluation dataset]
|
89 |
* **Metric 1 (e.g., Average Precision @ K):** [Value] (e.g., 0.85)
|
90 |
* **Metric 2 (e.g., Recall @ K):** [Value] (e.g., 0.92)
|