gauravprasadgp commited on
Commit
5e8ce65
·
verified ·
1 Parent(s): b7d808a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -10
README.md CHANGED
@@ -32,8 +32,8 @@ model-index:
32
  * **Base Model:** Qwen/Qwen3-Embedding-0.6B
33
  * **Model Type:** Embedding Model
34
  * **Fine-tuning Method:** Low-Rank Adaptation (LoRA)
35
- * **Developer:** [https://github.com/gauravprasadgp]
36
- * **Contact:** [https://github.com/gauravprasadgp]
37
  * **Date:** July 13, 2025
38
 
39
  ---
@@ -63,30 +63,28 @@ This model is intended for:
63
  * **Domain Specificity:** While fine-tuned, the model's performance may degrade on data significantly different from its training distribution.
64
  * **Inherited Biases:** As it is based on a pre-trained model, it may inherit biases present in the original training data. Users should be aware of potential biases related to gender, race, religion, etc., and test for them in their specific applications.
65
  * **Computational Resources:** While LoRA reduces resource demands for fine-tuning, inference still requires appropriate computational resources.
66
- * **Language:** Primarily designed for [Specify Language(s) if known, e.g., English] text. Performance on other languages may vary.
67
 
68
  ---
69
 
70
  ## Training Details
71
 
72
  * **Fine-tuning Method:** LoRA (Low-Rank Adaptation)
73
- * **LoRA Rank (r):** [Specify LoRA rank used, e.g., 8, 16, 32]
74
  * **LoRA Alpha (alpha):** [Specify LoRA alpha used, e.g., 16, 32, 64]
75
  * **Target Modules for LoRA:** [Specify which modules were targeted, e.g., `q_proj`, `k_proj`, `v_proj`, `out_proj`]
76
- * **Training Dataset:** [Brief description of your training dataset, e.g., "A proprietary dataset of customer reviews," "A collection of scientific abstracts," "Paragraphs from Wikipedia articles."]
77
- * **Dataset Size:** [Number of samples in your training dataset]
78
  * **Dataset Characteristics:** [Any relevant details about the dataset, e.g., "Contains highly technical language," "Focuses on conversational text," "Balanced across different topics."]
79
  * **Training Hardware:** [e.g., NVIDIA A100 GPU, Google Cloud TPU v3]
80
  * **Training Time:** [e.g., 4 hours, 2 days]
81
- * **Optimization Strategy:** [e.g., AdamW, learning rate schedule]
82
- * **Software Frameworks:** [e.g., PyTorch, Hugging Face Transformers, PEFT library]
83
 
84
  ---
85
 
86
  ## Performance Metrics
87
 
88
- *(Note: Provide actual metrics from your evaluation. Examples below are placeholders.)*
89
-
90
  * **Evaluation Dataset:** [Brief description of your evaluation dataset]
91
  * **Metric 1 (e.g., Average Precision @ K):** [Value] (e.g., 0.85)
92
  * **Metric 2 (e.g., Recall @ K):** [Value] (e.g., 0.92)
 
32
  * **Base Model:** Qwen/Qwen3-Embedding-0.6B
33
  * **Model Type:** Embedding Model
34
  * **Fine-tuning Method:** Low-Rank Adaptation (LoRA)
35
+ * **Developer:** https://github.com/gauravprasadgp
36
+ * **Contact:** https://github.com/gauravprasadgp
37
  * **Date:** July 13, 2025
38
 
39
  ---
 
63
  * **Domain Specificity:** While fine-tuned, the model's performance may degrade on data significantly different from its training distribution.
64
  * **Inherited Biases:** As it is based on a pre-trained model, it may inherit biases present in the original training data. Users should be aware of potential biases related to gender, race, religion, etc., and test for them in their specific applications.
65
  * **Computational Resources:** While LoRA reduces resource demands for fine-tuning, inference still requires appropriate computational resources.
66
+ * **Language:** Primarily designed for English text. Performance on other languages may vary.
67
 
68
  ---
69
 
70
  ## Training Details
71
 
72
  * **Fine-tuning Method:** LoRA (Low-Rank Adaptation)
73
+ * **LoRA Rank (r):** 16
74
  * **LoRA Alpha (alpha):** [Specify LoRA alpha used, e.g., 16, 32, 64]
75
  * **Target Modules for LoRA:** [Specify which modules were targeted, e.g., `q_proj`, `k_proj`, `v_proj`, `out_proj`]
76
+ * **Training Dataset:** Semantic Similar Sentences
77
+ * **Dataset Size:** 100
78
  * **Dataset Characteristics:** [Any relevant details about the dataset, e.g., "Contains highly technical language," "Focuses on conversational text," "Balanced across different topics."]
79
  * **Training Hardware:** [e.g., NVIDIA A100 GPU, Google Cloud TPU v3]
80
  * **Training Time:** [e.g., 4 hours, 2 days]
81
+ * **Optimization Strategy:** AdamW
82
+ * **Software Frameworks:** PyTorch, Hugging Face Transformers, PEFT library
83
 
84
  ---
85
 
86
  ## Performance Metrics
87
 
 
 
88
  * **Evaluation Dataset:** [Brief description of your evaluation dataset]
89
  * **Metric 1 (e.g., Average Precision @ K):** [Value] (e.g., 0.85)
90
  * **Metric 2 (e.g., Recall @ K):** [Value] (e.g., 0.92)