Update README.md
Browse files
README.md
CHANGED
@@ -15,7 +15,7 @@ tags:
|
|
15 |
---
|
16 |
|
17 |
|
18 |
-
# SmolLM2-
|
19 |
|
20 |
## Table of Contents
|
21 |
|
@@ -27,11 +27,11 @@ tags:
|
|
27 |
|
28 |
## Model Summary
|
29 |
|
30 |
-
**SmolLM2-
|
31 |
|
32 |
Unlike traditional fine-tuning datasets that aim to improve specific benchmarks or metrics, the Human-Like-DPO-Dataset focuses on aligning the model's behavior with human preferences. This process enhances the model's ability to generate more natural, human-like responses, making it particularly well-suited for conversational applications.
|
33 |
|
34 |
-
By emphasizing response quality and relatability, SmolLM2-
|
35 |
|
36 |
|
37 |
### How to use
|
|
|
15 |
---
|
16 |
|
17 |
|
18 |
+
# SmolLM2-360M-Humanized
|
19 |
|
20 |
## Table of Contents
|
21 |
|
|
|
27 |
|
28 |
## Model Summary
|
29 |
|
30 |
+
**SmolLM2-360M-Humanized** is a fine-tuned version of the [SmolLM2-360M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM2-360M-Instruct) model, optimized using the Direct Preference Optimization (DPO) method. To do this we used the "[Human-Like-DPO-Dataset](https://huggingface.co/datasets/HumanLLMs/Human-Like-DPO-Dataset)" from [Human-Like LLMs](https://huggingface.co/HumanLLMs).
|
31 |
|
32 |
Unlike traditional fine-tuning datasets that aim to improve specific benchmarks or metrics, the Human-Like-DPO-Dataset focuses on aligning the model's behavior with human preferences. This process enhances the model's ability to generate more natural, human-like responses, making it particularly well-suited for conversational applications.
|
33 |
|
34 |
+
By emphasizing response quality and relatability, SmolLM2-360M-Humanized is designed to deliver an engaging and intuitive user experience in dialogue-based scenarios.
|
35 |
|
36 |
|
37 |
### How to use
|