Update README.md
Browse files
README.md
CHANGED
|
@@ -1,4 +1,7 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
library_name: transformers
|
| 3 |
tags:
|
| 4 |
- 4-bit
|
|
@@ -6,10 +9,30 @@ tags:
|
|
| 6 |
- text-generation
|
| 7 |
- autotrain_compatible
|
| 8 |
- endpoints_compatible
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
pipeline_tag: text-generation
|
| 10 |
inference: false
|
| 11 |
quantized_by: Suparious
|
| 12 |
---
|
| 13 |
-
#
|
| 14 |
|
| 15 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
language:
|
| 3 |
+
- en
|
| 4 |
+
license: cc-by-nc-4.0
|
| 5 |
library_name: transformers
|
| 6 |
tags:
|
| 7 |
- 4-bit
|
|
|
|
| 9 |
- text-generation
|
| 10 |
- autotrain_compatible
|
| 11 |
- endpoints_compatible
|
| 12 |
+
- text-generation-inference
|
| 13 |
+
- transformers
|
| 14 |
+
- unsloth
|
| 15 |
+
- mistral
|
| 16 |
+
- trl
|
| 17 |
+
- sft
|
| 18 |
+
- Roleplay
|
| 19 |
+
- roleplay
|
| 20 |
+
base_model: SanjiWatsuki/Kunoichi-DPO-v2-7B
|
| 21 |
pipeline_tag: text-generation
|
| 22 |
inference: false
|
| 23 |
quantized_by: Suparious
|
| 24 |
---
|
| 25 |
+
# Alsebay/NarumashiRTS-7B-V2-1 AWQ
|
| 26 |
|
| 27 |
+
- Model creator: [Alsebay](https://huggingface.co/Alsebay)
|
| 28 |
+
- Original model: [NarumashiRTS-7B-V2-1](https://huggingface.co/Alsebay/NarumashiRTS-7B-V2-1)
|
| 29 |
+
|
| 30 |
+
## Model Summary
|
| 31 |
+
|
| 32 |
+
> [!Important]
|
| 33 |
+
> Still in experiment
|
| 34 |
+
|
| 35 |
+
Remake [version 2](https://huggingface.co/Alsebay/NarumashiRTS-V2) with safetensor format, more safety and stable method, nothing change too much (base on the model hash). But to be real, in the previous version 2, I used unsafety method to save pretrain model, which could lead apply Lora layer twice to model, that make model have terrible performance. (Thanks Unsloth community told me about this :D )
|
| 36 |
+
|
| 37 |
+
- **Finetuned with rough translate dataset, to increase the accuracy in TSF theme, which is not quite popular. (lewd dataset)**
|
| 38 |
+
- **Finetuned from model :** SanjiWatsuki/Kunoichi-DPO-v2-7B . Thank SanjiWatsuki a lot :)
|