Update README.md
Browse files
README.md
CHANGED
@@ -35,7 +35,8 @@ The model was also trained with a multi-user, multi-character paradigm, where us
|
|
35 |
- There is no real "system" role or label. System information or instructions can be added by using `::::user:` without any attached character name.
|
36 |
- OOC messages have been trained without character names, but in practice this doesn't seem to be a huge issue. The format is `(OOC: {{message}})`
|
37 |
- It is **necessary** to add the BOS token at the start of the prompt (`<|begin_of_text|>` in the case of Llama-3.1), otherwise performance will be significantly reduced.
|
38 |
-
- The model wasn't trained with an EOS token after each model/assistant turn. `::::` or `::::user` or `::::assistant` can be used as stopping strings just as effectively.
|
|
|
39 |
|
40 |
### Schematical example
|
41 |
The BOS token was omitted here. Messages and descriptions are short for the sake of brevity.
|
@@ -72,6 +73,17 @@ After meeting together, Anon and Nanashi have a little chat.
|
|
72 |
|
73 |
Note that in reality OOC messages might not necessarily always have the intended effect. This is something I'm trying to look into.
|
74 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
75 |
## Training details
|
76 |
[Unsloth](https://github.com/unslothai/unsloth) was used on a single RTX3090 24GB GPU with QLoRA finetuning.
|
77 |
|
@@ -102,6 +114,9 @@ lr_scheduler_kwargs = {
|
|
102 |
}
|
103 |
```
|
104 |
|
|
|
|
|
|
|
105 |
## Questions and answers
|
106 |
**Q. What's up with the name?**
|
107 |
|
|
|
35 |
- There is no real "system" role or label. System information or instructions can be added by using `::::user:` without any attached character name.
|
36 |
- OOC messages have been trained without character names, but in practice this doesn't seem to be a huge issue. The format is `(OOC: {{message}})`
|
37 |
- It is **necessary** to add the BOS token at the start of the prompt (`<|begin_of_text|>` in the case of Llama-3.1), otherwise performance will be significantly reduced.
|
38 |
+
- The model wasn't trained with an EOS token after each model/assistant turn. `::::` or `::::user` or `::::assistant` can be used as stopping strings just as effectively, at least in theory.
|
39 |
+
- Upon testing it appears that vLLM/Aphrodite cannot truly stop text generation without an EOS token, unlike text-generation-webui and llama.cpp. Your mileage may vary.
|
40 |
|
41 |
### Schematical example
|
42 |
The BOS token was omitted here. Messages and descriptions are short for the sake of brevity.
|
|
|
73 |
|
74 |
Note that in reality OOC messages might not necessarily always have the intended effect. This is something I'm trying to look into.
|
75 |
|
76 |
+
## Sampling settings
|
77 |
+
There are the settings I use for testing:
|
78 |
+
|
79 |
+
- Neutralize samplers
|
80 |
+
- Temperature: 1.0
|
81 |
+
- Min-p: 0.03
|
82 |
+
- DRY Repetition penalty
|
83 |
+
- Multiplier: 0.8
|
84 |
+
- Base: 1.75
|
85 |
+
- Allowed Length: 2
|
86 |
+
|
87 |
## Training details
|
88 |
[Unsloth](https://github.com/unslothai/unsloth) was used on a single RTX3090 24GB GPU with QLoRA finetuning.
|
89 |
|
|
|
114 |
}
|
115 |
```
|
116 |
|
117 |
+
### Eval / Train loss graph
|
118 |
+

|
119 |
+
|
120 |
## Questions and answers
|
121 |
**Q. What's up with the name?**
|
122 |
|