Safetensors
English
llama
Not-For-All-Audiences
lemonilia commited on
Commit
45ea8a0
·
verified ·
1 Parent(s): bf5e972

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -1
README.md CHANGED
@@ -35,7 +35,8 @@ The model was also trained with a multi-user, multi-character paradigm, where us
35
  - There is no real "system" role or label. System information or instructions can be added by using `::::user:` without any attached character name.
36
  - OOC messages have been trained without character names, but in practice this doesn't seem to be a huge issue. The format is `(OOC: {{message}})`
37
  - It is **necessary** to add the BOS token at the start of the prompt (`<|begin_of_text|>` in the case of Llama-3.1), otherwise performance will be significantly reduced.
38
- - The model wasn't trained with an EOS token after each model/assistant turn. `::::` or `::::user` or `::::assistant` can be used as stopping strings just as effectively.
 
39
 
40
  ### Schematical example
41
  The BOS token was omitted here. Messages and descriptions are short for the sake of brevity.
@@ -72,6 +73,17 @@ After meeting together, Anon and Nanashi have a little chat.
72
 
73
  Note that in reality OOC messages might not necessarily always have the intended effect. This is something I'm trying to look into.
74
 
 
 
 
 
 
 
 
 
 
 
 
75
  ## Training details
76
  [Unsloth](https://github.com/unslothai/unsloth) was used on a single RTX3090 24GB GPU with QLoRA finetuning.
77
 
@@ -102,6 +114,9 @@ lr_scheduler_kwargs = {
102
  }
103
  ```
104
 
 
 
 
105
  ## Questions and answers
106
  **Q. What's up with the name?**
107
 
 
35
  - There is no real "system" role or label. System information or instructions can be added by using `::::user:` without any attached character name.
36
  - OOC messages have been trained without character names, but in practice this doesn't seem to be a huge issue. The format is `(OOC: {{message}})`
37
  - It is **necessary** to add the BOS token at the start of the prompt (`<|begin_of_text|>` in the case of Llama-3.1), otherwise performance will be significantly reduced.
38
+ - The model wasn't trained with an EOS token after each model/assistant turn. `::::` or `::::user` or `::::assistant` can be used as stopping strings just as effectively, at least in theory.
39
+ - Upon testing it appears that vLLM/Aphrodite cannot truly stop text generation without an EOS token, unlike text-generation-webui and llama.cpp. Your mileage may vary.
40
 
41
  ### Schematical example
42
  The BOS token was omitted here. Messages and descriptions are short for the sake of brevity.
 
73
 
74
  Note that in reality OOC messages might not necessarily always have the intended effect. This is something I'm trying to look into.
75
 
76
+ ## Sampling settings
77
+ There are the settings I use for testing:
78
+
79
+ - Neutralize samplers
80
+ - Temperature: 1.0
81
+ - Min-p: 0.03
82
+ - DRY Repetition penalty
83
+ - Multiplier: 0.8
84
+ - Base: 1.75
85
+ - Allowed Length: 2
86
+
87
  ## Training details
88
  [Unsloth](https://github.com/unslothai/unsloth) was used on a single RTX3090 24GB GPU with QLoRA finetuning.
89
 
 
114
  }
115
  ```
116
 
117
+ ### Eval / Train loss graph
118
+ ![Train/loss graph](https://files.catbox.moe/y7rp37.png)
119
+
120
  ## Questions and answers
121
  **Q. What's up with the name?**
122