Triangle104 commited on
Commit
cea0c5f
·
verified ·
1 Parent(s): 43db93f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -0
README.md CHANGED
@@ -21,6 +21,63 @@ tags:
21
  This model was converted to GGUF format from [`EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1`](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
22
  Refer to the [original model card](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) for more details on the model.
23
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  ## Use with llama.cpp
25
  Install llama.cpp through brew (works on Mac and Linux)
26
 
 
21
  This model was converted to GGUF format from [`EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1`](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
22
  Refer to the [original model card](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) for more details on the model.
23
 
24
+ -
25
+ This model was converted to GGUF format from [`EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1`](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
26
+ Refer to the [original model card](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) for more details on the model.
27
+
28
+ A RP/storywriting specialist model, full-parameter finetune of Qwen2.5-7B on mixture of synthetic and natural data.
29
+
30
+ It uses Celeste 70B 0.1 data mixture, greatly expanding it to improve
31
+ versatility, creativity and "flavor" of the resulting model.
32
+
33
+ Version 0.1 notes:
34
+ Dataset was deduped and cleaned from
35
+ version 0.0, and learning rate was adjusted. Resulting model seems to be
36
+ stabler, and 0.0 problems with handling short inputs and min_p sampling
37
+ seem to be mostly gone.
38
+
39
+ Will be retrained once more, because this run crashed around e1.2 (out
40
+ of 3) (thanks, DeepSpeed, really appreciate it), and it's still
41
+ somewhat undertrained as a result.
42
+
43
+ Prompt format is ChatML.
44
+
45
+ Recommended sampler values:
46
+
47
+ Temperature: 0.87
48
+ Top-P: 0.81
49
+ Repetition Penalty: 1.03
50
+
51
+ Model appears to prefer lower temperatures (at least 0.9 and lower). Min-P seems to work now, as well.
52
+
53
+ Recommended SillyTavern presets (via CalamitousFelicitousness):
54
+
55
+ Context
56
+
57
+ Instruct and System Prompt
58
+
59
+ Training data:
60
+
61
+ Celeste 70B 0.1 data mixture minus Opus Instruct subset. See that model's card for details.
62
+ Kalomaze's Opus_Instruct_25k dataset, filtered for refusals.
63
+ A subset (1k rows) of ChatGPT-4o-WritingPrompts by Gryphe
64
+ A subset (2k rows) of Sonnet3.5-Charcards-Roleplay by Gryphe
65
+ A cleaned subset (~3k rows) of shortstories_synthlabels by Auri
66
+ Synthstruct and SynthRP datasets by Epiculous
67
+
68
+ Training time and hardware:
69
+
70
+ 2 days on 4x3090Ti (locally)
71
+
72
+ Model was trained by Kearm and Auri.
73
+
74
+ Special thanks:
75
+
76
+ to Gryphe, Lemmy, Kalomaze, Nopm and Epiculous for the data
77
+ to Alpindale for helping with FFT config for Qwen2.5
78
+ and to InfermaticAI's community for their continued support for our endeavors
79
+
80
+ ---
81
  ## Use with llama.cpp
82
  Install llama.cpp through brew (works on Mac and Linux)
83