cgus commited on
Commit
9763ed9
·
verified ·
1 Parent(s): 0e8e10b

Upload 7 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ muse.jpg filter=lfs diff=lfs merge=lfs -text
37
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - mistralai/Mistral-Nemo-Base-2407
7
+ tags:
8
+ - text adventure
9
+ - roleplay
10
+ library_name: transformers
11
+ ---
12
+
13
+ ![image/jpeg](muse.jpg)
14
+
15
+ # Muse-12B
16
+
17
+ Muse brings an extra dimension to any tale—whether you're exploring a fantastical realm, court intrigue, or slice-of-life scenarios where a conversation can be as meaningful as a quest. While it handles adventure capably, Muse truly shines when character relationships and emotions are at the forefront, delivering impressive narrative coherence over long contexts.
18
+
19
+ If you want to easily try this model for free, you can do so at [https://aidungeon.com](https://aidungeon.com/).
20
+
21
+ We plan to continue improving and open-sourcing similar models, so please share any and all feedback on how we can improve model behavior. Below we share more details on how Muse was created.
22
+
23
+ [Quantized GGUF weights can be downloaded here.](https://huggingface.co/LatitudeGames/Muse-12B-GGUF)
24
+
25
+ ## Model details
26
+
27
+ Muse 12B was trained using Mistral Nemo 12B as its foundation, with training occurring in three stages: SFT (supervised fine-tuning), followed by two distinct DPO (direct preference optimization) phases.
28
+
29
+ **SFT** - Various multi-turn datasets from a multitude of sources, combining text adventures of the kind used to finetune [our Wayfarer 12B model](https://huggingface.co/LatitudeGames/Wayfarer-12B), long emotional narratives and general roleplay, each carefully balanced and rewritten to be free of common AI cliches. A small single-turn instruct dataset was included to send a stronger signal during finetuning.
30
+
31
+ **DPO 1** - Gutenberg DPO, [credit to Jon Durbin](https://huggingface.co/datasets/jondurbin/gutenberg-dpo-v0.1) - This stage introduces human writing techniques, significantly enhancing the model's potential outputs, albeit trading some intelligence for the stylistic benefits of human-created text.
32
+
33
+ **DPO 2** - Reward Model User Preference Data, [detailed in our blog](https://blog.latitude.io/all-posts/synthetic-data-preference-optimization-and-reward-models) - This stage refines the Gutenberg stage's "wildness," restoring intelligence while maintaining enhanced writing quality and providing a final level of enhancement due to the reward model samples.
34
+
35
+ The result is a model that writes like no other: versatile across genres, natural in expression, and suited to emotional depth.
36
+
37
+ ## Inference
38
+
39
+ The Nemo architecture is known for being sensitive to higher temperatures, so the following settings are recommended as a baseline. Nothing stops you from experimenting with these, of course.
40
+
41
+ ```
42
+ "temperature": 0.8,
43
+ "repetition_penalty": 1.05,
44
+ "min_p": 0.025
45
+ ```
46
+
47
+ ## Limitations
48
+
49
+ Muse was trained exclusively on second-person present tense data (using “you”) in a narrative style. Other styles will work as well but may produce suboptimal results.
50
+
51
+ Average response lengths tend toward verbosity (1000+ tokens) due to the Gutenberg DPO influence, though this can be controlled through explicit instructions in the system prompt.
52
+
53
+ ## Prompt Format
54
+
55
+ ChatML was used during all training stages.
56
+
57
+ ```
58
+ <|im_start|>system
59
+ You're a masterful storyteller and gamemaster. Write in second person present tense (You are), crafting vivid, engaging narratives with authority and confidence.<|im_end|>
60
+ <|im_start|>user
61
+ > You peer into the darkness.
62
+ <|im_start|>assistant
63
+ You have been eaten by a grue.
64
+
65
+ GAME OVER
66
+ ```
67
+
68
+ ## Credits
69
+
70
+ Thanks to [Gryphe Padar](https://huggingface.co/Gryphe) for collaborating on this finetune with us!
config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Muse-12B",
3
+ "architectures": [
4
+ "MistralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 131072,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 5120,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 14336,
14
+ "max_position_embeddings": 131072,
15
+ "model_type": "mistral",
16
+ "num_attention_heads": 32,
17
+ "num_hidden_layers": 40,
18
+ "num_key_value_heads": 8,
19
+ "rms_norm_eps": 1e-05,
20
+ "rope_theta": 1000000.0,
21
+ "sliding_window": null,
22
+ "tie_word_embeddings": false,
23
+ "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.46.1",
25
+ "use_cache": false,
26
+ "vocab_size": 131074,
27
+ "quantization_config": {
28
+ "quant_method": "exl2",
29
+ "version": "0.3.1",
30
+ "bits": 4.0,
31
+ "head_bits": 6,
32
+ "calibration": {
33
+ "rows": 115,
34
+ "length": 2048,
35
+ "dataset": "(default)"
36
+ }
37
+ }
38
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": 131072,
6
+ "transformers_version": "4.46.1"
7
+ }
muse.jpg ADDED

Git LFS Details

  • SHA256: 828316e65e3afe01164ab517aa6080a40ad2aee97488b4c3649b5f07c1a47859
  • Pointer size: 131 Bytes
  • Size of remote file: 645 kB
special_tokens_map.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>"
4
+ ],
5
+ "bos_token": {
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "eos_token": {
13
+ "content": "<|im_end|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false
18
+ },
19
+ "pad_token": {
20
+ "content": "<pad>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false
25
+ },
26
+ "unk_token": {
27
+ "content": "<unk>",
28
+ "lstrip": false,
29
+ "normalized": false,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ }
33
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2fa2956478eaa353c6c4b1f47fdd6868cce6075e52e169c35ae8bd28524e7a8
3
+ size 17078668
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff