Update README.md
Browse files
README.md
CHANGED
@@ -2,46 +2,23 @@
|
|
2 |
pipeline_tag: text-generation
|
3 |
---
|
4 |
|
5 |
-
Collection of resources and models for storytelling and roleplay. Updated
|
6 |
-
|
7 |
-
**Current favorite**: [mistralai/Mistral-Small-24B-Instruct-2501](https://huggingface.co/mistralai/Mistral-Small-24B-Instruct-2501)
|
8 |
|
9 |
Some notes on best usage:
|
10 |
-
-
|
11 |
-
-
|
12 |
-
-
|
13 |
-
- Conclusion: use original instruct models with short prompts
|
14 |
-
- actually, on second thought: this idea that the original instruct models have "higher intelligence" might be flawed. The other perspective is that the models are merely a representation of their dataset, which is why you might prefer finetunes and monster merges.
|
15 |
-
# ⚒️ Base models
|
16 |
-
|
17 |
-
- Llama 3 (8B) - the OG
|
18 |
-
- [Mistral-Nemo-Base-2407](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407) (12B)
|
19 |
-
- Qwen2.5
|
20 |
-
- Mistral Small
|
21 |
-
|
22 |
-
# 🤖 Instruct models
|
23 |
-
|
24 |
-
- Llama 3 Instruct
|
25 |
-
- Qwen 2.5 Instruct
|
26 |
-
- [Mistral-Nemo-Instruct-2407](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407) (12B)
|
27 |
-
- **Mistral Small (22B)** (instruct) - a winner?
|
28 |
-
|
29 |
-
# 😈 Story/roleplay-tuned models
|
30 |
-
|
31 |
-
- [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2)
|
32 |
-
- [Sao10K/Fimbulvetr-11B-v2](https://huggingface.co/Sao10K/Fimbulvetr-11B-v2)
|
33 |
-
- [nothingiisreal/MN-12B-Celeste-V1.9](https://huggingface.co/nothingiisreal/MN-12B-Celeste-V1.9)
|
34 |
-
- [anthracite-org/magnum-12b-v2](https://huggingface.co/anthracite-org/magnum-12b-v2)
|
35 |
-
- [NeverSleep/Lumimaid-v0.2-8B](https://huggingface.co/NeverSleep/Lumimaid-v0.2-8B)
|
36 |
-
- [nbeerbower/mistral-nemo-gutenberg-12B-v2](https://huggingface.co/nbeerbower/mistral-nemo-gutenberg-12B-v2)
|
37 |
|
38 |
-
#
|
39 |
|
40 |
-
|
41 |
-
|
42 |
-
|
|
|
|
|
|
|
43 |
|
44 |
-
#
|
45 |
|
46 |
- [llama.cpp](https://github.com/ggerganov/llama.cpp) and [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) - **preferred LLM software**
|
47 |
- [/r/localllama](https://www.reddit.com/r/LocalLLaMA/)
|
|
|
2 |
pipeline_tag: text-generation
|
3 |
---
|
4 |
|
5 |
+
Collection of resources and models for storytelling and roleplay. Updated Mar 2025.
|
|
|
|
|
6 |
|
7 |
Some notes on best usage:
|
8 |
+
- dont waste time on sampler settings; use recommended and optimize the prompt
|
9 |
+
- don't "overparameterize" by writing too long a prompt
|
10 |
+
- model size/intelligence is important but they are just mimics, the dataset is very important
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
|
12 |
+
# model ranking
|
13 |
|
14 |
+
1. [lars1234/Mistral-Small-24B-Instruct-2501-writer](https://huggingface.co/lars1234/Mistral-Small-24B-Instruct-2501-writer) - **wow** this one is great
|
15 |
+
1. [Sao10K/MN-12B-Lyra-v4](https://huggingface.co/Sao10K/MN-12B-Lyra-v4)
|
16 |
+
1. [MarinaraSpaghetti/NemoMix-Unleashed-12B](https://huggingface.co/MarinaraSpaghetti/NemoMix-Unleashed-12B)
|
17 |
+
1. [Sao10K/Fimbulvetr-11B-v2](https://huggingface.co/Sao10K/Fimbulvetr-11B-v2)
|
18 |
+
1. [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2)
|
19 |
+
1. [Mistral-Nemo-Instruct-2407](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407) (12B)
|
20 |
|
21 |
+
# misc links
|
22 |
|
23 |
- [llama.cpp](https://github.com/ggerganov/llama.cpp) and [llama-cpp-python](https://github.com/abetlen/llama-cpp-python) - **preferred LLM software**
|
24 |
- [/r/localllama](https://www.reddit.com/r/LocalLLaMA/)
|