Triangle104 commited on
Commit
ef74610
·
verified ·
1 Parent(s): a728ae0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +39 -0
README.md CHANGED
@@ -16,6 +16,45 @@ base_model: DavidAU/Qwen3-8B-192k-Context-6X-Josiefied-Uncensored
16
  This model was converted to GGUF format from [`DavidAU/Qwen3-8B-192k-Context-6X-Josiefied-Uncensored`](https://huggingface.co/DavidAU/Qwen3-8B-192k-Context-6X-Josiefied-Uncensored) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
17
  Refer to the [original model card](https://huggingface.co/DavidAU/Qwen3-8B-192k-Context-6X-Josiefied-Uncensored) for more details on the model.
18
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  ## Use with llama.cpp
20
  Install llama.cpp through brew (works on Mac and Linux)
21
 
 
16
  This model was converted to GGUF format from [`DavidAU/Qwen3-8B-192k-Context-6X-Josiefied-Uncensored`](https://huggingface.co/DavidAU/Qwen3-8B-192k-Context-6X-Josiefied-Uncensored) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
17
  Refer to the [original model card](https://huggingface.co/DavidAU/Qwen3-8B-192k-Context-6X-Josiefied-Uncensored) for more details on the model.
18
 
19
+ ---
20
+ This repo is for Goekdeniz-Guelmez's excellent
21
+ "Josiefied-Qwen3-8B-abliterated-v1", modified from 32k (32768) context
22
+ to 192 k (196608) context modified using YARN as per tech notes at Qwen
23
+ repo.
24
+
25
+
26
+ ORG model repo for this fine tune:
27
+
28
+
29
+ [ https://huggingface.co/Goekdeniz-Guelmez/Josiefied-Qwen3-8B-abliterated-v1 ]
30
+
31
+
32
+ Max context on this version is : 192k (196608)
33
+
34
+
35
+ Suggest min context limit of : 8k to 16k for "thinking" / "output".
36
+
37
+
38
+ This model can output 2k to over 13k.
39
+
40
+
41
+ To improve long form output performance (especially creative):
42
+
43
+
44
+ Temp 1+, 2+ or higher.
45
+ Top k 100+
46
+ Rep pen 1.02-1.09
47
+
48
+
49
+ Use Jinja Template or CHATML template.
50
+
51
+
52
+ Please refer the QWEN model card for details, benchmarks, how to use, settings, turning reasoning on/off/ system roles etc etc :
53
+
54
+
55
+ [ https://huggingface.co/Qwen/Qwen3-8B ]
56
+
57
+ ---
58
  ## Use with llama.cpp
59
  Install llama.cpp through brew (works on Mac and Linux)
60