Triangle104 commited on
Commit
8fa6178
·
verified ·
1 Parent(s): 93e5d87

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -0
README.md CHANGED
@@ -12,6 +12,31 @@ tags:
12
  This model was converted to GGUF format from [`DavidAU/Llama-3.1-1million-ctx-Dark-Planet-v1.01-8B`](https://huggingface.co/DavidAU/Llama-3.1-1million-ctx-Dark-Planet-v1.01-8B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
13
  Refer to the [original model card](https://huggingface.co/DavidAU/Llama-3.1-1million-ctx-Dark-Planet-v1.01-8B) for more details on the model.
14
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  ## Use with llama.cpp
16
  Install llama.cpp through brew (works on Mac and Linux)
17
 
 
12
  This model was converted to GGUF format from [`DavidAU/Llama-3.1-1million-ctx-Dark-Planet-v1.01-8B`](https://huggingface.co/DavidAU/Llama-3.1-1million-ctx-Dark-Planet-v1.01-8B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
13
  Refer to the [original model card](https://huggingface.co/DavidAU/Llama-3.1-1million-ctx-Dark-Planet-v1.01-8B) for more details on the model.
14
 
15
+ ---
16
+ This model was converted to Nvidia's new "UltraLong8B" long context Llama 3.1 model structure (https://huggingface.co/nvidia/Llama-3.1-8B-UltraLong-1M-Instruct)
17
+ which allowed full transfer of "Dark Planet 8B" in all it's "glory" so
18
+ to speak. Due to Nvidia's structure, the new Dark Planet has attained
19
+ far greater long generation not only
20
+ in terms of context, but also coherence too. There is a also a bump in
21
+ overall performance as well.
22
+
23
+ This model has been designed to be relatively bullet proof and
24
+ operates with all parameters, including temp settings from 0 to 5.
25
+
26
+ It is an extraordinary compressed model, with a very low perplexity level (lower than Meta Llama3 Instruct).
27
+
28
+ It is for any writing, fiction or roleplay activity.
29
+
30
+ It requires Llama 3 template and/or "Command-R" template.
31
+
32
+ Suggest a context window of at least 8k, 16K is better... as this model will generate long outputs unless you set a hard limit.
33
+
34
+ Likewise, as this is an instruct model - the more instructions in
35
+ your prompt and/or system prompt - the greater the output quality.
36
+
37
+ IE: Less "guessing" equals far higher quality.
38
+
39
+ ---
40
  ## Use with llama.cpp
41
  Install llama.cpp through brew (works on Mac and Linux)
42