Text Generation
Transformers
llama
Inference Endpoints
bhenrym14 commited on
Commit
b8db8fb
1 Parent(s): 360ba9f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -30,6 +30,8 @@ This model employs [Partial NTK Rope Scaling](https://github.com/jquesnelle/scal
30
  2. Autogptq/GPTQ-for-Llama. Use these quantized weights. Make the same replacement as in 1.
31
  3. Use ExLLama, replacing the `model.py` file with the [modified version](https://github.com/bhenrym14/qlora-airoboros-longcontext/blob/main/exllama_pntk/model.py). Use `compress_pos_emb=1` and `alpha_value = 1` (defaults). The necessary scaling values should flow from the configuration file. If you have done this correctly, there should be a dump of indications in the console indicating the scaling factor used (should be 4). If not, be sure your client is importing exllama from where you replaced the file. (ooba imported from sitepackages for me). I hacked this together very quickly so don't be surprised if something goes wrong. It shouldn't break functionality with normal models (as long as the model config file does not have `original_max_embeddings` defined) but I haven't tested this.
32
 
 
 
33
  Please comment with any questions. This hasn't been extensively tested.
34
 
35
  ## Motivation
@@ -50,10 +52,10 @@ Here I explore whether training on long sequences that have clear conceptual dep
50
  | 8192 | **4.90** | 5.32 | Not Tested | 57.1 |
51
  | 12000 | **4.82** | 56.1 | Not Tested | Not Tested |
52
 
53
- - This model is very competitive with the Llama-1 33b extended context variants.
54
  - Not presented here, but this model outperforms the base llama-2-13b on MMLU-fs with a score of 54.9. While perhaps an insignificant difference, the fact there isn't a clear performance regression despite the context extension is notable.
55
  - Perplexity continues to decline to 12000 tokens, the longest context length I tested due to VRAM constraints.
56
- - Feedback regarding real-world performance is appreciated. I don't know if the first dolphin training phase really contributed much; many relevant modeling components changed here, so it's difficult to make any specific attributions. The base model improvement may very well be the most dominant change.
57
 
58
  ## Quantization:
59
 
 
30
  2. Autogptq/GPTQ-for-Llama. Use these quantized weights. Make the same replacement as in 1.
31
  3. Use ExLLama, replacing the `model.py` file with the [modified version](https://github.com/bhenrym14/qlora-airoboros-longcontext/blob/main/exllama_pntk/model.py). Use `compress_pos_emb=1` and `alpha_value = 1` (defaults). The necessary scaling values should flow from the configuration file. If you have done this correctly, there should be a dump of indications in the console indicating the scaling factor used (should be 4). If not, be sure your client is importing exllama from where you replaced the file. (ooba imported from sitepackages for me). I hacked this together very quickly so don't be surprised if something goes wrong. It shouldn't break functionality with normal models (as long as the model config file does not have `original_max_embeddings` defined) but I haven't tested this.
32
 
33
+ **If using ooba, be sure to increase the `Truncate the prompt up to this length` parameter under the `parameters` tab to 16384.**
34
+
35
  Please comment with any questions. This hasn't been extensively tested.
36
 
37
  ## Motivation
 
52
  | 8192 | **4.90** | 5.32 | Not Tested | 57.1 |
53
  | 12000 | **4.82** | 56.1 | Not Tested | Not Tested |
54
 
55
+ - This model is very competitive with the Llama-1 33b extended context variants. In particular, at 512 tokens it has lower perplexity. This is probably an improvement imparted (in part) by the NTK by parts scaling method.
56
  - Not presented here, but this model outperforms the base llama-2-13b on MMLU-fs with a score of 54.9. While perhaps an insignificant difference, the fact there isn't a clear performance regression despite the context extension is notable.
57
  - Perplexity continues to decline to 12000 tokens, the longest context length I tested due to VRAM constraints.
58
+ - Feedback regarding real-world performance is appreciated. I don't know if the first dolphin training phase really contributed much beyond what pile did for the 33b-lxctx model; many relevant modeling components changed here, so it's difficult to make any specific attributions. The base model improvement may very well be the most dominant change.
59
 
60
  ## Quantization:
61