jsaizant commited on
Commit
6e2b1b9
·
2 Parent(s): 87e61e4 107c1d3

Merging version 1.1

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/home/jsaiz/storage5/gpfs/projects/bsc88/text/models/salamandra_quantized_speculative/salamandra-7b-base-gptq",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -20,10 +20,12 @@
20
  "num_key_value_heads": 8,
21
  "pretraining_tp": 1,
22
  "quantization_config": {
 
23
  "batch_size": 1,
24
  "bits": 4,
25
  "block_name_to_quantize": null,
26
  "cache_block_outputs": true,
 
27
  "damp_percent": 0.1,
28
  "dataset": null,
29
  "desc_act": false,
@@ -32,6 +34,7 @@
32
  },
33
  "group_size": 128,
34
  "max_input_length": null,
 
35
  "model_seqlen": null,
36
  "module_name_preceding_first_block": null,
37
  "modules_in_block_to_quantize": null,
@@ -43,12 +46,12 @@
43
  "use_cuda_fp16": false,
44
  "use_exllama": true
45
  },
46
- "rms_norm_eps": 1e-06,
47
  "rope_scaling": null,
48
  "rope_theta": 10000.0,
49
  "tie_word_embeddings": false,
50
  "torch_dtype": "float16",
51
- "transformers_version": "4.46.1",
52
  "use_cache": true,
53
  "vocab_size": 256000
54
  }
 
1
  {
2
+ "_name_or_path": "/gpfs/projects/bsc88/text/models/salamandra_quantized_speculative/V1.1/salamandra-7b-base-gptq",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
20
  "num_key_value_heads": 8,
21
  "pretraining_tp": 1,
22
  "quantization_config": {
23
+ "backend": null,
24
  "batch_size": 1,
25
  "bits": 4,
26
  "block_name_to_quantize": null,
27
  "cache_block_outputs": true,
28
+ "checkpoint_format": "gptq",
29
  "damp_percent": 0.1,
30
  "dataset": null,
31
  "desc_act": false,
 
34
  },
35
  "group_size": 128,
36
  "max_input_length": null,
37
+ "meta": null,
38
  "model_seqlen": null,
39
  "module_name_preceding_first_block": null,
40
  "modules_in_block_to_quantize": null,
 
46
  "use_cuda_fp16": false,
47
  "use_exllama": true
48
  },
49
+ "rms_norm_eps": 1e-05,
50
  "rope_scaling": null,
51
  "rope_theta": 10000.0,
52
  "tie_word_embeddings": false,
53
  "torch_dtype": "float16",
54
+ "transformers_version": "4.49.0",
55
  "use_cache": true,
56
  "vocab_size": 256000
57
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.46.1"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.49.0"
6
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:93ae95d403dcc8a5b5734e2d28adec308195ad7052bb27ce8616201c24a3b5f5
3
  size 4978074496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6683058ba94aaa2ed56db83b12d0c7622bcc4f017e293f9096159d3e6b244e15
3
  size 4978074496
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ade8eedbd4c807739ead3bf727a957f540e006e223500270868ef96b8764704b
3
  size 2167530192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbd21065a25a08ed4d33c4737c0ef33e29488ba4cc71c489bd9856450bafafd5
3
  size 2167530192
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:990527d1e7b98c027d386c742250b2f8517bd3adf98c46cc6c1c2f35b234c224
3
- size 37007559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e90b85b3e3b3ebfc6b9bafeb954b37f2435eed595738337e53f2a746d23d5a2
3
+ size 37007416
tokenizer.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa490e57cebce5cb1a0a5b1a5d3fa4de05aee53dc3a44791f1c3401db44d802d
3
- size 4813274
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab94ddf46d14f0279254858d53770c5319c5129d47291ee2bada530271cb1292
3
+ size 4813276
tokenizer_config.json CHANGED
@@ -1087,7 +1087,9 @@
1087
  "bos_token": "<s>",
1088
  "clean_up_tokenization_spaces": false,
1089
  "eos_token": "</s>",
1090
- "legacy": true,
 
 
1091
  "model_max_length": 1000000000000000019884624838656,
1092
  "pad_token": null,
1093
  "sp_model_kwargs": {},
 
1087
  "bos_token": "<s>",
1088
  "clean_up_tokenization_spaces": false,
1089
  "eos_token": "</s>",
1090
+ "extra_special_tokens": {},
1091
+ "legacy": false,
1092
+ "local_files_only": true,
1093
  "model_max_length": 1000000000000000019884624838656,
1094
  "pad_token": null,
1095
  "sp_model_kwargs": {},