zhyncs commited on
Commit
964b1bc
·
verified ·
1 Parent(s): 514304e

Upload LlamaForCausalLMEagle3

Browse files
config.json CHANGED
@@ -4,25 +4,35 @@
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
- "bos_token_id": 1,
8
  "draft_vocab_size": 32000,
9
- "eos_token_id": 2,
 
 
 
 
 
 
 
 
 
 
 
10
  "head_dim": 128,
11
  "hidden_act": "silu",
12
  "hidden_size": 5120,
13
  "initializer_range": 0.02,
14
  "intermediate_size": 32768,
15
- "max_position_embeddings": 2048,
16
  "mlp_bias": false,
17
  "model_type": "llama",
18
  "num_attention_heads": 40,
19
  "num_hidden_layers": 1,
20
  "num_key_value_heads": 8,
21
- "pad_token_id": 0,
22
  "pretraining_tp": 1,
23
  "rms_norm_eps": 1e-05,
24
  "rope_scaling": null,
25
- "rope_theta": 10000.0,
26
  "tie_word_embeddings": false,
27
  "torch_dtype": "float32",
28
  "transformers_version": "4.53.2",
 
4
  ],
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
+ "bos_token_id": null,
8
  "draft_vocab_size": 32000,
9
+ "eagle_config": {
10
+ "eagle_aux_hidden_state_layer_ids": [
11
+ 1,
12
+ 23,
13
+ 44
14
+ ],
15
+ "use_aux_hidden_state": true,
16
+ "use_input_layernorm_in_first_layer": true,
17
+ "use_last_layernorm": true,
18
+ "use_mtp_layernorm": false
19
+ },
20
+ "eos_token_id": null,
21
  "head_dim": 128,
22
  "hidden_act": "silu",
23
  "hidden_size": 5120,
24
  "initializer_range": 0.02,
25
  "intermediate_size": 32768,
26
+ "max_position_embeddings": 1048576,
27
  "mlp_bias": false,
28
  "model_type": "llama",
29
  "num_attention_heads": 40,
30
  "num_hidden_layers": 1,
31
  "num_key_value_heads": 8,
 
32
  "pretraining_tp": 1,
33
  "rms_norm_eps": 1e-05,
34
  "rope_scaling": null,
35
+ "rope_theta": 500000.0,
36
  "tie_word_embeddings": false,
37
  "torch_dtype": "float32",
38
  "transformers_version": "4.53.2",
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a49d698ac7f1ec7c8771eec4d16473674f16c8dc1c1089d9d41384c10d3d247b
3
  size 4536860696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92f4188cdc145800df636a2c91f278f820fb56d25a8d3087b93cd55793f3b266
3
  size 4536860696
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a9df21baa435f3635340dd8686522e12398d45d9b5df1184e1c905611b08998
3
  size 2983281568
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c070461f189c1e7e893f8f8c0d88e641b5282c8190262fc0c5fba2196ad0ad3
3
  size 2983281568