File size: 575 Bytes
8882279 67b5bd5 8882279 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 |
{
"architectures": [
"ArcticLSTMSpeculatorPreTrainedModel"
],
"base_model_name_or_path": "Qwen/Qwen2.5-32B-Instruct",
"base_model_arch": "Qwen2ForCausalLM",
"input_hidden_dim": 5120,
"inner_dim": "5120",
"proj_dim": "5120",
"emb_dim": "5120",
"model_type": "mlp_speculator",
"n_candidates": 3,
"n_predict": 3,
"scale_input": true,
"tie_weights": true,
"tie_lstm_embs": true,
"top_k_tokens_per_head": [
1,
1,
1
],
"torch_dtype": "bfloat16",
"transformers_version": "4.47.0",
"vocab_size": 152064,
"method": "sum_lstm"
}
|