{ "architectures": [ "ArcticLSTMSpeculatorPreTrainedModel" ], "base_model_name_or_path": "Qwen/Qwen2.5-32B-Instruct", "base_model_archs": [ "Qwen2ForCausalLM" ], "input_hidden_dim": 5120, "inner_dim": "5120", "proj_dim": "5120", "emb_dim": "5120", "model_type": "mlp_speculator", "n_candidates": 3, "n_predict": 3, "scale_input": true, "tie_weights": true, "tie_lstm_embs": true, "top_k_tokens_per_head": [ 1, 1, 1 ], "torch_dtype": "bfloat16", "transformers_version": "4.47.0", "vocab_size": 152064, "method": "sum_lstm" }