frank6626's picture
Update config.json
52bbfab verified
raw
history blame
586 Bytes
{
"architectures": [
"ArcticLSTMSpeculatorPreTrainedModel"
],
"base_model_name_or_path": "Qwen/Qwen2.5-32B-Instruct",
"base_model_archs": [
"Qwen2ForCausalLM"
],
"input_hidden_dim": 5120,
"inner_dim": "5120",
"proj_dim": "5120",
"emb_dim": "5120",
"model_type": "mlp_speculator",
"n_candidates": 3,
"n_predict": 3,
"scale_input": true,
"tie_weights": true,
"tie_lstm_embs": true,
"top_k_tokens_per_head": [
1,
1,
1
],
"torch_dtype": "bfloat16",
"transformers_version": "4.47.0",
"vocab_size": 152064,
"method": "sum_lstm"
}