|
{ |
|
"act_fun": "silu", |
|
"add_bos_token": false, |
|
"architectures": [ |
|
"HgrnForCausalLM" |
|
], |
|
"auto_map": { |
|
"AutoConfig": "configuration_hgrn.HgrnConfig", |
|
"AutoModelForCausalLM": "modeling_hgrn.HgrnForCausalLM" |
|
}, |
|
"bias": false, |
|
"bos_token_id": 50260, |
|
"causal": true, |
|
"decoder_embed_dim": 1024, |
|
"decoder_layers": 22, |
|
"eos_token_id": 50260, |
|
"glu_act": "swish", |
|
"glu_dim": 2048, |
|
"init_std": 0.02, |
|
"model_type": "hgrn", |
|
"no_scale_embedding": false, |
|
"norm_type": "layernorm", |
|
"pad_token_id": null, |
|
"torch_dtype": "float32", |
|
"transformers_version": "4.31.0", |
|
"use_cache": true, |
|
"use_triton": false, |
|
"vocab_size": 50272 |
|
} |
|
|