Missing LM layers.

#1
by conscell - opened

Dear model creators,
It seems like this model doesn't include LM layers:

>>> model = AutoModelForMaskedLM.from_pretrained("nlp-waseda/roberta-base-japanese")
Some weights of RobertaForMaskedLM were not initialized from the model checkpoint at nlp-waseda/roberta-base-japanese and are newly initialized: ['lm_head.layer_norm.weight', 'lm_head.dense.bias', 'lm_head.dense.weight', 'lm_head.layer_norm.bias', 'lm_head.bias']

which makes impossible to do further fine-tuning on domain-specific data.
If possible, could you please include these layers to the base model. Large model doesn't have this problem, but infeasible to use due to GPU memory constraints.

Kawahara Lab at Waseda University org

Thank you for pointing this out. The LM head was missing when we changed the value of max_position_embeddings. We added the LM head again. Please try the latest model.

Prof. Kawahara, thank you very much for the update. New model worked perfectly.

conscell changed discussion status to closed

Sign up or log in to comment