Malaysian Finetuned Instruct LoRA
Collection
Continue finetuning Instruct model using LoRA from 0.5B up to 72B.
•
18 items
•
Updated
Continue finetuning https://huggingface.co/Qwen/Qwen3-14B on highly curated 1.5B tokens Malaysian instruction dataset.
Finetune on mesolitica/Malaysian-SFT to make the model understand Malaysian context.
["q_proj", "k_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj", "embed_tokens", "lm_head"]
.Source code at https://github.com/mesolitica/malaya/tree/master/session/qwen3
Based on 0-shot official MalayMMLU First token accuracy,
While the original model,
Based on 0-shot exact first token match using vLLM Guided Decoding,
While the original model,
Special thanks to https://www.sns.com.my and Nvidia for 8x H100 node!