Seeking Advice on Fine-tuning QWQ-32B Model

#62
by aaditya - opened

Hey Qwen Team, Great work as usual. It's a beast.

I'm planning to fine-tune the QWQ-32B model on a custom domain dataset and would appreciate some guidance from those with experience.

My Current Situation:
I have a dataset in Alpaca format

I'm unsure about the optimal fine-tuning approach for QWQ-32B

I do have few questions

  • Can QWQ-32B be effectively fine-tuned using the Alpaca format dataset, or would this be suboptimal?
  • Should I convert my data to use the format instead using DeepSeek or Claude?
  • Does QWQ-32B support QLoRA fine-tuning, or is full fine-tuning required?

Can you guide on these things. Yes, I would use QWQ template, but I am skeptical about whether I can still fine-tune with Alpaca format dataset with that template? I want the tag during inference for sure.

Thank you in advance for any insights!

The dataset format is {"instruction" : "", "input" : "", "output" : ""} `

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment