Comparison to other open-source models and alternative training scenarios

#1
by abilmansplus - opened

Thank you for a great model!
It seems to perform better on open-source Kazakh datasets compared to whisper-turbo out-of-the-box.
Here are the results of my tests:

model WER on CV21 Kazakh test-set WER on FLEURS Kazakh test-set
whisper-base-kk 23.8% 19.3%
whisper-turbo (beam_size=5) 41.9% 24.8%

I have a couple of questions.
Why did you choose to fine-tune the base model?
Have you tried fine-tuning the turbo model?

Thank you.

Thank you for the additional tests! I chose the base model for deployment on an edge device for real-time applications. I haven't fine-tuned the turbo model, but the training recipe should be the same.

akuzdeuov changed discussion status to closed
Your need to confirm your account before you can post a new comment.

Sign up or log in to comment