--- base_model: openai/whisper-small library_name: peft license: mit tags: - whisper-small - speech_to_text - ASR - french language: - fr demo: https://huggingface.co/spaces/visalkao/whisper-small-french-finetuned --- # Model Card for Model ID ## Model Details ### Model Description - **Developed by:** Visal KAO - **Model type:** Speech Recognition - **Language(s) (NLP):** French - **License:** MIT - **Finetuned from model :** Whisper-small ### Model Sources [optional] - **Repository:** openai/whisper-small ## Dataset This model is finetuned on 50% of French Single Speaker Speech Dataset on kaggle (Only lesmis). - **Link to dataset :** (https://www.kaggle.com/datasets/bryanpark/french-single-speaker-speech-dataset) ## Uses The goal of this project is to finetune whisper-small model to improve its accuracy for french transcription. The reason why I chose Whisper-small is due to its size and versatility. My primary objective is to build/finetune a small model to get acceptable results. ### Direct Use **Live Demo :** https://huggingface.co/spaces/visalkao/whisper-small-french-finetuned ## Bias, Risks, and Limitations As this model has less than 250 millions parameters, which is quite small considering its objective is to transcribe speech, it also has its own limitation. The Word Error Rate (WER) of this finetuned model is approximately 0.17 (17%). For reference, the original Whisper-small's WER is around 0.27 (27%) on the same dataset. ## Training Hyperparameters This model is trained using LoRa with these hyperparamters: * per_device_train_batch_size=3, * gradient_accumulation_steps=1, * learning_rate=1e-3, * num_train_epochs=7, * evaluation_strategy="epoch", * fp16=True, * per_device_eval_batch_size=1, * generation_max_length=225, * logging_steps=10, * remove_unused_columns=False, * label_names=["labels"], * predict_with_generate=True, ## Results Before finetuning, The Word Error Rate on this dataset is approximately 0.27. After finetuning, it drops down 0.1 to 0.17 or 17% wer (On testing data). Here is the training log: | Epoch | Training Loss | Validation Loss | WER | |-------|--------------|----------------|------------| | 1 | 0.369600 | 0.404414 | 26.665379 | | 2 | 0.273200 | 0.361762 | 22.793976 | | 3 | 0.308800 | 0.344289 | 24.454528 | | 4 | 0.131600 | 0.318023 | 21.847847 | | 5 | 0.117400 | 0.311023 | 19.134968 | | 6 | 0.035700 | 0.301410 | 18.922572 | | 7 | 0.013900 | 0.315151 | 16.972388 |