|
|
--- |
|
|
library_name: transformers |
|
|
license: apache-2.0 |
|
|
datasets: |
|
|
- mozilla-foundation/common_voice_17_0 |
|
|
language: |
|
|
- bn |
|
|
metrics: |
|
|
- wer |
|
|
base_model: |
|
|
- Da4ThEdge/base-bn-lora-adapter-cp10k |
|
|
model-index: |
|
|
- name: Whisper Base Bn (10k steps) - BanglaBridge |
|
|
results: |
|
|
- task: |
|
|
name: Automatic Speech Recognition |
|
|
type: automatic-speech-recognition |
|
|
dataset: |
|
|
name: Common Voice 17.0 |
|
|
type: mozilla-foundation/common_voice_17_0 |
|
|
config: bn |
|
|
split: None |
|
|
args: 'config: bn, split: test' |
|
|
metrics: |
|
|
- name: Wer |
|
|
type: wer |
|
|
value: 23.31617 |
|
|
--- |
|
|
|
|
|
# Whisper Base Bn (10k steps) - by BanglaBridge |
|
|
|
|
|
This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the Common Voice 17.0 dataset. |
|
|
|
|
|
It is the merged model from this fine-tuned PEFT LoRA adapter: [Da4ThEdge/base-bn-lora-adapter-cp10k](https://huggingface.co/Da4ThEdge/base-bn-lora-adapter-cp10k) |
|
|
|
|
|
After 10k steps it achieves the following results on the test set: |
|
|
- Wer: 46.25395 |
|
|
- Normalized Wer: 23.31617 |
|
|
|
|
|
Refer to the 20k full-trained adapter repository for more details on the finetuning: [banglabridge/base-bn-lora-adapter](https://huggingface.co/banglabridge/base-bn-lora-adapter) |
|
|
|
|
|
|
|
|
### Framework versions |
|
|
|
|
|
- Transformers 4.40.2 |
|
|
- Pytorch 2.6.0+cu124 |
|
|
- Tokenizers 0.19.1 |
|
|
- Peft 0.10.0 |