base-bn-cp10k / README.md
Da4ThEdge's picture
Update README.md
d2cef7a verified
metadata
library_name: transformers
license: apache-2.0
datasets:
  - mozilla-foundation/common_voice_17_0
language:
  - bn
metrics:
  - wer
base_model:
  - Da4ThEdge/base-bn-lora-adapter-cp10k
model-index:
  - name: Whisper Base Bn (10k steps) - BanglaBridge
    results:
      - task:
          name: Automatic Speech Recognition
          type: automatic-speech-recognition
        dataset:
          name: Common Voice 17.0
          type: mozilla-foundation/common_voice_17_0
          config: bn
          split: None
          args: 'config: bn, split: test'
        metrics:
          - name: Wer
            type: wer
            value: 23.31617

Whisper Base Bn (10k steps) - by BanglaBridge

This model is a fine-tuned version of openai/whisper-base on the Common Voice 17.0 dataset.

It is the merged model from this fine-tuned PEFT LoRA adapter: Da4ThEdge/base-bn-lora-adapter-cp10k

After 10k steps it achieves the following results on the test set:

  • Wer: 46.25395
  • Normalized Wer: 23.31617

Refer to the 20k full-trained adapter repository for more details on the finetuning: banglabridge/base-bn-lora-adapter

Framework versions

  • Transformers 4.40.2
  • Pytorch 2.6.0+cu124
  • Tokenizers 0.19.1
  • Peft 0.10.0