Model Card for whisper-large-v2-formosan-all

This model is a fine-tuned version of the Taiwanese indigenous openai/whisper-large-v2.
Note: we use indonesian as whisper language id

Training process

The training of the model was performed with the following hyperparameters

  • Batch size: 2*4 (on 4 L40s GPU)
  • Gradient accumulation steps: 64
  • Total steps: 4146
  • Learning rate: 5e-4
  • Data augmentation: No
  • Optimizer: schedule_free_radam
  • LR scheduler type: constant
Downloads last month
4
Safetensors
Model size
1.54B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for formospeech/whisper-large-v2-formosan-all

Finetuned
(193)
this model