GogetaBlueMUI commited on
Commit
f57632b
·
verified ·
1 Parent(s): b14b13e

End of training

Browse files
Files changed (2) hide show
  1. README.md +24 -23
  2. generation_config.json +1 -11
README.md CHANGED
@@ -3,40 +3,40 @@ library_name: transformers
3
  language:
4
  - ur
5
  license: apache-2.0
6
- base_model: GogetaBlueMUI/whisper-medium-ur
7
  tags:
8
  - generated_from_trainer
9
  datasets:
10
- - mozilla-foundation/common_voice_17_0
11
  metrics:
12
  - wer
13
  model-index:
14
- - name: Whisper Medium Ur - Muhammad Abdullah
15
  results:
16
  - task:
17
  name: Automatic Speech Recognition
18
  type: automatic-speech-recognition
19
  dataset:
20
- name: Common Voice 17.0
21
- type: mozilla-foundation/common_voice_17_0
22
  config: ur
23
  split: test
24
  args: 'config: ur, split: test'
25
  metrics:
26
  - name: Wer
27
  type: wer
28
- value: 28.35960496272455
29
  ---
30
 
31
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
32
  should probably proofread and complete it, then remove this comment. -->
33
 
34
- # Whisper Medium Ur - Muhammad Abdullah
35
 
36
- This model is a fine-tuned version of [GogetaBlueMUI/whisper-medium-ur](https://huggingface.co/GogetaBlueMUI/whisper-medium-ur) on the Common Voice 17.0 dataset.
37
  It achieves the following results on the evaluation set:
38
- - Loss: 0.4708
39
- - Wer: 28.3596
40
 
41
  ## Model description
42
 
@@ -55,29 +55,30 @@ More information needed
55
  ### Training hyperparameters
56
 
57
  The following hyperparameters were used during training:
58
- - learning_rate: 7e-06
59
- - train_batch_size: 32
60
- - eval_batch_size: 16
61
  - seed: 42
62
- - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
 
 
63
  - lr_scheduler_type: linear
64
- - lr_scheduler_warmup_steps: 90
65
- - training_steps: 800
66
  - mixed_precision_training: Native AMP
67
 
68
  ### Training results
69
 
70
  | Training Loss | Epoch | Step | Validation Loss | Wer |
71
  |:-------------:|:------:|:----:|:---------------:|:-------:|
72
- | 0.1373 | 0.5952 | 100 | 0.4251 | 27.3170 |
73
- | 0.081 | 1.1905 | 200 | 0.4459 | 30.8355 |
74
- | 0.0867 | 1.7857 | 300 | 0.4392 | 29.1326 |
75
- | 0.0352 | 2.3810 | 400 | 0.4708 | 28.3596 |
76
 
77
 
78
  ### Framework versions
79
 
80
- - Transformers 4.48.2
81
- - Pytorch 2.3.0+cu121
82
- - Datasets 3.2.0
83
  - Tokenizers 0.21.0
 
3
  language:
4
  - ur
5
  license: apache-2.0
6
+ base_model: openai/whisper-medium
7
  tags:
8
  - generated_from_trainer
9
  datasets:
10
+ - fsicoli/common_voice_19_0
11
  metrics:
12
  - wer
13
  model-index:
14
+ - name: Whisper Medium Ur - Your Name
15
  results:
16
  - task:
17
  name: Automatic Speech Recognition
18
  type: automatic-speech-recognition
19
  dataset:
20
+ name: Common Voice 19.0
21
+ type: fsicoli/common_voice_19_0
22
  config: ur
23
  split: test
24
  args: 'config: ur, split: test'
25
  metrics:
26
  - name: Wer
27
  type: wer
28
+ value: 27.720097349677363
29
  ---
30
 
31
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
32
  should probably proofread and complete it, then remove this comment. -->
33
 
34
+ # Whisper Medium Ur - Your Name
35
 
36
+ This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on the Common Voice 19.0 dataset.
37
  It achieves the following results on the evaluation set:
38
+ - Loss: 0.3564
39
+ - Wer: 27.7201
40
 
41
  ## Model description
42
 
 
55
  ### Training hyperparameters
56
 
57
  The following hyperparameters were used during training:
58
+ - learning_rate: 3e-06
59
+ - train_batch_size: 8
60
+ - eval_batch_size: 8
61
  - seed: 42
62
+ - gradient_accumulation_steps: 2
63
+ - total_train_batch_size: 16
64
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
65
  - lr_scheduler_type: linear
66
+ - lr_scheduler_warmup_steps: 150
67
+ - training_steps: 1500
68
  - mixed_precision_training: Native AMP
69
 
70
  ### Training results
71
 
72
  | Training Loss | Epoch | Step | Validation Loss | Wer |
73
  |:-------------:|:------:|:----:|:---------------:|:-------:|
74
+ | 0.3965 | 0.6557 | 500 | 0.3952 | 30.0288 |
75
+ | 0.3086 | 1.3108 | 1000 | 0.3665 | 27.9635 |
76
+ | 0.2877 | 1.9666 | 1500 | 0.3564 | 27.7201 |
 
77
 
78
 
79
  ### Framework versions
80
 
81
+ - Transformers 4.49.0
82
+ - Pytorch 2.5.1+cu121
83
+ - Datasets 3.4.1
84
  - Tokenizers 0.21.0
generation_config.json CHANGED
@@ -32,16 +32,6 @@
32
  "bos_token_id": 50257,
33
  "decoder_start_token_id": 50258,
34
  "eos_token_id": 50257,
35
- "forced_decoder_ids": [
36
- [
37
- 1,
38
- null
39
- ],
40
- [
41
- 2,
42
- 50359
43
- ]
44
- ],
45
  "is_multilingual": true,
46
  "lang_to_id": {
47
  "<|af|>": 50327,
@@ -246,5 +236,5 @@
246
  "transcribe": 50359,
247
  "translate": 50358
248
  },
249
- "transformers_version": "4.48.2"
250
  }
 
32
  "bos_token_id": 50257,
33
  "decoder_start_token_id": 50258,
34
  "eos_token_id": 50257,
 
 
 
 
 
 
 
 
 
 
35
  "is_multilingual": true,
36
  "lang_to_id": {
37
  "<|af|>": 50327,
 
236
  "transcribe": 50359,
237
  "translate": 50358
238
  },
239
+ "transformers_version": "4.49.0"
240
  }