whisper-large-sorani-v2

This model is a fine-tuned version of openai/whisper-large-v3 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.2079
  • Wer: 18.3752

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 3e-05
  • train_batch_size: 16
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 750
  • num_epochs: 15
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Wer
0.2256 0.3365 500 0.2478 42.8772
0.1959 0.6729 1000 0.2122 38.0271
0.1621 1.0094 1500 0.1883 34.3037
0.1345 1.3459 2000 0.1718 31.6828
0.1283 1.6824 2500 0.1607 29.1248
0.0689 2.0188 3000 0.1512 27.2824
0.0806 2.3553 3500 0.1542 27.9594
0.0827 2.6918 4000 0.1463 26.9971
0.0503 3.0283 4500 0.1472 25.0242
0.0417 3.3647 5000 0.1572 25.6141
0.0557 3.7012 5500 0.1466 24.4052
0.0294 4.0377 6000 0.1618 24.8017
0.0342 4.3742 6500 0.1564 23.3897
0.0349 4.7106 7000 0.1528 23.9942
0.0183 5.0471 7500 0.1598 24.6180
0.0264 5.3836 8000 0.1700 25.1402
0.0213 5.7201 8500 0.1660 22.6596
0.0113 6.0565 9000 0.1722 22.3211
0.0145 6.3930 9500 0.1791 23.3752
0.0146 6.7295 10000 0.1803 22.5048
0.0073 7.0659 10500 0.1766 21.4313
0.0081 7.4024 11000 0.1792 21.8617
0.0115 7.7389 11500 0.1792 21.6344
0.0051 8.0754 12000 0.1834 21.3153
0.0072 8.4118 12500 0.1829 21.3975
0.004 8.7483 13000 0.1846 21.1509
0.004 9.0848 13500 0.1845 21.1122
0.0053 9.4213 14000 0.1803 20.9768
0.004 9.7577 14500 0.1846 20.2176
0.002 10.0942 15000 0.1937 20.7930
0.0019 10.4307 15500 0.1940 20.4255
0.0029 10.7672 16000 0.1951 20.6190
0.0014 11.1036 16500 0.1897 19.8743
0.0009 11.4401 17000 0.1919 19.9662
0.0009 11.7766 17500 0.1990 19.7292
0.0013 12.1131 18000 0.1937 19.9275
0.0004 12.4495 18500 0.1974 19.2747
0.0006 12.7860 19000 0.2006 19.5116
0.0001 13.1225 19500 0.2016 18.9410
0.0002 13.4590 20000 0.2016 18.8346
0.0001 13.7954 20500 0.2016 18.4671
0.0002 14.1319 21000 0.2050 18.4091
0.0009 14.4684 21500 0.2074 18.4139
0.0 14.8048 22000 0.2079 18.3752

Framework versions

  • Transformers 4.51.3
  • Pytorch 2.6.0+cu124
  • Datasets 3.6.0
  • Tokenizers 0.21.4
Downloads last month
5
Safetensors
Model size
2B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for samil24/whisper-large-sorani-v2

Finetuned
(762)
this model

Evaluation results