--- license: apache-2.0 base_model: openai/whisper-tiny tags: - generated_from_trainer model-index: - name: whisper-tiny-kor-430k-hf-ep100 results: [] --- # whisper-tiny-kor-430k-hf-ep100 This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4802 - Cer: 7.0810 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 128 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 100 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Cer | |:-------------:|:-----:|:------:|:---------------:|:-------:| | 0.0276 | 1.0 | 2969 | 0.1394 | 10.5081 | | 0.0115 | 2.0 | 5938 | 0.1684 | 9.0941 | | 0.0082 | 3.0 | 8907 | 0.1932 | 8.4145 | | 0.0063 | 4.0 | 11876 | 0.2197 | 7.8659 | | 0.0052 | 5.0 | 14845 | 0.2516 | 7.7668 | | 0.0047 | 6.0 | 17814 | 0.2314 | 8.0565 | | 0.004 | 7.0 | 20783 | 0.2270 | 8.3521 | | 0.0037 | 8.0 | 23752 | 0.2540 | 7.8283 | | 0.0033 | 9.0 | 26721 | 0.2586 | 7.9098 | | 0.0028 | 10.0 | 29690 | 0.2891 | 7.4437 | | 0.0029 | 11.0 | 32659 | 0.2796 | 7.7976 | | 0.0025 | 12.0 | 35628 | 0.2630 | 8.2731 | | 0.0025 | 13.0 | 38597 | 0.2955 | 7.8518 | | 0.0025 | 14.0 | 41566 | 0.2812 | 7.4797 | | 0.002 | 15.0 | 44535 | 0.2859 | 7.9954 | | 0.0023 | 16.0 | 47504 | 0.3172 | 7.2374 | | 0.002 | 17.0 | 50473 | 0.3382 | 7.5966 | | 0.0018 | 18.0 | 53442 | 0.3320 | 7.6383 | | 0.0018 | 19.0 | 56411 | 0.3197 | 7.6900 | | 0.0015 | 20.0 | 59380 | 0.3305 | 8.3678 | | 0.0016 | 21.0 | 62349 | 0.3409 | 7.5117 | | 0.0015 | 22.0 | 65318 | 0.3382 | 7.8556 | | 0.0016 | 23.0 | 68287 | 0.3282 | 7.5863 | | 0.0015 | 24.0 | 71256 | 0.3220 | 8.2449 | | 0.0013 | 25.0 | 74225 | 0.3272 | 7.7731 | | 0.0015 | 26.0 | 77194 | 0.3557 | 7.8019 | | 0.0014 | 27.0 | 80163 | 0.3807 | 7.3311 | | 0.0012 | 28.0 | 83132 | 0.3398 | 7.8117 | | 0.0013 | 29.0 | 86101 | 0.3892 | 7.6089 | | 0.001 | 30.0 | 89070 | 0.3876 | 7.7875 | | 0.0011 | 31.0 | 92039 | 0.3942 | 7.3922 | | 0.0012 | 32.0 | 95008 | 0.3836 | 8.0308 | | 0.0011 | 33.0 | 97977 | 0.3745 | 7.9775 | | 0.001 | 34.0 | 100946 | 0.3605 | 8.0117 | | 0.001 | 35.0 | 103915 | 0.3615 | 7.4853 | | 0.001 | 36.0 | 106884 | 0.3563 | 7.5916 | | 0.0009 | 37.0 | 109853 | 0.3469 | 7.4750 | | 0.0009 | 38.0 | 112822 | 0.3940 | 7.5919 | | 0.0009 | 39.0 | 115791 | 0.3771 | 7.5443 | | 0.0009 | 40.0 | 118760 | 0.3392 | 7.6593 | | 0.0009 | 41.0 | 121729 | 0.3498 | 7.6393 | | 0.0009 | 42.0 | 124698 | 0.3705 | 7.4474 | | 0.0008 | 43.0 | 127667 | 0.3758 | 7.2274 | | 0.0008 | 44.0 | 130636 | 0.3944 | 7.6919 | | 0.0009 | 45.0 | 133605 | 0.3885 | 7.5565 | | 0.0008 | 46.0 | 136574 | 0.3830 | 7.4628 | | 0.0008 | 47.0 | 139543 | 0.3972 | 7.8546 | | 0.0008 | 48.0 | 142512 | 0.3875 | 7.4916 | | 0.0007 | 49.0 | 145481 | 0.3438 | 7.2606 | | 0.0007 | 50.0 | 148450 | 0.3540 | 7.1581 | | 0.0008 | 51.0 | 151419 | 0.3768 | 7.1712 | | 0.0007 | 52.0 | 154388 | 0.4050 | 7.2286 | | 0.0007 | 53.0 | 157357 | 0.3785 | 7.4637 | | 0.0008 | 54.0 | 160326 | 0.4145 | 7.4800 | | 0.0008 | 55.0 | 163295 | 0.4042 | 7.3791 | | 0.0006 | 56.0 | 166264 | 0.3885 | 7.6994 | | 0.0006 | 57.0 | 169233 | 0.4153 | 7.5440 | | 0.0006 | 58.0 | 172202 | 0.4111 | 7.3408 | | 0.0006 | 59.0 | 175171 | 0.4147 | 7.2872 | | 0.0006 | 60.0 | 178140 | 0.4209 | 7.6270 | | 0.0006 | 61.0 | 181109 | 0.4041 | 7.4258 | | 0.0006 | 62.0 | 184078 | 0.4032 | 7.5324 | | 0.0006 | 63.0 | 187047 | 0.4214 | 7.3687 | | 0.0005 | 64.0 | 190016 | 0.3991 | 7.2750 | | 0.0005 | 65.0 | 192985 | 0.3885 | 7.1731 | | 0.0006 | 66.0 | 195954 | 0.4087 | 7.5063 | | 0.0005 | 67.0 | 198923 | 0.3760 | 7.4913 | | 0.0005 | 68.0 | 201892 | 0.3929 | 7.3314 | | 0.0005 | 69.0 | 204861 | 0.4044 | 7.5173 | | 0.0005 | 70.0 | 207830 | 0.4075 | 7.2712 | | 0.0005 | 71.0 | 210799 | 0.4170 | 7.2415 | | 0.0005 | 72.0 | 213768 | 0.4148 | 7.1142 | | 0.0005 | 73.0 | 216737 | 0.4271 | 7.3020 | | 0.0005 | 74.0 | 219706 | 0.4281 | 7.1863 | | 0.0004 | 75.0 | 222675 | 0.4202 | 7.1543 | | 0.0005 | 76.0 | 225644 | 0.4320 | 7.2910 | | 0.0005 | 77.0 | 228613 | 0.4328 | 7.3995 | | 0.0005 | 78.0 | 231582 | 0.4304 | 7.2255 | | 0.0005 | 79.0 | 234551 | 0.4537 | 7.0023 | | 0.0005 | 80.0 | 237520 | 0.4544 | 7.2048 | | 0.0004 | 81.0 | 240489 | 0.4485 | 7.2167 | | 0.0005 | 82.0 | 243458 | 0.4564 | 7.1794 | | 0.0004 | 83.0 | 246427 | 0.4608 | 7.2145 | | 0.0004 | 84.0 | 249396 | 0.4724 | 7.2098 | | 0.0004 | 85.0 | 252365 | 0.4726 | 7.1424 | | 0.0004 | 86.0 | 255334 | 0.4754 | 7.2832 | | 0.0005 | 87.0 | 258303 | 0.4765 | 7.1709 | | 0.0004 | 88.0 | 261272 | 0.4610 | 7.1358 | | 0.0004 | 89.0 | 264241 | 0.4697 | 7.0797 | | 0.0004 | 90.0 | 267210 | 0.4717 | 7.0913 | | 0.0004 | 91.0 | 270179 | 0.4756 | 7.1017 | | 0.0004 | 92.0 | 273148 | 0.4766 | 7.2089 | | 0.0004 | 93.0 | 276117 | 0.4763 | 7.1057 | | 0.0004 | 94.0 | 279086 | 0.4764 | 7.1101 | | 0.0004 | 95.0 | 282055 | 0.4759 | 7.2170 | | 0.0004 | 96.0 | 285024 | 0.4772 | 7.1104 | | 0.0004 | 97.0 | 287993 | 0.4781 | 7.0819 | | 0.0004 | 98.0 | 290962 | 0.4798 | 7.0897 | | 0.0004 | 99.0 | 293931 | 0.4800 | 7.0872 | | 0.0004 | 100.0 | 296900 | 0.4802 | 7.0810 | ### Framework versions - Transformers 4.36.2 - Pytorch 2.0.1+cu117 - Datasets 2.16.1 - Tokenizers 0.15.0