File size: 1,280 Bytes
05e8ba9 62e1dd5 05e8ba9 62e1dd5 05e8ba9 62e1dd5 05e8ba9 62e1dd5 05e8ba9 62e1dd5 05e8ba9 62e1dd5 05e8ba9 62e1dd5 05e8ba9 62e1dd5 05e8ba9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 |
---
library_name: transformers
metrics:
- wer
- cer
model-index:
- name: Fine-Tuned-Whisper-Large-v3-Turbo-STT-Zeroth-KO-v2
results: []
datasets:
- o0dimplz0o/Zeroth-STT-Korean
language:
- ko
base_model:
- openai/whisper-large-v3-turbo
pipeline_tag: automatic-speech-recognition
---
## Whisper Large v2 Fine-Tuned for Korean ASR
This model is being fine-tuned from openai/whisper-large-v3-turbo on a custom dataset.
It currently achieves the following results on the evaluation set (still fine-tuning):
- Loss: 0.0164
- Wer: 19.9134
- Cer: 0.0660
## Model Description
This model is a version of `openai/whisper-large-v3-turbo`, currently still being incrementally fine-tune in stages, specifically optimized for Korean automatic speech recognition (ASR) tasks. The fine-tuning process aims to deliver high accuracy and timestamped transcriptions for Korean speech.
## Dataset Details
- **Dataset Source**: Custom dataset (https://huggingface.co/datasets/o0dimplz0o/Zeroth-STT-Korean)
- **Number of Samples**: 102,263
- **Split**: 93% train, 7% test
- **Data Augmentation**: 20% random, applied only to the training set
## Training Details
- **Hardware**: L40S GPU
- **Learning Rate Scheduler**: Cosine
- **Epochs**: [pending completion]
- **Optimizer**: AdamW Torch Fused
|