Update README.md
Browse files
README.md
CHANGED
@@ -104,13 +104,13 @@ arr, sampling_rate = librosa.load(file, sr=16000)
|
|
104 |
|
105 |
# Load whisper model and processor
|
106 |
processor = WhisperProcessor.from_pretrained("openai/whisper-small")
|
107 |
-
model = WhisperForConditionalGeneration.from_pretrained("
|
108 |
|
109 |
# Preprocessing
|
110 |
input_features = processor(arr, return_tensors="pt", sampling_rate=sampling_rate).input_features
|
111 |
|
112 |
# Prediction
|
113 |
-
forced_decoder_ids = processor.get_decoder_prompt_ids(language="
|
114 |
predicted_ids = model.generate(input_features, forced_decoder_ids=forced_decoder_ids)
|
115 |
transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)
|
116 |
|
|
|
104 |
|
105 |
# Load whisper model and processor
|
106 |
processor = WhisperProcessor.from_pretrained("openai/whisper-small")
|
107 |
+
model = WhisperForConditionalGeneration.from_pretrained("masters-thesis-vm/whispered_TIA_small_ad_tokenization_encoder_freezing_normal")
|
108 |
|
109 |
# Preprocessing
|
110 |
input_features = processor(arr, return_tensors="pt", sampling_rate=sampling_rate).input_features
|
111 |
|
112 |
# Prediction
|
113 |
+
forced_decoder_ids = processor.get_decoder_prompt_ids(language="en", task="transcribe")
|
114 |
predicted_ids = model.generate(input_features, forced_decoder_ids=forced_decoder_ids)
|
115 |
transcription = processor.batch_decode(predicted_ids, skip_special_tokens=True)
|
116 |
|