Issue with WER evaluation when using a pre-trained model #7893
-
I'm using a pre-trained model:
It loads correctly, I can play any wav file:
So, I need to evaluate the model on this dataset without training. How I do that?
But I can't import GreedyCTCDecoder from NeMo's library |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
If you just need transcription, you can simply use model.transcribe() and pass a list of files. If there's a whole manifest of files, then you can use the transcribe_speech.py script in ASR examples. For your code though, wer Decoding can be found for ctc under asr.metrics.wer.py - https://github.com/NVIDIA/NeMo/blob/main/nemo/collections/asr/metrics/wer.py#L993 This is Char decoding, for subwords use CTCBPEDecoding in ctc_bpe.py |
Beta Was this translation helpful? Give feedback.
If you just need transcription, you can simply use model.transcribe() and pass a list of files. If there's a whole manifest of files, then you can use the transcribe_speech.py script in ASR examples.
For your code though, wer Decoding can be found for ctc under asr.metrics.wer.py - https://github.com/NVIDIA/NeMo/blob/main/nemo/collections/asr/metrics/wer.py#L993
This is Char decoding, for subwords use CTCBPEDecoding in ctc_bpe.py