|
# Flashlight Decoder |
|
|
|
This script runs decoding for pre-trained speech recognition models. |
|
|
|
## Usage |
|
|
|
Assuming a few variables: |
|
|
|
```bash |
|
checkpoint=<path-to-checkpoint> |
|
data=<path-to-data-directory> |
|
lm_model=<path-to-language-model> |
|
lexicon=<path-to-lexicon> |
|
``` |
|
|
|
Example usage for decoding a fine-tuned Wav2Vec model: |
|
|
|
```bash |
|
python $FAIRSEQ_ROOT/examples/speech_recognition/new/infer.py --multirun \ |
|
task=audio_pretraining \ |
|
task.data=$data \ |
|
task.labels=ltr \ |
|
common_eval.path=$checkpoint \ |
|
decoding.type=kenlm \ |
|
decoding.lexicon=$lexicon \ |
|
decoding.lmpath=$lm_model \ |
|
dataset.gen_subset=dev_clean,dev_other,test_clean,test_other |
|
``` |
|
|
|
Example usage for using Ax to sweep WER parameters (requires `pip install hydra-ax-sweeper`): |
|
|
|
```bash |
|
python $FAIRSEQ_ROOT/examples/speech_recognition/new/infer.py --multirun \ |
|
hydra/sweeper=ax \ |
|
task=audio_pretraining \ |
|
task.data=$data \ |
|
task.labels=ltr \ |
|
common_eval.path=$checkpoint \ |
|
decoding.type=kenlm \ |
|
decoding.lexicon=$lexicon \ |
|
decoding.lmpath=$lm_model \ |
|
dataset.gen_subset=dev_other |
|
``` |
|
|