You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When I run the following script after I pretrained the xlsr model with my own dataset, Im getting error below,
FileNotFoundError: [Errno 2] No such file or directory: '/home/edo/self-supervised-speech-recognition/manifest/dev_other.tsv'
I have generated the dictionary file as well as labeled data according to the tutorial in the page with the following command, python3 gen_dict.py --transcript_file /home/edo/self-supervised-speech-recognition/examples/label_audio/asr/transcript.txt --save_dir dictionary
However, I run this command: python3 finetune.py --transcript_file home/edo/self-supervised-speech-recognition/examples/label_audio/asr/transcript.txt --pretrain_model ../models/07-46-31/checkpoints/checkpoint_best.pt --dict_file dictionary/dict.ltr.txt
and get the error above.
Environment
OS: Ubuntu 18.04 LTS
Cuda Version: 11.2
Additional Context
However, I tried to change the model to be used in fine-tuning process with wav2vec_small.pt instead of models/07-46-31/checkpoints/checkpoint_best.pt, I dont why it works.
Am I missing something? But im pretty sure I've followed the instructions. Or it is a bug from the fairseq itself?
The text was updated successfully, but these errors were encountered:
When I run the following script after I pretrained the xlsr model with my own dataset, Im getting error below,
FileNotFoundError: [Errno 2] No such file or directory: '/home/edo/self-supervised-speech-recognition/manifest/dev_other.tsv'
I have generated the dictionary file as well as labeled data according to the tutorial in the page with the following command,
python3 gen_dict.py --transcript_file /home/edo/self-supervised-speech-recognition/examples/label_audio/asr/transcript.txt --save_dir dictionary
However, I run this command:
python3 finetune.py --transcript_file home/edo/self-supervised-speech-recognition/examples/label_audio/asr/transcript.txt --pretrain_model ../models/07-46-31/checkpoints/checkpoint_best.pt --dict_file dictionary/dict.ltr.txt
and get the error above.
Environment
OS: Ubuntu 18.04 LTS
Cuda Version: 11.2
Additional Context
However, I tried to change the model to be used in fine-tuning process with wav2vec_small.pt instead of models/07-46-31/checkpoints/checkpoint_best.pt, I dont why it works.
Am I missing something? But im pretty sure I've followed the instructions. Or it is a bug from the fairseq itself?
The text was updated successfully, but these errors were encountered: