Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Whisper finetune #184

Open
Macsim2 opened this issue Jun 20, 2023 · 1 comment
Open

Whisper finetune #184

Macsim2 opened this issue Jun 20, 2023 · 1 comment

Comments

@Macsim2
Copy link

Macsim2 commented Jun 20, 2023

Hi I'm trying to train whisper fine-tune with multi-gpu
and I don't know what RANK to set
I just set WORLD_SIZE is numer of gpu and MASTER_ADDR is localhost, MASTER_PORT is idle port
When WORLD_SIZE is more than 2 and RANK is set 0, training is hanging
Probably it hanged in setting torch.distributed.TCPStore() part..

anyone who solved this problem?
let me know hint please

@sanchit-gandhi
Copy link
Contributor

Hey @Macsim2! You should just be able to launch multi-GPU training using torchrun, as shown here: https://github.com/huggingface/transformers/tree/main/examples/pytorch/speech-recognition#multi-gpu-whisper-training

Let me know if you encounter any difficulties!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants