Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

how to translate with multiple instances of the model using multiple GPUs #5

Open
StephennFernandes opened this issue Feb 15, 2024 · 1 comment

Comments

@StephennFernandes
Copy link

hey @pacman100

the translation code could take days given there are multiple SFT datasets and multiple languages to translate on.

is there a way to accelerate the code by launching multiple instances of the model on the same GPU, and launching the same on multiple GPUs such that it can collectively speed up the translation task ?

@StephennFernandes
Copy link
Author

found a partial solution to the problem here: huggingface/datasets#6186
still i believe there is a better way to populate all the vram on the single GPU until gpu is full by creating multiple instances of the same model on the same GPU multiple times, and further repeating the same on all available GPUs in the server to maximize flops

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant