Skip to content

No support for multi-GPU #106

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
parviste-fortum opened this issue Oct 13, 2023 · 3 comments
Open

No support for multi-GPU #106

parviste-fortum opened this issue Oct 13, 2023 · 3 comments

Comments

@parviste-fortum
Copy link

It seems that it's not possible to run models using multiple gpus, e.g. by passing device_map="auto" to pipelines.

Is there any way to work around this limitation?

@philschmid
Copy link
Collaborator

What model are you trying to run? We recommend for LLMs to use the LLM container

@parviste-fortum
Copy link
Author

Yes, I moved over to the TGI container. I had started with the generic container since it's what is used by https://registry.terraform.io/modules/philschmid/sagemaker-huggingface/aws/latest. The lack of multi-gpu support was just quite surprising, especially since it doesn't really seem to be any specific reason for not having it (I suppose no one has just implemented it).

@MoritzLaurer
Copy link

MoritzLaurer commented Jul 12, 2024

I think this blog post should respond to your question @parviste-fortum : https://www.philschmid.de/sagemaker-multi-replica

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants