Can we have LLMs loaded on multiple GPUs via the HuggingFacePipeline.from_model_id function? #4965
Unanswered
kiran-paul98
asked this question in
Q&A
Replies: 1 comment
-
Is there anything to be gained in ingest, embeddings or chatting from just on GPU? Are the instructions to get one GPU utilized pretty simple? E.G.
Am I missing anything? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I am trying to load mosaicml's mpt 7b by using the HuggingFacePipeline.from_model_id function using the device parameter. But I am not able to load the entire model into one GPU due to Out of Memory error. I have 2 GPUs. Is it possible to load the model into multiple GPUs? I know that currently device parameter takes only integer arguments. Is there a workaround so that we can pass multiple GPU devices to the function, so that models can be loaded?
Beta Was this translation helpful? Give feedback.
All reactions