Has anyone tried training the chat model with LLAMA-7B? #3230
alibabadoufu
started this conversation in
General
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I am wondering how much GPU memory needed for training the LLAMA-7B
My own experiment:
2 x V100 32GB running the LLAMA-7B model using lora implementation, I experienced the out of CUDA memory issue
Beta Was this translation helpful? Give feedback.
All reactions