-
Notifications
You must be signed in to change notification settings - Fork 56
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
multi gpu parallel computing #16
Comments
Hi, @ysm022 , firstly you can reduce your batch size and train on single GPU, secondly, if you want to use multi-GPUs training, you need to modify the following configuration:
|
Hi, |
hi have you solve this problem?
|
I run to paddle github searching for multi-gpu sample. I have tested this mnist-project link code:
|
Hello, I can run train.py with very little dataset. 6 pics as train input, including 3 real and 3 fake. 2 pics as val. But when I use big dataset to train, there are total number of data: 11071 | pos: 5705, neg: 5366, total number of data: 1231 | pos: 634, neg: 597.
I get error as follow:
Error Message Summary:
ResourceExhaustedError:
Out of memory error on GPU 0. Cannot allocate 1.158715GB memory on GPU 0, available memory is only 199.500000MB.
Please check whether there is any other process using GPU 0.
at (/paddle/paddle/fluid/memory/allocation/cuda_allocator.cc:69)
I use a nvidia 1080ti to train, the memory is about 11G. The error is ResourceExhaustedError.
I have 4 pieces 1080ti. So how can I do multi gpu parallel computing?
Thank you!
The text was updated successfully, but these errors were encountered: