-
-
Notifications
You must be signed in to change notification settings - Fork 44
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CUDA out of memory on a 40G GPU #11
Comments
I also encountered the same problem, hope to receive help |
I faced a similar challenge. Maybe the authors @zhangxiao696 can provide a better solution. |
I change |
@Little-Podi @kashyap7x @YTEP-ZHI, may I ask if you guys have any alternate solution? |
Sorry, we do not have any other useful tips to provide at the moment. Currently, only |
There is actually a way to further reduce memory usage without reducing the quality. The encoder processes all frames in parallel, but this can be changed to process them in sequence instead. After this change the model works fine with less than 20GB memory. Vista/vwm/modules/encoders/modules.py Line 127 in cea9cd9
with https://github.com/rerun-io/hf-example-vista/blob/381b9d574befe0e9a60e9130980d8da0aec5c6ec/vista/vwm/modules/encoders/modules.py#L129-L134 |
I attempted your approach with reduced resolution (64*64), |
Maybe try the fork I linked and see if that works. It works fine for me 25 frames, any number of segments, full resolution. You also have to use the low memory mode if you aren't yet. |
Thanks for replying! |
solve my problem !!! |
Thank you for your sharing, I can successfully use 40G GPU memory for reasoning, but can not train, even if the resolution is 320*576, may I ask how you successfully test training in small GPU memory ? |
@TianDianXin Seems at this time, we can only switch A100 40G to A100 80G😂 |
Hello, everyone! there maybe some batch normalization operations in the encoder, and encode the @Little-Podi need your help😖 |
Hi @SEU-zxj, I think you can do that modification confidently. There are no batchnorm included in the model, thereby encoding the batch sequentially will NOT hurt the performance. |
OK, Thanks for your reply! @YTEP-ZHI |
Following
docs/ISSUES.md
anddocs/SAMPLING.md
set, but still out of memory. Here's my config and instructionIn
configs/inference/vista.yaml
, changeen_and_decode_n_samples_a_time
to1
then run sample with
caught by
Did I miss some useful setting? Any help would be appreciated.
The text was updated successfully, but these errors were encountered: