You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I attempted evaluations using the llava-hf and llama-vision models but encountered out-of-memory (OOM) errors.
For the llava-hf model, the OOM error occurred when using the VQAv2 dataset, while it did not occur when using the MMMU dataset.
For the llama-vision model, OOM errors occurred with all datasets, even when utilizing nine 40GB GPUs. Applying device_map=auto did not resolve the issue.
Is there a way to perform evaluations without encountering OOM errors?
Hiii! You may try to start the task directly using lmms-eval, because if you haven't run accelerate config, accelerate will, by default, load one model on each GPU:
I attempted evaluations using the llava-hf and llama-vision models but encountered out-of-memory (OOM) errors.
For the llava-hf model, the OOM error occurred when using the VQAv2 dataset, while it did not occur when using the MMMU dataset.
For the llama-vision model, OOM errors occurred with all datasets, even when utilizing nine 40GB GPUs. Applying device_map=auto did not resolve the issue.
Is there a way to perform evaluations without encountering OOM errors?
Below is the shell script I used for reference.
python3 -m accelerate.commands.launch \ --num_processes=9 \ -m lmms_eval \ --model llava_hf \ --model_args pretrained="llava-hf/llava-v1.6-vicuna-7b-hf" \ --tasks vqav2 \ --batch_size 1 \ --log_samples \ --log_samples_suffix llava_v1.5_vqav2 \ --output_path ./logs/
python3 -m accelerate.commands.launch \ --num_processes=9 \ -m lmms_eval \ --model llama_vision \ --model_args pretrained="meta-llama/Llama-3.2-11B-Vision-Instruct",device_map=auto \ --tasks mmmu \ --batch_size 1 \ --log_samples \ --log_samples_suffix llama_11b_mmmu \ --output_path ./logs/ \
The text was updated successfully, but these errors were encountered: