Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Train LAVILA (L) to perform action recognition on the EPIC-100 dataset? #9

Open
daiguangzhao opened this issue Feb 17, 2023 · 1 comment

Comments

@daiguangzhao
Copy link

Thank you and your team for bringing such great work. I currently have only one node (8 cards in total), how should I fine-tune the model on the epic-100 dataset? Is the correct script like the one below?

TimeSformer-Large

python run_with_submitit_finetune_classification.py
--pretrain-model $PATH
--use-vn-classifier --num-classes 97 300 3806
--use-sgd --wd 4e-5 --lr-multiplier-on-backbone 0.1
--use-checkpoint --node 1

@zhaoyue-zephyrus
Copy link
Contributor

zhaoyue-zephyrus commented Feb 17, 2023

Hi @daiguangzhao ,

The command you attached should work. To be more close to our setting, you may also try to either (1) add --update-freq 4 OR (2) linearly scale your learning rate by 1/4x, namely --lr 7.5e-4 if you are using --node 1. Note that if your machine is not scheduled by slurm, you can simply use torchrun nproc_per_node=8 main_finetune_classification.py ... to kick off your job.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants