You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I have been trying to train the FullSubNet model for a while using the code in this repo. What I experienced is that I must use a batch size of maximum 12, resulting in a very slow and inefficient training (the loss decreases quite slowly). If I try with a larger batch size, I get a GPU out-of-memory message.
I have two Nvidia RTX 2080 Ti with 11 GB each. I see from train.toml that the default batch size is 48, any suggestion on that?
The text was updated successfully, but these errors were encountered:
Hi, I have been trying to train the FullSubNet model for a while using the code in this repo. What I experienced is that I must use a batch size of maximum 12, resulting in a very slow and inefficient training (the loss decreases quite slowly). If I try with a larger batch size, I get a GPU out-of-memory message.
I have two Nvidia RTX 2080 Ti with 11 GB each. I see from
train.toml
that the default batch size is 48, any suggestion on that?The text was updated successfully, but these errors were encountered: