You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I inscreased batchsize by refill the shape of query(Instances in code) in different batches by 0 or -1, which made it possible to train in batchsize>1 during pre process, and keep the post process batch by batch. But I met the problem that loss is large, grad_norm is small and the loss descents slowly. I wonder if anyone else tried to achieve larger batchsize, and did you met the same case?
The text was updated successfully, but these errors were encountered:
DengYizhe
changed the title
I tried to inscrease batchsize
I tried to inscrease batchsize but met some difficulties
Dec 23, 2024
I inscreased batchsize by refill the shape of query(Instances in code) in different batches by 0 or -1, which made it possible to train in batchsize>1 during pre process, and keep the post process batch by batch. But I met the problem that loss is large, grad_norm is small and the loss descents slowly. I wonder if anyone else tried to achieve larger batchsize, and did you met the same case?
The text was updated successfully, but these errors were encountered: