Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

packed_data问题 #208

Open
WhyDwelledOnAi opened this issue Dec 13, 2024 · 1 comment
Open

packed_data问题 #208

WhyDwelledOnAi opened this issue Dec 13, 2024 · 1 comment

Comments

@WhyDwelledOnAi
Copy link

使用generate_packed_dataset.py后的packed_data训练时,训练会卡在accessory/engine_pretrain.py 的metric_logger.synchronize_between_processes()不动,然后ddp超时结束。
在使用*.parquet文件时则没有问题。

[E ProcessGroupNCCL.cpp:828] [Rank 1] Watchdog caught collective operation timeout: WorkNCCL(SeqNum=495104, OpType=_ALLGATHER_BASE, Timeout(ms)=1800000) ran for 1805715 milliseconds before timing out.
[E ProcessGroupNCCL.cpp:455] Some NCCL operations have failed or timed out. Due to the asynchronous nature of CUDA kernels, subsequent GPU operations might run on corrupted/incomplete data.
[E ProcessGroupNCCL.cpp:460] To avoid data inconsistency, we are taking the entire process down.
terminate called after throwing an instance of 'std::runtime_error'

环境完全遵循文档中的requirement.txt.

@WhyDwelledOnAi
Copy link
Author

WhyDwelledOnAi commented Dec 13, 2024

在accessory/util/misc.py中修改SmoothedValue类的synchronize_between_processes函数。

rank = torch.distributed.get_rank(group=None)
t = torch.tensor([self.count, self.total], dtype=torch.float64, device=f'cuda:{rank}')

doesn't work

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant