Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

There was no increase in speed after installing flash-attn and xformer. #59

Open
kongwanbianjinyu opened this issue May 16, 2024 · 0 comments

Comments

@kongwanbianjinyu
Copy link

Hello,

I installed the flash-attn and xformer packages to accelerate the training process, but noticed no speed improvement compared to slow_attn. I'm using 4 RTX 6000 GPUs to train VAR-d16 for 200 epochs, with a global_batch_size of 96 due to memory constraints. The whole training process would take roughly 1000 hours. Could you kindly share your hardware settings and the time required to train VAR-d16?

Best,
Jiachen

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant