-
Notifications
You must be signed in to change notification settings - Fork 1.3k
Issues: Dao-AILab/flash-attention
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
In unit test,how is the dropout_fraction diff tolerance selected?
#1286
opened Oct 18, 2024 by
muoshuosha
FlashAttention installation error: "CUDA 11.6 and above" requirement issue
#1282
opened Oct 17, 2024 by
21X5122
Unable to import my new kernel function after compilation success.
#1278
opened Oct 15, 2024 by
jpli02
Why does the flash_attn_varlen_func method increase GPU memory usage?
#1277
opened Oct 15, 2024 by
shaonan1993
Is there a way to install flash-attention without specific cuda version ?
#1276
opened Oct 14, 2024 by
HuangChiEn
Concurrent Warp Group Execution in FA3: Tensor Core Resource Limitation?
#1275
opened Oct 13, 2024 by
ziyuhuang123
How to use the function of flash-attn-1 to mimic the behavior of flash_attn_func in flash-attn-2?
#1270
opened Oct 11, 2024 by
jpWang
Unable to compile for MI300X (gfx942) with ROCm 6.2.2 due to getCurrentHIPStream().stream();
#1269
opened Oct 10, 2024 by
lhl
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.