-
Notifications
You must be signed in to change notification settings - Fork 2.4k
Pull requests: Dao-AILab/flash-attention
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Add SM120 (Blackwell GeForce / DGX Spark) forward pass support
#2268
opened Feb 20, 2026 by
blake-snc
Loading…
2 of 4 tasks
[WIP,Cute,Flex,Sm100] vectorized mask mod application
#2261
opened Feb 17, 2026 by
reubenconducts
•
Draft
[Cute] Handle window_size=(-1, -1) for non-local attention
#2251
opened Feb 11, 2026 by
henrylhtsang
Loading…
Add two-level accumulation for SM90 FP8 FWD to mitigate long-context degradation
#2250
opened Feb 11, 2026 by
jmkuebler
Loading…
[ROCM] Add support with Infinity Cache (LLC) awareness for performance improvement - [PR#2147 rebased on PR#2178]
#2217
opened Jan 29, 2026 by
tianwyan
Loading…
Add shift scheduler for deterministic full‑mask FA3 bwd on Hopper (sm90)
#2207
opened Jan 23, 2026 by
tie-pilot-qxw
Loading…
Fix compute_block_sparsity import in benchmark_mask_mod
#2190
opened Jan 17, 2026 by
blueberrycongee
Loading…
[Cute,Fwd,Sm100] support irregular qhead / kvhead ratios
#2186
opened Jan 16, 2026 by
timmy-feng
•
Draft
Update mha_fwd.cpp, Normalize the commented-out parameters
#2160
opened Jan 9, 2026 by
breakfei
Loading…
Previous Next
ProTip!
Find all pull requests that aren't related to any open issues with -linked:issue.