Skip to content

Pull requests: Dao-AILab/flash-attention

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

Add SM120 (Blackwell GeForce / DGX Spark) forward pass support
#2268 opened Feb 20, 2026 by blake-snc Loading…
2 of 4 tasks
Deterministic backward for blocksparse impl
#2253 opened Feb 12, 2026 by drisspg Loading…
correct max_seqlen_q for performance.
#2252 opened Feb 12, 2026 by lightzhan-intellif Loading…
Fix directory path in README instructions
#2249 opened Feb 11, 2026 by Chuge0335 Loading…
support the flash api for Ascend
#2246 opened Feb 10, 2026 by AnyFree813 Loading…
Add gfx1150/gfx1151 (RDNA 3.5) to RDNA_ARCHS
#2243 opened Feb 9, 2026 by rwfsmith Loading…
[AMD] Migrate to Triton Backend to Aiter
#2230 opened Feb 4, 2026 by micmelesse Loading…
Nicer headdim error message
#2227 opened Feb 4, 2026 by drisspg Loading…
[Ai-assisted] CLC work stealing
#2218 opened Jan 31, 2026 by drisspg Loading…
Add loc info & Fix api changes for CuTeDSL 4.4
#2204 opened Jan 23, 2026 by keithzzzzz Loading…
[Cute, SM100] Fix comment in tmem_p_offset
#2201 opened Jan 22, 2026 by Edenzzzz Loading…
Warn when ninja is missing
#2191 opened Jan 17, 2026 by blueberrycongee Loading…
[Cute][Testing] Protyping a fast test mode for Cute
#2188 opened Jan 16, 2026 by drisspg Loading…
[Cute] Add torch.compile support for FA4
#2164 opened Jan 9, 2026 by gilfordting Loading…
ProTip! Find all pull requests that aren't related to any open issues with -linked:issue.