Skip to content

Pull requests: Dao-AILab/flash-attention

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

Support AMD ROCm on FlashAttention 2
#1010 opened Jun 26, 2024 by rocking5566 Loading…
Add return_softmax_lse in flash_attn_with_kvcache
#989 opened Jun 13, 2024 by ovowei Loading…
Fix +/-inf in LSE returned by forward
#978 opened Jun 3, 2024 by sgrigory Loading…
fix typo
#974 opened May 31, 2024 by jslhcl Loading…
add pyproject.toml with build dependencies
#958 opened May 17, 2024 by dhellmann Loading…
Relative position encoding
#956 opened May 14, 2024 by b-albar Loading…
1 of 4 tasks
Add softmax_d in mha_bwd
#905 opened Apr 1, 2024 by MayDomine Loading…
ALiBi for the non-flash code path
#858 opened Feb 29, 2024 by Markus28 Loading…
Fix typos of statement about shape.
#837 opened Feb 19, 2024 by 66RING Loading…
Add support for small page sizes
#824 opened Feb 13, 2024 by skrider Loading…
Add C++ build support for use with LibTorch
#819 opened Feb 9, 2024 by shaltielshmid Loading…
meta tensor stuff
#769 opened Jan 15, 2024 by tsengalb99 Loading…
Jetson (aarch64) support
#724 opened Dec 14, 2023 by jasl Loading…
Update utils.py
#710 opened Dec 8, 2023 by adarshxs Loading…
Add flash_attn_varlen_func_with_kvcache.
#685 opened Nov 22, 2023 by garrett4wade Loading…
Custom attention bias
#617 opened Oct 19, 2023 by b-albar Loading…
2 of 5 tasks
Setup: Add extra compute targets
#605 opened Oct 15, 2023 by bdashore3 Loading…
ProTip! Exclude everything labeled bug with -label:bug.