-
Notifications
You must be signed in to change notification settings - Fork 1k
Pull requests: Dao-AILab/flash-attention
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
remove swizzle part of
sV.data()
to get a completely non-swizzle sVtNoSwizzle
#984
opened Jun 11, 2024 by
soundOfDestiny
Loading…
[Draft] support qk head_dim different from vo head_dim
#980
opened Jun 6, 2024 by
defei-coder
Loading…
Fixing argument checking when using
seqlenq_ngroups_swapped
.
#976
opened Jun 3, 2024 by
Narsil
Loading…
Fix KeyError handling for non-existing key in state_dict.pop()
#898
opened Mar 23, 2024 by
JiedaokouWangguan
Loading…
Add local version identifier to package metadata for pre-built wheels
#856
opened Feb 28, 2024 by
yundai424
Loading…
Animations for Flash Attention, Flash Attention2, and Standard Attention
#736
opened Dec 24, 2023 by
LuisAVasquez
Loading…
feat(attention): add Bi-Directional MLM attention model
#721
opened Dec 12, 2023 by
TamirFriedman-RecoLabs
•
Draft
[fix bug] Llama-2-70B crashed when prompt_len < ngroups
#708
opened Dec 7, 2023 by
li2haipeng
Loading…
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.