index
:
forks/candle.git
main
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
candle-flash-attn
/
src
Commit message (
Expand
)
Author
Age
Files
Lines
*
Flash-Attn upgrade / SoftCap Candle-FlashAttn [3/n] (#2690)
Michael Feil
2024-12-31
2
-4
/
+5
*
Flash-Attn upgrade / SoftCap Candle-FlashAttn [2/n] (#2689)
Michael Feil
2024-12-31
2
-0
/
+117
*
Use flash-attn in gemma. (#2195)
Laurent Mazare
2024-05-18
1
-1
/
+3
*
chore: update flash attention kernels (#1518)
OlivierDehaene
2024-01-05
2
-14
/
+428
*
Fix for flash-attn. (#1310)
Laurent Mazare
2023-11-10
1
-2
/
+2
*
Properly set the is_bf16 flag. (#738)
Laurent Mazare
2023-09-04
1
-6
/
+10
*
BF16 support for flash-attn. (#737)
Laurent Mazare
2023-09-04
1
-41
/
+81
*
Add back the bf16 flash-attn kernels. (#730)
Laurent Mazare
2023-09-04
2
-0
/
+3
*
Relax the requirements on CustomOp. (#486)
Laurent Mazare
2023-08-17
1
-2
/
+2
*
Fix the flash-attention function names. (#282)
Laurent Mazare
2023-07-31
1
-2
/
+2
*
Flash attention without padding (varlen). (#281)
Laurent Mazare
2023-07-31
2
-1
/
+232
*
Add some flash attn test (#253)
Laurent Mazare
2023-07-26
1
-10
/
+24
*
Use bail rather than wrapping a string where possible. (#249)
Laurent Mazare
2023-07-26
1
-2
/
+2
*
Lining up the flash attn version with the non-flash one. (#248)
Laurent Mazare
2023-07-26
1
-1
/
+18
*
Again set a few extra params in flash-attn. (#245)
Laurent Mazare
2023-07-26
1
-8
/
+8
*
Proper flash-attn parameters. (#244)
Laurent Mazare
2023-07-26
2
-7
/
+100
*
Add flash attention (#241)
Laurent Mazare
2023-07-26
2
-0
/
+94