Commit message (Expand) | Author | Age | Files | Lines | |
---|---|---|---|---|---|
* | Flash-Attn upgrade / SoftCap Candle-FlashAttn [3/n] (#2690) | Michael Feil | 2024-12-31 | 1 | -0/+2 |
* | Flash-Attn upgrade / SoftCap Candle-FlashAttn [2/n] (#2689) | Michael Feil | 2024-12-31 | 1 | -3/+13 |
* | Update the flash attn kernels. (#2333) | Laurent Mazare | 2024-07-15 | 1 | -10/+10 |
* | chore: update flash attention kernels (#1518) | OlivierDehaene | 2024-01-05 | 1 | -14/+26 |
* | Add back the bf16 flash-attn kernels. (#730) | Laurent Mazare | 2023-09-04 | 1 | -13/+13 |
* | Flash attention without padding (varlen). (#281) | Laurent Mazare | 2023-07-31 | 1 | -3/+6 |
* | Again set a few extra params in flash-attn. (#245) | Laurent Mazare | 2023-07-26 | 1 | -0/+109 |