Commit message (Expand) | Author | Age | Files | Lines | |
---|---|---|---|---|---|
* | Update the flash attn kernels. (#2333) | Laurent Mazare | 2024-07-15 | 49 | -898/+2257 |
* | Use flash-attn in gemma. (#2195) | Laurent Mazare | 2024-05-18 | 1 | -0/+4 |
* | chore: update flash attention kernels (#1518) | OlivierDehaene | 2024-01-05 | 26 | -451/+658 |
* | Add back the bf16 flash-attn kernels. (#730) | Laurent Mazare | 2023-09-04 | 1 | -13/+13 |
* | Flash attention without padding (varlen). (#281) | Laurent Mazare | 2023-07-31 | 1 | -3/+6 |
* | Again set a few extra params in flash-attn. (#245) | Laurent Mazare | 2023-07-26 | 17 | -91/+379 |
* | Proper flash-attn parameters. (#244) | Laurent Mazare | 2023-07-26 | 1 | -1/+22 |
* | Add flash attention (#241) | Laurent Mazare | 2023-07-26 | 10 | -0/+2361 |