Commit message (Expand) | Author | Age | Files | Lines | |
---|---|---|---|---|---|
* | chore: update flash attention kernels (#1518) | OlivierDehaene | 2024-01-05 | 1 | -1/+7 |
* | Add back the bf16 flash-attn kernels. (#730) | Laurent Mazare | 2023-09-04 | 1 | -0/+1 |
* | Flash attention without padding (varlen). (#281) | Laurent Mazare | 2023-07-31 | 1 | -0/+2 |
* | Proper flash-attn parameters. (#244) | Laurent Mazare | 2023-07-26 | 1 | -0/+6 |
* | Add flash attention (#241) | Laurent Mazare | 2023-07-26 | 1 | -0/+35 |