summaryrefslogtreecommitdiff
path: root/candle-flash-attn/tests
Commit message (Collapse)AuthorAgeFilesLines
* Flash-Attn upgrade / SoftCap Candle-FlashAttn [2/n] (#2689)Michael Feil2024-12-311-0/+52
| | | | | | | | | | | | | | | | | * update flash-attn v1 * restore: hdim224 * add 224 flash_fwd_template * remove whitespace * softcap is working, including test and api. * make softcap test case better --------- Co-authored-by: laurent <laurent.mazare@gmail.com>
* Flash attention without padding (varlen). (#281)Laurent Mazare2023-07-311-0/+45
| | | | | | | | | | | | | * Expose the seqlen variable for flash-attn without padding. * Fix the batched call. * Adapt for the varlen variant. * No need to set the batch strides when in varlen mode. * Add a test (disabled at the moment). * Get the test to work properly.
* Softmax numerical stability. (#267)Laurent Mazare2023-07-281-1/+1
| | | | | * Softmax numerical stability. * Fix the flash-attn test.
* Add some flash attn test (#253)Laurent Mazare2023-07-261-0/+90
* Add some flash-attn test. * Add the cpu test. * Fail when the head is not a multiple of 8. * Polish the flash attention test.