[package] name = "candle-flash-attn" version = "0.1.0" edition = "2021" description = "Flash attention layer for the candle ML framework." repository = "https://github.com/LaurentMazare/candle" keywords = ["blas", "tensor", "machine-learning"] categories = ["science"] license = "MIT/Apache-2.0" readme = "README.md" [dependencies] candle = { path = "../candle-core", features = ["cuda"] } half = { version = "2.3.1", features = ["num-traits"] } [build-dependencies] anyhow = { version = "1", features = ["backtrace"] } num_cpus = "1.15.0" rayon = "1.7.0" [dev-dependencies] anyhow = { version = "1", features = ["backtrace"] }