[package] name = "candle-flash-attn" version = "0.8.1" edition = "2021" description = "Flash attention layer for the candle ML framework." repository = "https://github.com/huggingface/candle" keywords = ["blas", "tensor", "machine-learning"] categories = ["science"] license = "MIT OR Apache-2.0" readme = "README.md" [dependencies] candle = { path = "../candle-core", features = ["cuda"], package = "candle-core", version = "0.8.1" } half = { version = "2.3.1", features = ["num-traits"] } [build-dependencies] bindgen_cuda = "0.1.1" anyhow = { version = "1", features = ["backtrace"] } [dev-dependencies] anyhow = { version = "1", features = ["backtrace"] } candle-nn = { path = "../candle-nn", features = ["cuda"] }