summaryrefslogtreecommitdiff
path: root/candle-transformers/src/models/falcon.rs
Commit message (Collapse)AuthorAgeFilesLines
* Documentation Pass for Models (#2617)zachcp2024-11-151-0/+6
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | * links in chinese_clip * links for clip model * add mod docs for flux and llava * module doc for MMDIT and MIMI * add docs for a few more modesl * mod docs for bert naser and beit * add module docs for convmixer colpali codegeex and chatglm * add another series of moddocs * add fastvit-llama2_c * module docs mamba -> mobileone * module docs from moondream-phi3 * mod docs for quantized and qwen * update to yi * fix long names * Update llama2_c.rs * Update llama2_c_weights.rs * Fix the link for mimi + tweaks --------- Co-authored-by: Laurent Mazare <laurent.mazare@gmail.com>
* Clippy fixes for 1.81.0. (#2461)Laurent Mazare2024-09-051-1/+1
| | | | | * Clippy fixes for 1.81.0. * Another fix.
* Make the falcon model cloneable. (#2067)Laurent Mazare2024-04-151-5/+5
|
* Add a function to clear the KV cache in falcon. (#2066)Laurent Mazare2024-04-151-0/+14
| | | | | * Add a function to clear the KV cache in falcon. * Clippy.
* Support alternative dtypes for mamba (#2036)Laurent Mazare2024-04-101-1/+3
| | | | | * Allow different dtypes in mamba. * Add a dtype flag.
* first commit (#1994)Jorge António2024-04-021-1/+2
|
* Avoid the attention mask where possible. (#1933)Laurent Mazare2024-03-251-12/+20
|
* Support for attention bias in gemma + refactor things a bit. (#1744)Laurent Mazare2024-02-221-11/+1
| | | | | * Support for attention bias in gemma + refactor things a bit. * Fix the cuda tests.
* Use candle_nn::embedding instead of local copies in a few models. (#1562)Jani Monoses2024-01-101-6/+1
|
* Move some models to candle-transformers so that it's easier to re-use. (#794)Laurent Mazare2023-09-101-0/+484
* Move some models to candle-transformers so that they can be shared. * Also move falcon. * Move Llama. * Move whisper (partial).