| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* links in chinese_clip
* links for clip model
* add mod docs for flux and llava
* module doc for MMDIT and MIMI
* add docs for a few more modesl
* mod docs for bert naser and beit
* add module docs for convmixer colpali codegeex and chatglm
* add another series of moddocs
* add fastvit-llama2_c
* module docs mamba -> mobileone
* module docs from moondream-phi3
* mod docs for quantized and qwen
* update to yi
* fix long names
* Update llama2_c.rs
* Update llama2_c_weights.rs
* Fix the link for mimi + tweaks
---------
Co-authored-by: Laurent Mazare <laurent.mazare@gmail.com>
|
|
|
|
|
| |
* Clippy fixes for 1.81.0.
* Another fix.
|
| |
|
|
|
|
|
| |
* Add a function to clear the KV cache in falcon.
* Clippy.
|
|
|
|
|
| |
* Allow different dtypes in mamba.
* Add a dtype flag.
|
| |
|
| |
|
|
|
|
|
| |
* Support for attention bias in gemma + refactor things a bit.
* Fix the cuda tests.
|
| |
|
|
* Move some models to candle-transformers so that they can be shared.
* Also move falcon.
* Move Llama.
* Move whisper (partial).
|