summaryrefslogtreecommitdiff
path: root/README.md
Commit message (Collapse)AuthorAgeFilesLines
* add link to README (#2701)Andrei Fajardo2025-01-041-0/+1
|
* Update README.md (#2577)sashaphmn2024-10-261-1/+2
|
* improve (#2548)Jorge António2024-10-071-0/+1
|
* FastViT fixes. (#2452)Jani Monoses2024-08-281-1/+1
| | | | | * correct optional SE layer dimensions. * head_dim instead of num_heads is 32. * update test example output.
* Add a readme for the parler-tts example. (#2434)Laurent Mazare2024-08-191-0/+3
| | | | | | | | | * Add a readme for the parler-tts example. * Remove the python decode script. * mp4 tweaks. * Another readme tweak.
* Add support for gemma-2. (#2425)Laurent Mazare2024-08-171-2/+2
| | | | | | | | | | | * Add gemma-2. * Support a couple more models. * Sliding window support. * Example + readme updates. * Update the main readme.
* add models support and example for THUDM/glm-4 (#2362)唐璜2024-08-051-0/+2
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | * add models support and example for THUDM/glm-4 * fix the ci report * fmt * fix * Update README.org * Update README.org * fmt * Update README.org * README.md add codegeex4 * README.md add glm4 * Typo. * change expect into ? --------- Co-authored-by: Laurent Mazare <laurent.mazare@gmail.com>
* Add Hiera vision model. (#2382)Jani Monoses2024-08-011-1/+1
|
* Add Mobilenet v4 (#2325)Jani Monoses2024-07-091-1/+1
| | | | | | | * Support different resolutions in load_image() * Added MobilenetV4 model. * Add MobileNetv4 to README
* chore: add fix for windows cudarc into the readme (#2189)Martin Stefcek2024-05-161-0/+7
|
* chore: fix some typos in comments (#2121)hardlydearly2024-04-281-1/+1
| | | Signed-off-by: hardlydearly <799511800@qq.com>
* Mention phi-v3 in the readmes. (#2122)Laurent Mazare2024-04-241-2/+3
|
* Use llama v3 by default + add to readme. (#2094)Laurent Mazare2024-04-201-2/+2
|
* Minor fix to the readme. (#2080)Laurent Mazare2024-04-171-2/+2
| | | Co-authored-by: Jane Doe <jane.doe@example.org>
* Add a quantized version of recurrent-gemma. (#2054)Laurent Mazare2024-04-131-2/+3
| | | | | | | * Add a quantized version of recurrent-gemma. * Share the rglru part. * Get the quantized gemma model to work.
* Added link to the Coursera ML algorithm implementations (#1989)Vishal Patil2024-04-031-0/+1
| | | | | * Added link to the coursera ML algo implementations * Fixed link
* Add Moondream transformer implementation and example (#1970)Santiago Medina2024-03-311-0/+2
| | | | | | | | | | | | | | | | | | | * moondream implementation * add moondream example * change config default activation * Add assets and integrate phi mixformer with example * Make use of kv cache and fix seq_len bug; Clean up example code * Add README link to example * Remove pos_embed scaling; Remove assets; Add to README; Expand VisionConfig * Delete image * Use apply instead of forward
* Qwen MoE model. (#1960)Laurent Mazare2024-03-281-1/+3
| | | | | | | | | | | * Qwen MoE model. * Add the MoE model to the example. * Fix the scaling. * Readme updates. * Readme tweaks.
* Update README.md (#1840)Thomas Santerre2024-03-131-0/+1
| | | Adds the candle-einops to the readme as an external resource
* Add the new models to the main readme. (#1797)Laurent Mazare2024-03-031-3/+8
|
* Mention rwkv v6 in the readmes. (#1784)Laurent Mazare2024-03-011-2/+2
|
* EfficientVit (MSRA) model (#1783)Jani Monoses2024-03-011-1/+1
| | | | | * Add EfficientVit (Microsoft Research Asia) model. * Mention models in README
* Add the StarCoder2 model. (#1779)Laurent Mazare2024-02-281-2/+3
| | | | | | | * Add the StarCoder2 model. * Add the example code and get things to work. * And also tweak the readme.
* Add a flag to force running the quantized model on CPUs. (#1778)Laurent Mazare2024-02-281-1/+5
| | | | | * Add a flag to force running the quantized model on CPUs. * Add encodec to the readme.
* Add the Gemma models. (#1741)Laurent Mazare2024-02-211-0/+3
| | | | | | | | | | | | | | | | | | | | | * Add the Gemma models. * Add the gemma example. * Adapt the RmsNorm. * Get the 2b model to work. * 7b support. * Use the config head dim. * Yet another fix. * Make the matrixes contiguous. * Also get the 7b model to work. * And add to the readme.
* Add the RWKV model (v5). (#1707)Laurent Mazare2024-02-141-1/+7
| | | | | | | | | | | | | | | | | | | | | | | * Start adding the RWKV model. * More of the forward step. * Handle rescaling. * FeedForward. * More work on RWKV. * Better state tracking. * Finish a first pass on forward. * Fix the shape mismatches. * Do not rescale in f32. * Rename to rwkv-v5. * Add the new models to the readme.
* Improved mamba model optimized for inference (#1694)Laurent Mazare2024-02-111-2/+2
| | | | | | | | | | | | | | | * Sketch the mamba model for inference. * Complete the forward pass. * Add the mamba example. * Optimize the selective-scan part. * Fix a couple shape mismatches and get inference to work. * Tweak the readmes. * More readme tweaks.
* Mention TrOCR in the readmes. (#1691)Laurent Mazare2024-02-101-1/+3
|
* Mention the new models in the readme. (#1651)Laurent Mazare2024-02-031-3/+4
|
* Fix "Minimal Mamba" link in README. (#1577)SebastianRueClausen2024-01-121-1/+1
|
* Mention VGG in the readme. (#1573)Laurent Mazare2024-01-121-1/+4
|
* add link to gpt-from-scratch-rs (#1525)Jeroen Vlek2024-01-051-0/+1
|
* Add some mention to SOLAR-10.7B in the readme. (#1487)Laurent Mazare2023-12-271-2/+3
|
* Support different mamba models. (#1471)Laurent Mazare2023-12-231-0/+3
|
* Fix a couple typos (#1451)Laurent Mazare2023-12-171-2/+4
| | | | | * Mixtral quantized instruct. * Fix a couple typos.
* Readme tweaks. (#1446)Laurent Mazare2023-12-161-3/+6
|
* Update the readme to mention mixtral. (#1443)Laurent Mazare2023-12-151-0/+2
|
* Mention phi-2 in the readme. (#1434)Laurent Mazare2023-12-141-4/+4
|
* Stable Diffusion Turbo Support (#1395)Edwin Cheng2023-12-031-1/+1
| | | | | | | | | | | * Add support for SD Turbo * Set Leading as default in euler_ancestral discrete * Use the appropriate default values for n_steps and guidance_scale. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* Add candle-lora transformers to readme? (#1356)Eric Buehler2023-11-211-5/+5
| | | | | * Demonstrate lora transformers in readme * Shorten readme
* Update the readme. (#1354)Laurent Mazare2023-11-211-0/+5
|
* Fix the kalosm link (#1353)Laurent Mazare2023-11-211-1/+1
|
* Mention the Yi-6b/Yi-34b models in the readme. (#1321)Laurent Mazare2023-11-111-0/+3
|
* Add support to UL2 model family (#1300)Juarez Bochi2023-11-091-1/+1
| | | | | | | | | | | | | | | | | | | | | | | | | * Add support to UL2 model family * Update docs with UL2 * Create ActivationWithOptionalGating to avoid polluting activations * Also refactor quantized t5 * Remove useless conversion * Revert Activation::NewGelu name change * Remove useless return * Apply rustfmt and clippy recommendations * Reuse t5::ActivationWithOptionalGating in quantized version * (cosmetic change) use a match rather than ifs + avoid early returns. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* Update README: Move T5 to Text to Text section (#1288)Juarez Bochi2023-11-071-3/+3
| | | I think it makes more sense to have it there, since it's a seq2seq model with cross attention, and not a LM. There are also Decoder only T5 models that work as LMs, but that's not the standard.
* Add info about MADLAD-400 in readme files (#1287)Juarez Bochi2023-11-071-1/+1
|
* Add candle-sampling (#1278)Eric Buehler2023-11-061-1/+2
|
* Add a link to candle-ext to README.md (#1277)YiiSh2023-11-061-0/+1
|
* Put the onnx example behind a feature flag. (#1276)Laurent Mazare2023-11-061-19/+1
| | | | | | | * Put the onnx example behind a feature flag. * Exclude the onnx bits from the workspace. * README tweaks.
* Added info about missing protoc (#1275)figgefigge2023-11-061-0/+19
| | | Co-authored-by: figgefigge <fredric.1337mail.com>