summaryrefslogtreecommitdiff
path: root/candle-examples
Commit message (Expand)AuthorAgeFilesLines
* Add a flag to force running the quantized model on CPUs. (#1778)Laurent Mazare2024-02-281-1/+5
* Support more modes in the encodec example. (#1777)Laurent Mazare2024-02-287-641/+159
* Make some dependencies optional in the examples. (#1776)Laurent Mazare2024-02-281-2/+14
* Encodec encoding demo. (#1775)Laurent Mazare2024-02-281-1/+16
* Encodec model. (#1771)Laurent Mazare2024-02-274-0/+114
* Add an option to split the prompt. (#1766)Laurent Mazare2024-02-271-1/+14
* add quantized rwkv v5 model (#1743)Jack Shih2024-02-251-4/+38
* Cuda acceleration for quantized model. (#1754)Laurent Mazare2024-02-251-0/+1
* Fix the eos token for gemma. (#1753)Laurent Mazare2024-02-241-2/+2
* Fix typo in README (#1740)Daniel Varga2024-02-221-1/+1
* Make the cache for the llama model explicit too. (#1745)Laurent Mazare2024-02-221-3/+3
* Explicit caching in llama2.c.laurent2024-02-222-20/+21
* Add the Gemma models. (#1741)Laurent Mazare2024-02-212-0/+281
* Use the tokenizer-output-stream in the llama example. (#1715)Laurent Mazare2024-02-154-20/+17
* Add a readme for rwkv. (#1712)Laurent Mazare2024-02-141-0/+17
* Custom tokenizer for rwkv. (#1711)Laurent Mazare2024-02-141-38/+13
* Add the RWKV model (v5). (#1707)Laurent Mazare2024-02-141-0/+290
* Add ConvNeXt-V2 and smaller model variants. (#1709)Jani Monoses2024-02-142-15/+40
* Detach the tensors on batch-norm eval. (#1702)Laurent Mazare2024-02-132-5/+5
* feat: support microphone whisper streaming (#1678)drbh2024-02-123-0/+816
* Improved mamba model optimized for inference (#1694)Laurent Mazare2024-02-113-0/+319
* Fixing the qwen tokenizer location. (#1693)Nicolas Patry2024-02-111-3/+1
* docs: add trocr examples (#1692)Todsaporn Banjerdkit2024-02-102-2/+11
* Mention TrOCR in the readmes. (#1691)Laurent Mazare2024-02-101-1/+7
* Use the repo config for trocr rather than hardcoding it + small tweaks. (#1689)Laurent Mazare2024-02-101-40/+62
* ChatGLM custom tokenizer. (#1687)Laurent Mazare2024-02-101-1/+3
* Add the custom tokenizer. (#1686)Laurent Mazare2024-02-091-1/+3
* Use the proper endoftext token for gwen. (#1685)Laurent Mazare2024-02-091-2/+2
* Add the Qwen2 model (#1684)Laurent Mazare2024-02-091-0/+281
* Add the ChatGLM model. (#1237)Laurent Mazare2024-02-091-0/+235
* Fix clippy lints for 1.76. (#1682)Laurent Mazare2024-02-081-1/+1
* Fix token generation in bilingual models (non-English outputs) (#1668)Guoqing Bao2024-02-062-1/+2
* Update docs to reflect current usage of example (#1610)Tarek2024-02-041-4/+33
* Quantized support for stable-lm2. (#1654)Laurent Mazare2024-02-042-6/+27
* Add StableLM-2, StableLM Code and Zephyr variants (#1650)Jani Monoses2024-02-032-10/+56
* Supports more audio formats (#1628)Hubert Shelley2024-02-033-12/+81
* Add ConvNeXt model. (#1604)Jani Monoses2024-02-032-0/+124
* Quantized GGUF style (#1523)Nicolas Patry2024-01-179-35/+43
* Add MobileOne model. (#1595)Jani Monoses2024-01-162-0/+118
* Use the new phi model by default. (#1589)Laurent Mazare2024-01-151-26/+29
* Update the Phi model to use the updated architecture. (#1580)Laurent Mazare2024-01-131-11/+35
* Metal: f16 and bf16 where_cond + benchmark (#1545)ivarflakstad2024-01-121-1/+0
* Mention VGG in the readme. (#1573)Laurent Mazare2024-01-121-2/+4
* Pin the revision used for phi-v2 + make it the default. (#1572)Laurent Mazare2024-01-122-10/+3
* Add RepVGG model. (#1561)Jani Monoses2024-01-112-0/+131
* Use bindgen-cuda for the custom-kernel example. (#1536)Laurent Mazare2024-01-074-236/+20
* Simplifying our internal cargo dependencies. (#1529)Nicolas Patry2024-01-071-6/+6
* fix index_pos bug when kv cache is disabled. (#1517)optman2024-01-061-4/+4
* Format properly the Stable Diffusion example run with params (#1511)stano2024-01-011-1/+1
* Do not implement Module for BatchNorm. (#1513)Laurent Mazare2024-01-011-1/+1