summaryrefslogtreecommitdiff
path: root/candle-examples/examples
Commit message (Collapse)AuthorAgeFilesLines
* Fix the helium weights download. (#2717)Laurent Mazare2025-01-131-1/+1
|
* Helium repo update. (#2716)Laurent Mazare2025-01-132-2/+8
|
* Add the helium model. (#2715)Laurent Mazare2025-01-132-0/+303
|
* Fixes for running Phi-4 quantized. (#2714)Jani Monoses2025-01-131-1/+5
|
* ModernBERT model (#2713)Jani Monoses2025-01-132-0/+192
| | | | | | | | | | | * layer_norm_no_bias * Modernbert model. * Format + cleanup error. --------- Co-authored-by: laurent <laurent.mazare@gmail.com>
* UniPC for diffusion sampling (#2684)Nick Senger2025-01-011-2/+2
| | | | | | | | | | | | | | | | | | | * feat: Add unipc multistep scheduler * chore: Clippy and formatting * chore: Update comments * chore: Avoid unsafety in float ordering * refactor: Update Scheduler::step mutability requirements * fix: Corrector img2img * chore: Update unipc ref link to latest diffusers release * chore: Deduplicate float ordering * fix: Panic when running with dev profile
* Actually remove the default hf-hub cache path for glm. (#2696)Laurent Mazare2024-12-311-1/+1
|
* Use the default hf-hub cache for glm. (#2695)Laurent Mazare2024-12-311-7/+10
|
* Streamline the glm4 example. (#2694)Laurent Mazare2024-12-313-147/+99
|
* Update README.org (#2670)jetsung2024-12-301-1/+1
| | | The command line error in the CPU section of the documentation.
* Added XLMRobertaModel for Reranking (#2686)Akshay Ballal2024-12-302-0/+307
| | | | | | | | | | | | | | | | * add xlm-roberta-base * Add task enum for fill-mask and reranker in xlm-roberta example; update README and fix attention mask dimensions - Introduced a new `Task` enum to replace string task identifiers in the xlm-roberta example. - Updated the logic in `main.rs` to handle tasks using the new enum. - Enhanced README with example output for fill-mask task. - Fixed dimension retrieval in `prepare_4d_attention_mask` function for better clarity and safety. * Clippy fix. --------- Co-authored-by: laurent <laurent.mazare@gmail.com>
* make DepthAnythingV2 more reusable (#2675)Edgar Riba2024-12-211-4/+2
| | | | | | | | | * make DepthAnythingV2 more reusable * Fix clippy lints. --------- Co-authored-by: laurent <laurent.mazare@gmail.com>
* Add Nvembed v2 model (#2649)cdoko2024-12-032-0/+257
| | | | | | | | | | | | | | | | | | | | | | | * Update mod.rs * Create mod.rs * Create decoder.rs * Create model.rs * Create main.rs * Create README.md * Update README.md * Update main.rs * Update and rename decoder.rs to embedding.rs * Update mod.rs * Update model.rs
* Adds support for stella_en_v5 embedding model -400M variant (#2608)iskng2024-11-292-25/+73
| | | | | | | | | | | | | | | | | | | | * Adds support for stella_en_v5 embedding model -400M variant * Unified stella * WIP: Unified Stella * Combined stella for both 1.5B and 400M variants * Cargo fmt for the CI * removed redundant stella-400m model and example after merge into stella-en-v5 * cargo fmt --all --------- Co-authored-by: Anubhab Bandyopadhyay <4890833+AnubhabB@users.noreply.github.com> Co-authored-by: laurent <laurent.mazare@gmail.com>
* Lint fixes introduced with Rust 1.83 (#2646)Anubhab Bandyopadhyay2024-11-281-1/+1
| | | | | | | | | | | * Fixes for lint errors introduced with Rust 1.83 * rustfmt * Fix more lints. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* Fix for whisper-microphone example failure if audio isn't chunk aligned (#2645)Adam Nelson2024-11-271-3/+17
| | | | | | | | | | | | | | | | | | | | | At least on my macOS Sequoia system (MBP 14" 2021, M1 Pro), when I run the `whisper-microphone` example after it has gathered 10 seconds of audio, it fails before the transcription: ``` Error: Insufficient buffer size 384 for input channel 0, expected 1024 ``` At least for the audio device I'm using (Airpods Pro Max), there is no guarantee that each audio buffer is a multiple of 1024 samples. Thus at the end of the 10 seconds, `buffered_pcm` can have some samples at the end that do not form a complete 1024 sample chunk. This fixes that by tracking when there is a partial chunk at the end of the buffer, and leaving it in `buffered_pcm` to be processed on the next loop iteration. Note that, in the interest of keeping this PR as small as possible, I didn't make any other changes to this example.
* Remove some unused macros. (#2618)Laurent Mazare2024-11-155-9/+9
| | | | | * Remove some unused macros. * More unused fixes.
* Add the SmolLM2 models. (#2595)Laurent Mazare2024-11-032-15/+67
| | | | | * Add the SmolLM2 models. * More SmolLM2 support.
* Support Skip Layer Guidance (SLG) for Stable Diffusion 3.5 Medium (#2590)Czxck0012024-11-012-6/+57
| | | | | | | | | | | | | * support skip layer guidance (slg) for stable diffusion 3.5 medium * Tweak the comments formatting. * Proper error message. * Cosmetic tweaks. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* Lazy upcasting for t5. (#2589)Laurent Mazare2024-10-302-31/+11
|
* Support sd3.5 medium and MMDiT-X (#2587)Czxck0012024-10-302-12/+52
| | | | | | | | | | | * extract attn out of joint_attn * further adjust attn and joint_attn * add mmdit-x support * support sd3.5-medium in the example * update README.md
* Reduce memory usage for sd 3.5. (#2582)Laurent Mazare2024-10-281-0/+2
|
* Release the mmdit model earlier to reduce memory usage. (#2581)Laurent Mazare2024-10-281-16/+17
| | | | | | | | | | | | | * Stable diffusion 3.5 support. * Clippy fixes. * CFG fix. * Remove some unnecessary clones. * Avoid duplicating some of the code. * Release the mmdit model earlier to reduce memory usage.
* Stable diffusion 3.5 support. (#2578)Laurent Mazare2024-10-273-84/+166
| | | | | | | | | | | * Stable diffusion 3.5 support. * Clippy fixes. * CFG fix. * Remove some unnecessary clones. * Avoid duplicating some of the code.
* Fix the repo name for llama 3.1. (#2576)Laurent Mazare2024-10-261-2/+2
| | | | | * Fix the repo name for llama 3.1. * Fix the book.
* Enable stable-diffusion 3 on metal. (#2560)Laurent Mazare2024-10-142-7/+10
|
* Adds support for Stella_en_v5 embedding model - 1.5B variant (#2551)Anubhab Bandyopadhyay2024-10-132-0/+404
| | | | | | | | | | | | | * Stella_en_1.5B_v5 * Separated creation. This is a critical step for numerical accuracy and would be documented in the readme * EmbedDim would require clone and copy * WIP: example * Examples added * a litte more in README
* Fix the guide to gain access to Stable Diffusion 3 Medium (#2559)Czxck0012024-10-131-2/+9
|
* Add Stable Diffusion 3 Example (#2558)Czxck0012024-10-136-0/+588
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | * Add stable diffusion 3 example Add get_qkv_linear to handle different dimensionality in linears Add stable diffusion 3 example Add use_quant_conv and use_post_quant_conv for vae in stable diffusion adapt existing AutoEncoderKLConfig to the change add forward_until_encoder_layer to ClipTextTransformer rename sd3 config to sd3_medium in mmdit; minor clean-up Enable flash-attn for mmdit impl when the feature is enabled. Add sd3 example codebase add document crediting references pass the cargo fmt test pass the clippy test * fix typos * expose cfg_scale and time_shift as options * Replace the sample image with JPG version. Change image output format accordingly. * make meaningful error messages * remove the tail-end assignment in sd3_vae_vb_rename * remove the CUDA requirement * use default_value in clap args * add use_flash_attn to turn on/off flash-attn for MMDiT at runtime * resolve clippy errors and warnings * use default_value_t * Pin the web-sys dependency. * Clippy fix. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* feat: intergrate chinese clip and add example (#2555)SethWen2024-10-101-0/+224
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | * start to impl chinese clip * impl vision model * copy code from bert * refactor use * refactor use again * fix text model * refactor * try to fix text model * tuning * tuning chinese clip * delete useless code * revert code * Clippy fixes. * Also apply cargo fmt. --------- Co-authored-by: laurent <laurent.mazare@gmail.com>
* Add BertForMaskedLM to support SPLADE Models (#2550)Akshay Ballal2024-10-072-0/+238
| | | | | | | | | | | | | | | * add bert for masked lm * working example * add example readme * Clippy fix. * And apply rustfmt. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* Add required feature for whisper example in Readme (#2539)dengelt2024-10-041-1/+1
|
* Make the RNN configs accessible from the models. (#2541)Laurent Mazare2024-10-042-2/+0
|
* Support whisper large-v3 turbo in the whisper-microphone example. (#2533)Laurent Mazare2024-10-021-0/+3
|
* Add whisper large-v3 turbo to the example. (#2531)Laurent Mazare2024-10-021-0/+3
|
* Add a seed to the flux example. (#2529)Laurent Mazare2024-10-021-3/+10
|
* Add ColPali (#2524)Akshay Ballal2024-10-012-0/+286
| | | | | | | * add colpali * cleanup * fix clippy
* Refactor the whisper microphone example. (#2523)Laurent Mazare2024-10-011-81/+73
| | | | | * Refactor the whisper microphone example. * Tweak the whisper microphone example more.
* Pixtral polishing. (#2522)Laurent Mazare2024-09-301-12/+3
| | | | | * Pixtral polishing. * Clippy fix.
* Add Pixtral. (#2521)Laurent Mazare2024-09-302-0/+364
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | * Add Pixtral. * More pixtral vision encoder. * Sketch a pixtral example. * Sketch a pixtral example. * Better image loading. * Support loading images embedded in safetensor files. * Clippy fixes. * Add the llava multimodal adapter. * Add more of the llava bits. * Add the pixtral config. * More pixtral inference. * Add the text generation bits. * Get the example to work. * Bugfix. * Run some bits of the model in f32. * Blessed version :) * Better rope frequency computations. * README update.
* Add PaliGemma. (#2519)Laurent Mazare2024-09-292-0/+304
| | | | | | | | | * Add PaliGemma. * PaliGemma inference loop. * Running PaliGemma example. * Tweak the prompt.
* Add the SigLIP model. (#2515)Laurent Mazare2024-09-283-41/+180
| | | | | | | | | | | | | | | | | * Add the SigLIP model. * Add more to the forward pass of the vision model. * Complete the forward pass. * Add the siglip example. * Fix. * Another fix. * Get everything in place. * Add a readme.
* Remove some extra whitelines. (#2513)Laurent Mazare2024-09-281-22/+0
|
* Add some llama-3.2 examples. (#2508)Laurent Mazare2024-09-261-1/+13
| | | | | * Add some llama-3.2 examples. * Support tie-word-embeddings for llama.
* Quantized version of flux. (#2500)Laurent Mazare2024-09-262-20/+65
| | | | | | | | | | | | | * Quantized version of flux. * More generic sampling. * Hook the quantized model. * Use the newly minted gguf file. * Fix for the quantized model. * Default to avoid the faster cuda kernels.
* Add a RotatingKVCache. (#2493)Laurent Mazare2024-09-231-5/+39
| | | | | | | | | | | | | | | | | | | | | | | | | | | * Add a RotatingKVCache. * Add some KvCache tests. * Test the reset too. * More kv-cache testing. * More tests for the rotating kv-cache. * Improve the api for the rotating cache so that the whole src tensor gets returned when it's overlarge. * Handle contiguity + bugfix + use in mimi. * Add a way to test the mimi streaming mode. * Mimi streaming fixes. * More rotating kv-cache. * Fix the attn mask generation. * Handle the abs case. * Add some tests for the generated mask.
* Adding Granite 7b Instruct model example (#2487)Juan Gomez2024-09-212-0/+271
| | | | | | | | | | | | | | | * Adding Granite 7b Instruct model example * Minor refactoring to make it a little more idiomatic * Clippy fixes. * * Adding a README with some information about supported Granite models * Changing the default prompt to accomodate better the Language modality of the Granite 7b Instruct model --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* Add the mimi audio-tokenizer. (#2488)Laurent Mazare2024-09-203-0/+426
| | | | | | | | | | | | | | | * Add the mimi audio-tokenizer. * Formatting tweaks. * Add a full example. * Use the transformers names. * More renamings. * Get encoding and decoding to work. * Clippy fixes.
* Clippy fixes for 1.81.0. (#2461)Laurent Mazare2024-09-054-8/+7
| | | | | * Clippy fixes for 1.81.0. * Another fix.
* Fix FLUX.1 weights (#2457)Eugene Hauptmann2024-08-291-3/+3
| | | | | * fix FLUX.1 weights * added flux1-dev.safetensors