summaryrefslogtreecommitdiff
path: root/candle-transformers/src
Commit message (Collapse)AuthorAgeFilesLines
* Add the helium model. (#2715)Laurent Mazare2025-01-132-0/+396
|
* Fixes for running Phi-4 quantized. (#2714)Jani Monoses2025-01-131-1/+1
|
* ModernBERT model (#2713)Jani Monoses2025-01-132-0/+408
| | | | | | | | | | | * layer_norm_no_bias * Modernbert model. * Format + cleanup error. --------- Co-authored-by: laurent <laurent.mazare@gmail.com>
* Fix mistral attention on Metal (#2699)Luka Zakrajšek2025-01-041-1/+2
| | | Co-authored-by: Luka Zakrajsek <luka.zakrajsek@soniox.com>
* UniPC for diffusion sampling (#2684)Nick Senger2025-01-015-3/+1009
| | | | | | | | | | | | | | | | | | | * feat: Add unipc multistep scheduler * chore: Clippy and formatting * chore: Update comments * chore: Avoid unsafety in float ordering * refactor: Update Scheduler::step mutability requirements * fix: Corrector img2img * chore: Update unipc ref link to latest diffusers release * chore: Deduplicate float ordering * fix: Panic when running with dev profile
* Added XLMRobertaModel for Reranking (#2686)Akshay Ballal2024-12-302-0/+546
| | | | | | | | | | | | | | | | * add xlm-roberta-base * Add task enum for fill-mask and reranker in xlm-roberta example; update README and fix attention mask dimensions - Introduced a new `Task` enum to replace string task identifiers in the xlm-roberta example. - Updated the logic in `main.rs` to handle tasks using the new enum. - Enhanced README with example output for fill-mask task. - Fixed dimension retrieval in `prepare_4d_attention_mask` function for better clarity and safety. * Clippy fix. --------- Co-authored-by: laurent <laurent.mazare@gmail.com>
* Fix bug in whisper transformer (#2681)mert-kurttutan2024-12-241-0/+1
| | | | | | | | | | | * Fix bug in whisper transformer - due to num_threads going to zero in single threaded case * Apply rustfmt. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* Fix position encodings for Pixtral (#2678)Amélie Royer2024-12-231-13/+55
| | | | | | | | | * init commit: add position id in meshgrid * pass in subsampled positions * clippy fix * clippy fix
* Add a Context trait similar to anyhow::Context. (#2676)Laurent Mazare2024-12-227-25/+21
| | | | | * Add a Context trait similar to anyhow::Context. * Switch two unwrap to context.
* make DepthAnythingV2 more reusable (#2675)Edgar Riba2024-12-211-19/+25
| | | | | | | | | * make DepthAnythingV2 more reusable * Fix clippy lints. --------- Co-authored-by: laurent <laurent.mazare@gmail.com>
* Change/bert encoder public (#2658)Justin Sing2024-12-041-21/+30
| | | | | | | | | | | | | | | * change: BertEncoder struct to public * change: make certain fields in Config struct public * change: all fields in bert config struct to be public * change: add clone to bert encoder and others * Clippy fix. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* Add Nvembed v2 model (#2649)cdoko2024-12-034-0/+546
| | | | | | | | | | | | | | | | | | | | | | | * Update mod.rs * Create mod.rs * Create decoder.rs * Create model.rs * Create main.rs * Create README.md * Update README.md * Update main.rs * Update and rename decoder.rs to embedding.rs * Update mod.rs * Update model.rs
* Adds support for stella_en_v5 embedding model -400M variant (#2608)iskng2024-11-291-87/+482
| | | | | | | | | | | | | | | | | | | | * Adds support for stella_en_v5 embedding model -400M variant * Unified stella * WIP: Unified Stella * Combined stella for both 1.5B and 400M variants * Cargo fmt for the CI * removed redundant stella-400m model and example after merge into stella-en-v5 * cargo fmt --all --------- Co-authored-by: Anubhab Bandyopadhyay <4890833+AnubhabB@users.noreply.github.com> Co-authored-by: laurent <laurent.mazare@gmail.com>
* Lint fixes introduced with Rust 1.83 (#2646)Anubhab Bandyopadhyay2024-11-286-8/+8
| | | | | | | | | | | * Fixes for lint errors introduced with Rust 1.83 * rustfmt * Fix more lints. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* 20241118 docs (#2629)zachcp2024-11-195-0/+25
| | | | | | | | | | | | | | | | | * module docs * varbuilder gguf docs * add a link to gguf files * small additonal mod doc titles * safetensor docs * more core docs * more module docs in canlde_core * 2 more link fixes
* Fix for clippy. (#2626)Laurent Mazare2024-11-181-1/+1
|
* Module Docs (#2624)zachcp2024-11-1839-115/+170
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | * update whisper * update llama2c * update t5 * update phi and t5 * add a blip model * qlamma doc * add two new docs * add docs and emoji * additional models * openclip * pixtral * edits on the model docs * update yu * update a fe wmore models * add persimmon * add model-level doc * names * update module doc * links in heira * remove empty URL * update more hyperlinks * updated hyperlinks * more links * Update mod.rs --------- Co-authored-by: Laurent Mazare <laurent.mazare@gmail.com>
* More Model Module Docs (#2623)zachcp2024-11-1711-72/+285
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | * dinov2 * add another example * ad dinov2reg4 * eva2 * efficientvit * moondream * update t5 * update t5 * rwkv * stable diffusion docs * add wasm link * add segment_anything * adjsut for clippy * ignore bertdoc * dinov2 ignore * update block to be text * remove the rust blocks for the moment * bump python to 3.11 * add a setup-python step * add py311 to test as well
* Module Docs (#2620)zachcp2024-11-165-10/+126
| | | | | | | | | | | * update bert docs * update based * update bigcode * add pixtral * add flux as well
* Remove some unused macros. (#2618)Laurent Mazare2024-11-152-3/+2
| | | | | * Remove some unused macros. * More unused fixes.
* Documentation Pass for Models (#2617)zachcp2024-11-1594-51/+1001
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | * links in chinese_clip * links for clip model * add mod docs for flux and llava * module doc for MMDIT and MIMI * add docs for a few more modesl * mod docs for bert naser and beit * add module docs for convmixer colpali codegeex and chatglm * add another series of moddocs * add fastvit-llama2_c * module docs mamba -> mobileone * module docs from moondream-phi3 * mod docs for quantized and qwen * update to yi * fix long names * Update llama2_c.rs * Update llama2_c_weights.rs * Fix the link for mimi + tweaks --------- Co-authored-by: Laurent Mazare <laurent.mazare@gmail.com>
* Add some missing index-select metal kernels. (#2613)Laurent Mazare2024-11-121-1/+2
| | | | | * Add some missing index-select metal kernels. * Make some matrix contiguous pre-matmul.
* Add some fast Metal MLX SDPA kernels (#2584)Eric Buehler2024-11-051-14/+20
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | * Add some fast Metal MLX SDPA kernels (#32) * Sketch the sdpa kernel * Add full sdpa kernel, * Add test * Add vectorized kernel for decoding * Update tests * Add some docs * Fix sdpa_vector names * Add softcapping for vectorized sdpa * Add softcapping for full sdpa * Add support for head dim 32, 96, 256 * Add support for head dim 32, 96, 256 * Update docs * Add update notice * Clippy and format * Conditional compilation for bf16 * Use it in quantized llama * Some review comments * Use set_params! * Remove unused * Remove feature * Fix metal sdpa for v stride * Remove comma * Add the dim method to layout and shape. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* Add the SmolLM2 models. (#2595)Laurent Mazare2024-11-031-3/+6
| | | | | * Add the SmolLM2 models. * More SmolLM2 support.
* Support Skip Layer Guidance (SLG) for Stable Diffusion 3.5 Medium (#2590)Czxck0012024-11-011-4/+22
| | | | | | | | | | | | | * support skip layer guidance (slg) for stable diffusion 3.5 medium * Tweak the comments formatting. * Proper error message. * Cosmetic tweaks. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* Lazy upcasting for t5. (#2589)Laurent Mazare2024-10-301-3/+48
|
* Support sd3.5 medium and MMDiT-X (#2587)Czxck0012024-10-302-23/+217
| | | | | | | | | | | * extract attn out of joint_attn * further adjust attn and joint_attn * add mmdit-x support * support sd3.5-medium in the example * update README.md
* Stable diffusion 3.5 support. (#2578)Laurent Mazare2024-10-272-1/+43
| | | | | | | | | | | * Stable diffusion 3.5 support. * Clippy fixes. * CFG fix. * Remove some unnecessary clones. * Avoid duplicating some of the code.
* use softmax_last_dim (metal and cuda kernel) in llama attention layer (#2572)Zack Angelo2024-10-231-1/+2
|
* Enable stable-diffusion 3 on metal. (#2560)Laurent Mazare2024-10-141-2/+1
|
* Adds support for Stella_en_v5 embedding model - 1.5B variant (#2551)Anubhab Bandyopadhyay2024-10-132-0/+400
| | | | | | | | | | | | | * Stella_en_1.5B_v5 * Separated creation. This is a critical step for numerical accuracy and would be documented in the readme * EmbedDim would require clone and copy * WIP: example * Examples added * a litte more in README
* fix: Allow marian configs to deserialize from json. (#2556)Mikarific2024-10-131-1/+2
|
* Add Stable Diffusion 3 Example (#2558)Czxck0012024-10-137-33/+158
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | * Add stable diffusion 3 example Add get_qkv_linear to handle different dimensionality in linears Add stable diffusion 3 example Add use_quant_conv and use_post_quant_conv for vae in stable diffusion adapt existing AutoEncoderKLConfig to the change add forward_until_encoder_layer to ClipTextTransformer rename sd3 config to sd3_medium in mmdit; minor clean-up Enable flash-attn for mmdit impl when the feature is enabled. Add sd3 example codebase add document crediting references pass the cargo fmt test pass the clippy test * fix typos * expose cfg_scale and time_shift as options * Replace the sample image with JPG version. Change image output format accordingly. * make meaningful error messages * remove the tail-end assignment in sd3_vae_vb_rename * remove the CUDA requirement * use default_value in clap args * add use_flash_attn to turn on/off flash-attn for MMDiT at runtime * resolve clippy errors and warnings * use default_value_t * Pin the web-sys dependency. * Clippy fix. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* feat: intergrate chinese clip and add example (#2555)SethWen2024-10-104-0/+1134
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | * start to impl chinese clip * impl vision model * copy code from bert * refactor use * refactor use again * fix text model * refactor * try to fix text model * tuning * tuning chinese clip * delete useless code * revert code * Clippy fixes. * Also apply cargo fmt. --------- Co-authored-by: laurent <laurent.mazare@gmail.com>
* Add BertForMaskedLM to support SPLADE Models (#2550)Akshay Ballal2024-10-071-0/+97
| | | | | | | | | | | | | | | * add bert for masked lm * working example * add example readme * Clippy fix. * And apply rustfmt. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* Add ColPali (#2524)Akshay Ballal2024-10-014-1/+103
| | | | | | | * add colpali * cleanup * fix clippy
* Pixtral polishing. (#2522)Laurent Mazare2024-09-301-0/+26
| | | | | * Pixtral polishing. * Clippy fix.
* Add Pixtral. (#2521)Laurent Mazare2024-09-306-5/+436
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | * Add Pixtral. * More pixtral vision encoder. * Sketch a pixtral example. * Sketch a pixtral example. * Better image loading. * Support loading images embedded in safetensor files. * Clippy fixes. * Add the llava multimodal adapter. * Add more of the llava bits. * Add the pixtral config. * More pixtral inference. * Add the text generation bits. * Get the example to work. * Bugfix. * Run some bits of the model in f32. * Blessed version :) * Better rope frequency computations. * README update.
* Add PaliGemma. (#2519)Laurent Mazare2024-09-293-0/+130
| | | | | | | | | * Add PaliGemma. * PaliGemma inference loop. * Running PaliGemma example. * Tweak the prompt.
* Paligemma siglip vision config (#2518)Laurent Mazare2024-09-291-0/+54
| | | | | * Add the paligemma siglip vision config. * More paligemma configs.
* Add the SigLIP model. (#2515)Laurent Mazare2024-09-285-13/+617
| | | | | | | | | | | | | | | | | * Add the SigLIP model. * Add more to the forward pass of the vision model. * Complete the forward pass. * Add the siglip example. * Fix. * Another fix. * Get everything in place. * Add a readme.
* Remove some extra whitelines. (#2513)Laurent Mazare2024-09-282-5/+0
|
* Add some llama-3.2 examples. (#2508)Laurent Mazare2024-09-262-1/+13
| | | | | * Add some llama-3.2 examples. * Support tie-word-embeddings for llama.
* Quantized version of flux. (#2500)Laurent Mazare2024-09-264-6/+490
| | | | | | | | | | | | | * Quantized version of flux. * More generic sampling. * Hook the quantized model. * Use the newly minted gguf file. * Fix for the quantized model. * Default to avoid the faster cuda kernels.
* Add a RotatingKVCache. (#2493)Laurent Mazare2024-09-231-32/+7
| | | | | | | | | | | | | | | | | | | | | | | | | | | * Add a RotatingKVCache. * Add some KvCache tests. * Test the reset too. * More kv-cache testing. * More tests for the rotating kv-cache. * Improve the api for the rotating cache so that the whole src tensor gets returned when it's overlarge. * Handle contiguity + bugfix + use in mimi. * Add a way to test the mimi streaming mode. * Mimi streaming fixes. * More rotating kv-cache. * Fix the attn mask generation. * Handle the abs case. * Add some tests for the generated mask.
* Adding Granite 7b Instruct model example (#2487)Juan Gomez2024-09-212-0/+459
| | | | | | | | | | | | | | | * Adding Granite 7b Instruct model example * Minor refactoring to make it a little more idiomatic * Clippy fixes. * * Adding a README with some information about supported Granite models * Changing the default prompt to accomodate better the Language modality of the Granite 7b Instruct model --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* Add the mimi audio-tokenizer. (#2488)Laurent Mazare2024-09-207-0/+2593
| | | | | | | | | | | | | | | * Add the mimi audio-tokenizer. * Formatting tweaks. * Add a full example. * Use the transformers names. * More renamings. * Get encoding and decoding to work. * Clippy fixes.
* Clippy fixes for 1.81.0. (#2461)Laurent Mazare2024-09-0511-19/+19
| | | | | * Clippy fixes for 1.81.0. * Another fix.
* MobileCLIP models S1 and S2 (#2454)Jani Monoses2024-08-294-0/+358
| | | | | | | | | | | | | * Allow loading images with given std and mean * OpenCLIP text encoder component * Two MobileCLIP models * Clippy fixes. --------- Co-authored-by: Laurent <laurent.mazare@gmail.com>
* FastViT fixes. (#2452)Jani Monoses2024-08-281-3/+3
| | | | | * correct optional SE layer dimensions. * head_dim instead of num_heads is 32. * update test example output.