| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
|
| |
* Sync upstream mlx sdpa vector kernels with mask
* Dispatch to the 2pass kernel
* Format
|
|
|
|
|
|
|
| |
* Bump the ug dependency.
* Fix some test.
* Fix the ug test.
|
| |
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
| |
* layer_norm_no_bias
* Modernbert model.
* Format + cleanup error.
---------
Co-authored-by: laurent <laurent.mazare@gmail.com>
|
| |
|
| |
|
| |
|
| |
|
|
|
| |
Co-authored-by: Luka Zakrajsek <luka.zakrajsek@soniox.com>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* feat: Add unipc multistep scheduler
* chore: Clippy and formatting
* chore: Update comments
* chore: Avoid unsafety in float ordering
* refactor: Update Scheduler::step mutability requirements
* fix: Corrector img2img
* chore: Update unipc ref link to latest diffusers release
* chore: Deduplicate float ordering
* fix: Panic when running with dev profile
|
|
|
|
|
|
|
| |
* Update the hf-hub dependency to 0.4.0.
* Fix the book.
* Use 0.4.1.
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* update flash-attn v1
* restore: hdim224
* add 224 flash_fwd_template
* remove whitespace
* softcap is working, including test and api.
* make softcap test case better
* unpadded lse added
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* update flash-attn v1
* restore: hdim224
* add 224 flash_fwd_template
* remove whitespace
* softcap is working, including test and api.
* make softcap test case better
---------
Co-authored-by: laurent <laurent.mazare@gmail.com>
|
|
|
|
|
|
|
|
|
| |
* update flash-attn v1
* restore: hdim224
* add 224 flash_fwd_template
* remove whitespace
|
| |
|
| |
|
|
|
| |
The command line error in the CPU section of the documentation.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* add xlm-roberta-base
* Add task enum for fill-mask and reranker in xlm-roberta example; update README and fix attention mask dimensions
- Introduced a new `Task` enum to replace string task identifiers in the xlm-roberta example.
- Updated the logic in `main.rs` to handle tasks using the new enum.
- Enhanced README with example output for fill-mask task.
- Fixed dimension retrieval in `prepare_4d_attention_mask` function for better clarity and safety.
* Clippy fix.
---------
Co-authored-by: laurent <laurent.mazare@gmail.com>
|
|
|
|
|
|
|
|
|
|
|
| |
* Fix bug in whisper transformer
- due to num_threads going to zero
in single threaded case
* Apply rustfmt.
---------
Co-authored-by: Laurent <laurent.mazare@gmail.com>
|
|
|
|
| |
items.is_empty (#2654) (#2655)
|
|
|
|
|
|
|
|
|
| |
* init commit: add position id in meshgrid
* pass in subsampled positions
* clippy fix
* clippy fix
|
|
|
|
|
| |
* Add a Context trait similar to anyhow::Context.
* Switch two unwrap to context.
|
|
|
|
|
|
|
|
|
| |
* make DepthAnythingV2 more reusable
* Fix clippy lints.
---------
Co-authored-by: laurent <laurent.mazare@gmail.com>
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* change: BertEncoder struct to public
* change: make certain fields in Config struct public
* change: all fields in bert config struct to be public
* change: add clone to bert encoder and others
* Clippy fix.
---------
Co-authored-by: Laurent <laurent.mazare@gmail.com>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Update mod.rs
* Create mod.rs
* Create decoder.rs
* Create model.rs
* Create main.rs
* Create README.md
* Update README.md
* Update main.rs
* Update and rename decoder.rs to embedding.rs
* Update mod.rs
* Update model.rs
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Adds support for stella_en_v5 embedding model -400M variant
* Unified stella
* WIP: Unified Stella
* Combined stella for both 1.5B and 400M variants
* Cargo fmt for the CI
* removed redundant stella-400m model and example after merge into stella-en-v5
* cargo fmt --all
---------
Co-authored-by: Anubhab Bandyopadhyay <4890833+AnubhabB@users.noreply.github.com>
Co-authored-by: laurent <laurent.mazare@gmail.com>
|
|
|
|
|
|
|
|
|
|
|
| |
* Fixes for lint errors introduced with Rust 1.83
* rustfmt
* Fix more lints.
---------
Co-authored-by: Laurent <laurent.mazare@gmail.com>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
At least on my macOS Sequoia system (MBP 14" 2021, M1 Pro), when I run
the `whisper-microphone` example after it has gathered 10 seconds of
audio, it fails before the transcription:
```
Error: Insufficient buffer size 384 for input channel 0, expected 1024
```
At least for the audio device I'm using (Airpods Pro Max), there is no
guarantee that each audio buffer is a multiple of 1024 samples. Thus at
the end of the 10 seconds, `buffered_pcm` can have some samples at the
end that do not form a complete 1024 sample chunk.
This fixes that by tracking when there is a partial chunk at the end of
the buffer, and leaving it in `buffered_pcm` to be processed on the next
loop iteration.
Note that, in the interest of keeping this PR as small as possible, I
didn't make any other changes to this example.
|
|
|
|
|
|
|
|
|
| |
* Support for Sign operation #2641
* Apply rustfmt.
---------
Co-authored-by: Laurent <laurent.mazare@gmail.com>
|
|
|
|
|
|
|
| |
* Provide a method to allow PTH files iwth state maps to be loaded.
* add a line to the doc
* String-. &str
|
| |
|
|
|
|
|
| |
* Tweak the CI to avoid running out of disk space.
* Linux only.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* module docs
* varbuilder gguf docs
* add a link to gguf files
* small additonal mod doc titles
* safetensor docs
* more core docs
* more module docs in canlde_core
* 2 more link fixes
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* update whisper
* update llama2c
* update t5
* update phi and t5
* add a blip model
* qlamma doc
* add two new docs
* add docs and emoji
* additional models
* openclip
* pixtral
* edits on the model docs
* update yu
* update a fe wmore models
* add persimmon
* add model-level doc
* names
* update module doc
* links in heira
* remove empty URL
* update more hyperlinks
* updated hyperlinks
* more links
* Update mod.rs
---------
Co-authored-by: Laurent Mazare <laurent.mazare@gmail.com>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* dinov2
* add another example
* ad dinov2reg4
* eva2
* efficientvit
* moondream
* update t5
* update t5
* rwkv
* stable diffusion docs
* add wasm link
* add segment_anything
* adjsut for clippy
* ignore bertdoc
* dinov2 ignore
* update block to be text
* remove the rust blocks for the moment
* bump python to 3.11
* add a setup-python step
* add py311 to test as well
|
|
|
|
|
|
|
|
|
|
|
| |
* update bert docs
* update based
* update bigcode
* add pixtral
* add flux as well
|
|
|
|
|
| |
* Remove some unused macros.
* More unused fixes.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* links in chinese_clip
* links for clip model
* add mod docs for flux and llava
* module doc for MMDIT and MIMI
* add docs for a few more modesl
* mod docs for bert naser and beit
* add module docs for convmixer colpali codegeex and chatglm
* add another series of moddocs
* add fastvit-llama2_c
* module docs mamba -> mobileone
* module docs from moondream-phi3
* mod docs for quantized and qwen
* update to yi
* fix long names
* Update llama2_c.rs
* Update llama2_c_weights.rs
* Fix the link for mimi + tweaks
---------
Co-authored-by: Laurent Mazare <laurent.mazare@gmail.com>
|
| |
|