| Commit message (Collapse) | Author | Age | Files | Lines |
... | |
| |
|
| |
|
| |
|
| |
|
|
|
|
|
| |
* Metal part 1 - Scaffolding for metal.
* Remove tracing.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Add support to UL2 model family
* Update docs with UL2
* Create ActivationWithOptionalGating to avoid polluting activations
* Also refactor quantized t5
* Remove useless conversion
* Revert Activation::NewGelu name change
* Remove useless return
* Apply rustfmt and clippy recommendations
* Reuse t5::ActivationWithOptionalGating in quantized version
* (cosmetic change) use a match rather than ifs + avoid early returns.
---------
Co-authored-by: Laurent <laurent.mazare@gmail.com>
|
|
|
|
|
|
|
| |
* Put the onnx example behind a feature flag.
* Exclude the onnx bits from the workspace.
* README tweaks.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Add the onnx protos.
* Move the reading bits.
* Install protoc on the CI.
* Install protoc on the cuda CI too.
* Use clap for the onnx tool.
* Tweak the CI protoc install.
* Add some simple evalution function.
* Add some binary operator support.
|
|
|
|
|
|
|
|
|
|
|
| |
* blip wasm start
* fix dependency issue, move token stream here
* vanilla js worker
* roll back vscode
* spell
|
| |
|
| |
|
|
|
|
|
| |
* Update to gemm-0.16.
* Enable wasm-simd128.
|
|
|
|
|
| |
* Bump the version to 0.3.0.
* Changelog update.
|
|
|
|
|
|
|
|
|
|
|
| |
* Simd128 vec-dot for q4_0.
* Bugfix.
* Add wasm tests.
* Bugfix for the q40 vecdot.
* More quantization tests.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* add phi wasm module
* replace input with textarea
* trim input prompt
* stop on <|endoftext|>
* formatting
* clean up
* add blurb, and syntax highlighting
* add phi-v1.5 wasm
* add note
* hide Options on details
* add first token to generated text
* whitespaces for new line
* fix: abort -> aborted
|
|
|
|
|
|
|
|
|
| |
* Use yoke to provide a self-referential container for mmaped safetensor files.
* Add the new self-owned type for safetensor files without removing the previous version.
* Add routing.
* Add an initializer for the case of multiple files.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* init t5 wasm model
* split workers for each model
* clean up
* add some ui
* readme
* index
* typo
* remove cache param, clear_kv_cache
* add max_length as param
* add model tasks option to ui
* add method to load quantized gguf from buffer
* Add quantized wasm module
* add quantized models to UI, dynamic import wasms
* link to quantized
* fix copy
* fix ModelEncoder
* fix README.md
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* implement wasm module
* add example to workspace
* add UI explore semantic similiarity
* change status messages
* formatting
* minor changes
|
|
|
|
|
| |
* Bump the crate version.
* Also update the python bindings.
|
| |
|
| |
|
| |
|
|
|
|
|
| |
* Add some documentation.
* Bump the crate version.
|
| |
|
|
|
|
|
| |
- Moved around book from `examples` to `candle-book` proper (overlapping
the book and the lib structures)
|
| |
|
|
|
|
|
| |
* Add some configurable legend for yolo detection.
* Clippyness.
|
|
|
|
|
|
|
| |
* Move the yolo model bits in a separate file.
* Improve the drawing.
* Bugfix.
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Add some group parameter to convolutions.
* Avoid some unnecessary groups checks.
* Move the tensor convolution bits.
* Properh handling of groups.
* Bump the crate version.
* And add a changelog.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Sketch the yolo wasm example.
* Web ui.
* Get the web ui to work.
* UI tweaks.
* More UI tweaks.
* Use the natural width/height.
* Add a link to the hf space in the readme.
|
| |
|
|
|
|
|
|
|
| |
* Rename vec-dot to vec-ops.
* Also bump the crate version.
* Add a currently empty readme.
|
|
|
|
|
|
|
|
|
| |
* Import the simd intrinsics in candle-core.
* simd version of reduce-sum.
* Bugfix.
* Fix some clippy lints.
|
|
|
|
|
| |
* Add a cuda kernel for upsampling.
* Update for the latest tokenizers version.
|
|
|
|
|
|
|
|
|
| |
* Parallelise the conv2d op.
* Tighter control on threading.
* Also parallelise conv1d.
* Add some safety comment.
|
| |
|
| |
|
|
|
|
|
|
|
|
|
| |
* Reorder the conv1d loops in the cpu backend.
* Optimize the 1d convolution.
* Conv1D optimize.
* Fix some clippy lints.
|
|
|
|
|
|
|
|
|
|
|
| |
* Change distributions
Standard generates in [0, 1), Normal is correct.
* Add test
Not sure if this is the best place to put the test
* Remove unnecessary use
|
| |
|
|
|
|
|
| |
* Use the image crate to write the generated images.
* Make the dependency optional.
|
| |
|
|
|
|
|
| |
* Add the accelerate feature.
* Ffi tweaks.
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Move the vision datasets to a separate crate.
* Move the batcher bits.
* Update the readme.
* Move the tiny-stories bits.
---------
Co-authored-by: Jane Doe <jane.doe@example.org>
|
| |
|
|
|
|
|
| |
* Switch to candle-gemm for the time being.
* Add the missing versions.
|
| |
|
| |
|
|\
| |
| | |
TP sharding v2
|