summaryrefslogtreecommitdiff
path: root/candle-nn/src
Commit message (Expand)AuthorAgeFilesLines
* Add fn 'get_with_hints_dtype' in VarBuilder (#1877) (#1897)yinqiwen2024-04-011-4/+15
* Fix detail in new RoPE implementation (#1935)Hugo Abonizio2024-03-251-1/+1
* Contiguous variant of the rope kernel. (#1929)Laurent Mazare2024-03-251-0/+252
* Fast kernels for rotary embeddings. (#1928)Laurent Mazare2024-03-242-0/+248
* RmsNorm kernel for metal. (#1895)Laurent Mazare2024-03-211-1/+46
* Custom op for RmsNorm (#1890)Laurent Mazare2024-03-211-4/+167
* add clone to candle dropout (#1814)Kirpal Grewal2024-03-081-1/+1
* Improve metal buffer usage (#1807)ivarflakstad2024-03-071-1/+2
* Add the StarCoder2 model. (#1779)Laurent Mazare2024-02-281-0/+4
* Encodec model. (#1771)Laurent Mazare2024-02-271-1/+1
* Tweak the VarMap set type. (#1758)Laurent Mazare2024-02-251-1/+1
* Support for attention bias in gemma + refactor things a bit. (#1744)Laurent Mazare2024-02-222-6/+19
* Bugfix for applying the bias in conv1d-transpose. (#1732)Laurent Mazare2024-02-181-1/+1
* Support for groups in conv-transpose1d. (#1731)Laurent Mazare2024-02-181-3/+13
* Expose the weights and biases in transposed convolutions. (#1727)Laurent Mazare2024-02-181-0/+16
* Expose more conv1d functions/structs. (#1726)Laurent Mazare2024-02-172-2/+19
* feat: add silu activation function (#1706)OlivierDehaene2024-02-142-4/+3
* Detach the tensors on batch-norm eval. (#1702)Laurent Mazare2024-02-131-2/+12
* Fix clippy lints for 1.76. (#1682)Laurent Mazare2024-02-081-1/+1
* Enhance pickle to retrieve state_dict with a given key (#1671)Dilshod Tadjibaev2024-02-061-1/+1
* Add `VarBuilder::from_backend` (#1670)Daniƫl de Kok2024-02-061-8/+17
* Update the Phi model to use the updated architecture. (#1580)Laurent Mazare2024-01-131-0/+1
* Simplify the one-hot implementation, support arbitrary rank. (#1514)Laurent Mazare2024-01-011-181/+38
* Add one-hot/cold encoding (#1489)Ryan Tate2024-01-012-0/+294
* Do not implement Module for BatchNorm. (#1513)Laurent Mazare2024-01-011-13/+13
* Small tweaks to batch-norm. (#1505)Laurent Mazare2023-12-301-19/+16
* [Breaking] Add training to batchnorm with exponential moving average (#1504)nkoppel2023-12-301-50/+158
* Merge pull request #1318 from huggingface/metal4Nicolas Patry2023-12-201-0/+41
|\
| * Clippy pass.Nicolas Patry2023-12-181-3/+3
| * Addressing a lot of comments.Nicolas Patry2023-12-151-1/+2
| * Remove `unwrap()`.Nicolas Patry2023-12-151-2/+2
| * Renamed all kernel names.Nicolas Patry2023-12-151-3/+3
| * Fixing softmax.Nicolas Patry2023-12-151-1/+1
| * Working with merging encoders and using fences.Nicolas Patry2023-12-141-2/+0
| * Lots of updates including some stack of command buffers.nicolas2023-12-121-1/+3
| * Starting to fix some tests.Nicolas Patry2023-11-301-0/+40
* | Fix a couple typos (#1451)Laurent Mazare2023-12-171-1/+1
* | Expose AdamW parameters (#1449)Dave Lage2023-12-161-0/+8
* | Speedup ShardedSafeTensors to load Tensors with default hints (#1384)YiiSh2023-12-141-1/+7
* | Another prelu bugfix. (#1407)Laurent Mazare2023-12-061-1/+1
* | Use the proper broadcasting for prelu. (#1406)Laurent Mazare2023-12-051-5/+16
* | Add the prelu layer. (#1402)Laurent Mazare2023-12-033-4/+51
|/
* Add support to UL2 model family (#1300)Juarez Bochi2023-11-091-1/+0
* Add weight and bias functions to LayerNorm (#1306)jwnz2023-11-091-0/+8
* Transposed conv1d in candle-nn. (#1252)Laurent Mazare2023-11-031-0/+94
* Add the swiglu activation from the chatglm PR. (#1246)Laurent Mazare2023-11-022-0/+7
* Add hard-sigmoid and hard-swish activations (#1244)jamjamjon2023-11-022-0/+9
* Add support for the marian base model. (#1221)Laurent Mazare2023-10-301-0/+2
* Allow for different behavior between training and eval (#1213)Laurent Mazare2023-10-293-2/+43
* Add the relu2 and relu6 activations. (#1201)Laurent Mazare2023-10-271-0/+4