summaryrefslogtreecommitdiff
path: root/candle-nn
Commit message (Expand)AuthorAgeFilesLines
...
* feat: add silu activation function (#1706)OlivierDehaene2024-02-142-4/+3
* Detach the tensors on batch-norm eval. (#1702)Laurent Mazare2024-02-131-2/+12
* Fix clippy lints for 1.76. (#1682)Laurent Mazare2024-02-081-1/+1
* Enhance pickle to retrieve state_dict with a given key (#1671)Dilshod Tadjibaev2024-02-061-1/+1
* Add `VarBuilder::from_backend` (#1670)Daniƫl de Kok2024-02-061-8/+17
* Quantized GGUF style (#1523)Nicolas Patry2024-01-171-1/+4
* Update the Phi model to use the updated architecture. (#1580)Laurent Mazare2024-01-131-0/+1
* Simplifying our internal cargo dependencies. (#1529)Nicolas Patry2024-01-071-2/+2
* Simplify the one-hot implementation, support arbitrary rank. (#1514)Laurent Mazare2024-01-011-181/+38
* Add one-hot/cold encoding (#1489)Ryan Tate2024-01-013-0/+414
* Do not implement Module for BatchNorm. (#1513)Laurent Mazare2024-01-012-15/+15
* Small tweaks to batch-norm. (#1505)Laurent Mazare2023-12-301-19/+16
* [Breaking] Add training to batchnorm with exponential moving average (#1504)nkoppel2023-12-302-50/+169
* Bump the crate version to 0.3.3. (#1490)Laurent Mazare2023-12-281-1/+1
* Merge pull request #1318 from huggingface/metal4Nicolas Patry2023-12-202-0/+44
|\
| * Clippy pass.Nicolas Patry2023-12-181-3/+3
| * Addressing a lot of comments.Nicolas Patry2023-12-151-1/+2
| * Remove `unwrap()`.Nicolas Patry2023-12-151-2/+2
| * Renamed all kernel names.Nicolas Patry2023-12-151-3/+3
| * Fixing softmax.Nicolas Patry2023-12-151-1/+1
| * Working with merging encoders and using fences.Nicolas Patry2023-12-141-2/+0
| * Lots of updates including some stack of command buffers.nicolas2023-12-122-2/+5
| * Starting to fix some tests.Nicolas Patry2023-11-302-0/+42
* | Bump the crate version to 0.3.2. (#1452)Laurent Mazare2023-12-171-1/+1
* | Fix a couple typos (#1451)Laurent Mazare2023-12-171-1/+1
* | Expose AdamW parameters (#1449)Dave Lage2023-12-161-0/+8
* | Speedup ShardedSafeTensors to load Tensors with default hints (#1384)YiiSh2023-12-141-1/+7
* | Another prelu bugfix. (#1407)Laurent Mazare2023-12-061-1/+1
* | Use the proper broadcasting for prelu. (#1406)Laurent Mazare2023-12-051-5/+16
* | Add the prelu layer. (#1402)Laurent Mazare2023-12-033-4/+51
|/
* Implement the module trait directly for QMatMul. (#1372)Laurent Mazare2023-11-251-1/+1
* Update for 0.3.1. (#1324)Laurent Mazare2023-11-111-1/+1
* Add support to UL2 model family (#1300)Juarez Bochi2023-11-091-1/+0
* Add weight and bias functions to LayerNorm (#1306)jwnz2023-11-091-0/+8
* Transposed conv1d in candle-nn. (#1252)Laurent Mazare2023-11-031-0/+94
* Add the swiglu activation from the chatglm PR. (#1246)Laurent Mazare2023-11-022-0/+7
* Add hard-sigmoid and hard-swish activations (#1244)jamjamjon2023-11-022-0/+9
* Add support for the marian base model. (#1221)Laurent Mazare2023-10-301-0/+2
* Allow for different behavior between training and eval (#1213)Laurent Mazare2023-10-293-2/+43
* Add the relu2 and relu6 activations. (#1201)Laurent Mazare2023-10-271-0/+4
* Add fuse-conv-bn method for Conv2d (#1196)jamjamjon2023-10-272-0/+25
* Expose the fields from batch-norm. (#1176)Laurent Mazare2023-10-251-2/+12
* Add Binary Cross Entropy With Logit Loss to nn crate (#1157)Ogundepo Odunayo2023-10-232-0/+69
* Make func cloneable. (#1137)Laurent Mazare2023-10-202-6/+8
* Add the sequential layer. (#1136)Laurent Mazare2023-10-202-0/+64
* Experiment with resnet (#1128)Laurent Mazare2023-10-191-0/+9
* feat: add pth varbuilder (#1108)OlivierDehaene2023-10-161-0/+41
* Add a matvec cpu benchmark. (#1076)Laurent Mazare2023-10-121-3/+22
* Convmixer (#1073)Laurent Mazare2023-10-111-2/+2
* Only optimize float tensors. (#1069)Laurent Mazare2023-10-101-0/+5