index
:
forks/candle.git
main
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
candle-nn
/
src
Commit message (
Expand
)
Author
Age
Files
Lines
*
Support for groups in conv-transpose1d. (#1731)
Laurent Mazare
2024-02-18
1
-3
/
+13
*
Expose the weights and biases in transposed convolutions. (#1727)
Laurent Mazare
2024-02-18
1
-0
/
+16
*
Expose more conv1d functions/structs. (#1726)
Laurent Mazare
2024-02-17
2
-2
/
+19
*
feat: add silu activation function (#1706)
OlivierDehaene
2024-02-14
2
-4
/
+3
*
Detach the tensors on batch-norm eval. (#1702)
Laurent Mazare
2024-02-13
1
-2
/
+12
*
Fix clippy lints for 1.76. (#1682)
Laurent Mazare
2024-02-08
1
-1
/
+1
*
Enhance pickle to retrieve state_dict with a given key (#1671)
Dilshod Tadjibaev
2024-02-06
1
-1
/
+1
*
Add `VarBuilder::from_backend` (#1670)
Daniƫl de Kok
2024-02-06
1
-8
/
+17
*
Update the Phi model to use the updated architecture. (#1580)
Laurent Mazare
2024-01-13
1
-0
/
+1
*
Simplify the one-hot implementation, support arbitrary rank. (#1514)
Laurent Mazare
2024-01-01
1
-181
/
+38
*
Add one-hot/cold encoding (#1489)
Ryan Tate
2024-01-01
2
-0
/
+294
*
Do not implement Module for BatchNorm. (#1513)
Laurent Mazare
2024-01-01
1
-13
/
+13
*
Small tweaks to batch-norm. (#1505)
Laurent Mazare
2023-12-30
1
-19
/
+16
*
[Breaking] Add training to batchnorm with exponential moving average (#1504)
nkoppel
2023-12-30
1
-50
/
+158
*
Merge pull request #1318 from huggingface/metal4
Nicolas Patry
2023-12-20
1
-0
/
+41
|
\
|
*
Clippy pass.
Nicolas Patry
2023-12-18
1
-3
/
+3
|
*
Addressing a lot of comments.
Nicolas Patry
2023-12-15
1
-1
/
+2
|
*
Remove `unwrap()`.
Nicolas Patry
2023-12-15
1
-2
/
+2
|
*
Renamed all kernel names.
Nicolas Patry
2023-12-15
1
-3
/
+3
|
*
Fixing softmax.
Nicolas Patry
2023-12-15
1
-1
/
+1
|
*
Working with merging encoders and using fences.
Nicolas Patry
2023-12-14
1
-2
/
+0
|
*
Lots of updates including some stack of command buffers.
nicolas
2023-12-12
1
-1
/
+3
|
*
Starting to fix some tests.
Nicolas Patry
2023-11-30
1
-0
/
+40
*
|
Fix a couple typos (#1451)
Laurent Mazare
2023-12-17
1
-1
/
+1
*
|
Expose AdamW parameters (#1449)
Dave Lage
2023-12-16
1
-0
/
+8
*
|
Speedup ShardedSafeTensors to load Tensors with default hints (#1384)
YiiSh
2023-12-14
1
-1
/
+7
*
|
Another prelu bugfix. (#1407)
Laurent Mazare
2023-12-06
1
-1
/
+1
*
|
Use the proper broadcasting for prelu. (#1406)
Laurent Mazare
2023-12-05
1
-5
/
+16
*
|
Add the prelu layer. (#1402)
Laurent Mazare
2023-12-03
3
-4
/
+51
|
/
*
Add support to UL2 model family (#1300)
Juarez Bochi
2023-11-09
1
-1
/
+0
*
Add weight and bias functions to LayerNorm (#1306)
jwnz
2023-11-09
1
-0
/
+8
*
Transposed conv1d in candle-nn. (#1252)
Laurent Mazare
2023-11-03
1
-0
/
+94
*
Add the swiglu activation from the chatglm PR. (#1246)
Laurent Mazare
2023-11-02
2
-0
/
+7
*
Add hard-sigmoid and hard-swish activations (#1244)
jamjamjon
2023-11-02
2
-0
/
+9
*
Add support for the marian base model. (#1221)
Laurent Mazare
2023-10-30
1
-0
/
+2
*
Allow for different behavior between training and eval (#1213)
Laurent Mazare
2023-10-29
3
-2
/
+43
*
Add the relu2 and relu6 activations. (#1201)
Laurent Mazare
2023-10-27
1
-0
/
+4
*
Add fuse-conv-bn method for Conv2d (#1196)
jamjamjon
2023-10-27
2
-0
/
+25
*
Expose the fields from batch-norm. (#1176)
Laurent Mazare
2023-10-25
1
-2
/
+12
*
Add Binary Cross Entropy With Logit Loss to nn crate (#1157)
Ogundepo Odunayo
2023-10-23
1
-0
/
+22
*
Make func cloneable. (#1137)
Laurent Mazare
2023-10-20
2
-6
/
+8
*
Add the sequential layer. (#1136)
Laurent Mazare
2023-10-20
2
-0
/
+64
*
Experiment with resnet (#1128)
Laurent Mazare
2023-10-19
1
-0
/
+9
*
feat: add pth varbuilder (#1108)
OlivierDehaene
2023-10-16
1
-0
/
+41
*
Only optimize float tensors. (#1069)
Laurent Mazare
2023-10-10
1
-0
/
+5
*
More general seq forward functions for RNNs. (#1050)
Laurent Mazare
2023-10-07
1
-27
/
+25
*
Use AsRef<str> for set_one. (#1033)
Laurent Mazare
2023-10-05
1
-1
/
+1
*
Bump the version to 0.3.0. (#1014)
Laurent Mazare
2023-10-01
1
-20
/
+0
*
Use a silu activation in mistral. (#991)
Laurent Mazare
2023-09-29
1
-0
/
+4
*
Use the gelu-erf activation. (#969)
Laurent Mazare
2023-09-26
1
-3
/
+1
[next]