index
:
forks/candle.git
main
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
candle-nn
/
src
Commit message (
Expand
)
Author
Age
Files
Lines
*
Add fn 'get_with_hints_dtype' in VarBuilder (#1877) (#1897)
yinqiwen
2024-04-01
1
-4
/
+15
*
Fix detail in new RoPE implementation (#1935)
Hugo Abonizio
2024-03-25
1
-1
/
+1
*
Contiguous variant of the rope kernel. (#1929)
Laurent Mazare
2024-03-25
1
-0
/
+252
*
Fast kernels for rotary embeddings. (#1928)
Laurent Mazare
2024-03-24
2
-0
/
+248
*
RmsNorm kernel for metal. (#1895)
Laurent Mazare
2024-03-21
1
-1
/
+46
*
Custom op for RmsNorm (#1890)
Laurent Mazare
2024-03-21
1
-4
/
+167
*
add clone to candle dropout (#1814)
Kirpal Grewal
2024-03-08
1
-1
/
+1
*
Improve metal buffer usage (#1807)
ivarflakstad
2024-03-07
1
-1
/
+2
*
Add the StarCoder2 model. (#1779)
Laurent Mazare
2024-02-28
1
-0
/
+4
*
Encodec model. (#1771)
Laurent Mazare
2024-02-27
1
-1
/
+1
*
Tweak the VarMap set type. (#1758)
Laurent Mazare
2024-02-25
1
-1
/
+1
*
Support for attention bias in gemma + refactor things a bit. (#1744)
Laurent Mazare
2024-02-22
2
-6
/
+19
*
Bugfix for applying the bias in conv1d-transpose. (#1732)
Laurent Mazare
2024-02-18
1
-1
/
+1
*
Support for groups in conv-transpose1d. (#1731)
Laurent Mazare
2024-02-18
1
-3
/
+13
*
Expose the weights and biases in transposed convolutions. (#1727)
Laurent Mazare
2024-02-18
1
-0
/
+16
*
Expose more conv1d functions/structs. (#1726)
Laurent Mazare
2024-02-17
2
-2
/
+19
*
feat: add silu activation function (#1706)
OlivierDehaene
2024-02-14
2
-4
/
+3
*
Detach the tensors on batch-norm eval. (#1702)
Laurent Mazare
2024-02-13
1
-2
/
+12
*
Fix clippy lints for 1.76. (#1682)
Laurent Mazare
2024-02-08
1
-1
/
+1
*
Enhance pickle to retrieve state_dict with a given key (#1671)
Dilshod Tadjibaev
2024-02-06
1
-1
/
+1
*
Add `VarBuilder::from_backend` (#1670)
Daniƫl de Kok
2024-02-06
1
-8
/
+17
*
Update the Phi model to use the updated architecture. (#1580)
Laurent Mazare
2024-01-13
1
-0
/
+1
*
Simplify the one-hot implementation, support arbitrary rank. (#1514)
Laurent Mazare
2024-01-01
1
-181
/
+38
*
Add one-hot/cold encoding (#1489)
Ryan Tate
2024-01-01
2
-0
/
+294
*
Do not implement Module for BatchNorm. (#1513)
Laurent Mazare
2024-01-01
1
-13
/
+13
*
Small tweaks to batch-norm. (#1505)
Laurent Mazare
2023-12-30
1
-19
/
+16
*
[Breaking] Add training to batchnorm with exponential moving average (#1504)
nkoppel
2023-12-30
1
-50
/
+158
*
Merge pull request #1318 from huggingface/metal4
Nicolas Patry
2023-12-20
1
-0
/
+41
|
\
|
*
Clippy pass.
Nicolas Patry
2023-12-18
1
-3
/
+3
|
*
Addressing a lot of comments.
Nicolas Patry
2023-12-15
1
-1
/
+2
|
*
Remove `unwrap()`.
Nicolas Patry
2023-12-15
1
-2
/
+2
|
*
Renamed all kernel names.
Nicolas Patry
2023-12-15
1
-3
/
+3
|
*
Fixing softmax.
Nicolas Patry
2023-12-15
1
-1
/
+1
|
*
Working with merging encoders and using fences.
Nicolas Patry
2023-12-14
1
-2
/
+0
|
*
Lots of updates including some stack of command buffers.
nicolas
2023-12-12
1
-1
/
+3
|
*
Starting to fix some tests.
Nicolas Patry
2023-11-30
1
-0
/
+40
*
|
Fix a couple typos (#1451)
Laurent Mazare
2023-12-17
1
-1
/
+1
*
|
Expose AdamW parameters (#1449)
Dave Lage
2023-12-16
1
-0
/
+8
*
|
Speedup ShardedSafeTensors to load Tensors with default hints (#1384)
YiiSh
2023-12-14
1
-1
/
+7
*
|
Another prelu bugfix. (#1407)
Laurent Mazare
2023-12-06
1
-1
/
+1
*
|
Use the proper broadcasting for prelu. (#1406)
Laurent Mazare
2023-12-05
1
-5
/
+16
*
|
Add the prelu layer. (#1402)
Laurent Mazare
2023-12-03
3
-4
/
+51
|
/
*
Add support to UL2 model family (#1300)
Juarez Bochi
2023-11-09
1
-1
/
+0
*
Add weight and bias functions to LayerNorm (#1306)
jwnz
2023-11-09
1
-0
/
+8
*
Transposed conv1d in candle-nn. (#1252)
Laurent Mazare
2023-11-03
1
-0
/
+94
*
Add the swiglu activation from the chatglm PR. (#1246)
Laurent Mazare
2023-11-02
2
-0
/
+7
*
Add hard-sigmoid and hard-swish activations (#1244)
jamjamjon
2023-11-02
2
-0
/
+9
*
Add support for the marian base model. (#1221)
Laurent Mazare
2023-10-30
1
-0
/
+2
*
Allow for different behavior between training and eval (#1213)
Laurent Mazare
2023-10-29
3
-2
/
+43
*
Add the relu2 and relu6 activations. (#1201)
Laurent Mazare
2023-10-27
1
-0
/
+4
[next]