index
:
forks/candle.git
main
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
candle-nn
/
tests
Commit message (
Expand
)
Author
Age
Files
Lines
*
Add some fast Metal MLX SDPA kernels (#2584)
Eric Buehler
2024-11-05
1
-0
/
+206
*
Improved launch config for layer-norm/rms-norm. (#2591)
Laurent Mazare
2024-11-04
1
-0
/
+45
*
Add a RotatingKVCache. (#2493)
Laurent Mazare
2024-09-23
1
-0
/
+110
*
Enable the new layer-norm. (#2213)
Laurent Mazare
2024-05-24
1
-3
/
+12
*
Add the layernorm specialized op. (#2212)
Laurent Mazare
2024-05-24
1
-0
/
+27
*
Bug Fix: When converting a tensor to a variable, clone if the tensor is alrea...
Jeffrey Dallatezza
2024-04-29
1
-2
/
+44
*
Fix sigmoid gradient calculation and move sigmoid into a specialized op (#2114)
MilkFather
2024-04-29
1
-0
/
+11
*
Add the rope THD kernel. (#2014)
Laurent Mazare
2024-04-05
1
-0
/
+31
*
Contiguous variant of the rope kernel. (#1929)
Laurent Mazare
2024-03-25
1
-2
/
+30
*
Fast kernels for rotary embeddings. (#1928)
Laurent Mazare
2024-03-24
1
-0
/
+28
*
Custom op for RmsNorm (#1890)
Laurent Mazare
2024-03-21
1
-4
/
+30
*
Tweak the VarMap set type. (#1758)
Laurent Mazare
2024-02-25
1
-1
/
+38
*
Add one-hot/cold encoding (#1489)
Ryan Tate
2024-01-01
1
-0
/
+120
*
Do not implement Module for BatchNorm. (#1513)
Laurent Mazare
2024-01-01
1
-2
/
+2
*
[Breaking] Add training to batchnorm with exponential moving average (#1504)
nkoppel
2023-12-30
1
-0
/
+11
*
Add Binary Cross Entropy With Logit Loss to nn crate (#1157)
Ogundepo Odunayo
2023-10-23
1
-0
/
+47
*
Add a custom softmax implementation. (#744)
Laurent Mazare
2023-09-05
1
-0
/
+10
*
Avoid some redundant clone. (#731)
Laurent Mazare
2023-09-04
1
-2
/
+2
*
Add the optimizer trait. (#702)
Laurent Mazare
2023-09-01
1
-3
/
+3
*
Add a GRU layer. (#688)
Laurent Mazare
2023-08-31
1
-0
/
+44
*
Add a python variant for the lstm test. (#682)
Laurent Mazare
2023-08-30
1
-0
/
+15
*
Add a LSTM test. (#681)
Laurent Mazare
2023-08-30
1
-0
/
+42
*
Move the test-utils bits to a shared place. (#619)
Laurent Mazare
2023-08-27
7
-62
/
+13
*
Some fixes for yolo-v3. (#529)
Laurent Mazare
2023-08-20
1
-5
/
+9
*
Add a yolo-v3 example. (#528)
Laurent Mazare
2023-08-20
6
-6
/
+29
*
Add a batch normalization layer (#508)
Laurent Mazare
2023-08-18
1
-0
/
+70
*
Add a simple Module trait and implement it for the various nn layers (#500)
Laurent Mazare
2023-08-18
3
-3
/
+3
*
Fix the tests for mkl. (#437)
Laurent Mazare
2023-08-14
1
-2
/
+4
*
Fixes for the stable diffusion example. (#342)
Laurent Mazare
2023-08-08
1
-2
/
+2
*
Implement group-norm. (#334)
Laurent Mazare
2023-08-07
1
-0
/
+103
*
Add the AdamW optimizer. (#307)
Laurent Mazare
2023-08-02
3
-15
/
+100
*
Llama more training (#297)
Laurent Mazare
2023-08-01
1
-2
/
+2
*
Add the cross-entropy loss. (#287)
Laurent Mazare
2023-07-31
1
-1
/
+4
*
Make the nll op closer to the pytorch version + add a test. (#286)
Laurent Mazare
2023-07-31
1
-0
/
+31
*
Softmax numerical stability. (#267)
Laurent Mazare
2023-07-28
1
-0
/
+62
*
Simplify the parameters used by sum and sum_keepdim. (#165)
Laurent Mazare
2023-07-14
1
-2
/
+2
*
Use the same default as pytorch for sum. (#164)
Laurent Mazare
2023-07-13
1
-2
/
+2
*
Add the pytorch version of the linear regression as a comment. (#163)
Laurent Mazare
2023-07-13
1
-0
/
+24
*
Add the gradient for reduce-sum. (#162)
Laurent Mazare
2023-07-13
1
-2
/
+26
*
Add the SGD optimizer (#160)
Laurent Mazare
2023-07-13
1
-0
/
+19
*
Add some layer-norm tests. (#121)
Laurent Mazare
2023-07-10
1
-0
/
+43