index
:
forks/candle.git
main
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
candle-core
/
examples
Commit message (
Expand
)
Author
Age
Files
Lines
*
Add a basic metal example with capture (#2324)
Laurent Mazare
2024-07-09
1
-0
/
+28
*
Fix the fast bf16 gemm cublas kernels. (#2274)
Laurent Mazare
2024-06-18
1
-1
/
+4
*
Make it possible to use TF32 accumulation in F32 matmuls. (#2178)
Laurent Mazare
2024-05-11
1
-24
/
+18
*
Move the tensor-tools binary in a separate crate. (#1969)
Laurent Mazare
2024-03-30
1
-513
/
+0
*
Add a print command to tensor-tools. (#1967)
Laurent Mazare
2024-03-30
1
-0
/
+124
*
Cuda kernel for dequantizing q8k. (#1760)
Laurent Mazare
2024-02-26
1
-4
/
+4
*
Cuda acceleration for quantized model. (#1754)
Laurent Mazare
2024-02-25
1
-16
/
+23
*
Enhance pickle to retrieve state_dict with a given key (#1671)
Dilshod Tadjibaev
2024-02-06
1
-1
/
+1
*
Quantized GGUF style (#1523)
Nicolas Patry
2024-01-17
1
-70
/
+52
*
Add a dequantize command to tensor-tools. (#1565)
Laurent Mazare
2024-01-11
1
-1
/
+24
*
Implement the module trait directly for QMatMul. (#1372)
Laurent Mazare
2023-11-25
1
-6
/
+5
*
Quantized version of mistral. (#1009)
Laurent Mazare
2023-09-30
1
-9
/
+27
*
Use yoke to provide a self-referential container for mmaped safetenso… (#939)
Laurent Mazare
2023-09-23
1
-2
/
+1
*
Use the proper block size for quantizing models. (#933)
Laurent Mazare
2023-09-22
1
-2
/
+17
*
T5 quantized example (#922)
Laurent Mazare
2023-09-21
1
-0
/
+53
*
Add a custom softmax implementation. (#744)
Laurent Mazare
2023-09-05
1
-166
/
+0
*
Dilated convolutions (#657)
Laurent Mazare
2023-08-29
3
-6
/
+6
*
Llama quantization. (#625)
Laurent Mazare
2023-08-27
1
-15
/
+75
*
Add the quantize command. (#624)
Laurent Mazare
2023-08-27
1
-1
/
+75
*
More pickle support. (#588)
Laurent Mazare
2023-08-24
1
-1
/
+1
*
Add to the cuda example a reproduction of the issue. (#579)
Laurent Mazare
2023-08-24
1
-2
/
+11
*
Add a test for conv2d with padding + bugfix the random number generation on c...
Laurent Mazare
2023-08-24
1
-0
/
+3
*
Add some group parameter to convolutions. (#566)
Laurent Mazare
2023-08-23
3
-4
/
+4
*
Handle GGUF files in tensor-tools. (#558)
Laurent Mazare
2023-08-23
1
-1
/
+20
*
Small tweaks to tensor-tools. (#517)
Laurent Mazare
2023-08-19
1
-9
/
+15
*
Retrieve tensor data from PyTorch files. (#516)
Laurent Mazare
2023-08-19
1
-5
/
+7
*
Retrieve more information from PyTorch checkpoints. (#515)
Laurent Mazare
2023-08-19
1
-3
/
+9
*
Add ggml support to tensor-tools (#512)
Laurent Mazare
2023-08-19
1
-15
/
+59
*
Preliminary support for importing PyTorch weights. (#511)
Laurent Mazare
2023-08-19
1
-0
/
+16
*
Add the tensor-tools binary. (#510)
Laurent Mazare
2023-08-19
1
-0
/
+72
*
Tensor -> QTensor conversion (#496)
Laurent Mazare
2023-08-18
1
-1
/
+1
*
AVX version of the vecdot for q4_0. (#474)
Laurent Mazare
2023-08-17
1
-0
/
+24
*
Cudnn support (#445)
Laurent Mazare
2023-08-14
1
-5
/
+4
*
Add a softmax bench. (#433)
Laurent Mazare
2023-08-13
1
-1
/
+29
*
Add a matmul benchmark. (#429)
Laurent Mazare
2023-08-13
1
-0
/
+19
*
More accelerate optimizations (#427)
Laurent Mazare
2023-08-13
2
-0
/
+6
*
Small example for benchmarking some cpu ops (#394)
Laurent Mazare
2023-08-10
2
-24
/
+95
*
Add a conv1d benchmark based on the whisper sizes. (#377)
Laurent Mazare
2023-08-09
1
-0
/
+24
*
Add some conv1d test + bugfix using padding. (#349)
Laurent Mazare
2023-08-08
1
-20
/
+6
*
Support the Accelerate BLAS on macOS. (#325)
Laurent Mazare
2023-08-05
1
-0
/
+3
*
Rename the candle crate to candle-core (#301)
Laurent Mazare
2023-08-02
3
-3
/
+3
*
Simplify Tensor::randn. (#255)
Laurent Mazare
2023-07-27
1
-0
/
+5
*
Simplify the parameters used by sum and sum_keepdim. (#165)
Laurent Mazare
2023-07-14
2
-6
/
+6
*
Use the same default as pytorch for sum. (#164)
Laurent Mazare
2023-07-13
2
-10
/
+10
*
Sketch a fast cuda kernel for reduce-sum. (#109)
Laurent Mazare
2023-07-08
1
-0
/
+15
*
Add some very simple sum benchmark. (#108)
Laurent Mazare
2023-07-08
2
-34
/
+51
*
Add mkl support for matrix multiply. (#86)
Laurent Mazare
2023-07-06
2
-0
/
+6
*
Move llama in a cargo-examples directory.
laurent
2023-07-03
4
-912
/
+0
*
Adding a bit more docs around safety.
Nicolas Patry
2023-07-03
1
-1
/
+1
*
Move more safetensors bits to the shared module.
laurent
2023-07-03
1
-16
/
+8
[next]