index
:
forks/candle.git
main
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
candle-core
/
examples
Commit message (
Expand
)
Author
Age
Files
Lines
*
Add a dequantize command to tensor-tools. (#1565)
Laurent Mazare
2024-01-11
1
-1
/
+24
*
Implement the module trait directly for QMatMul. (#1372)
Laurent Mazare
2023-11-25
1
-6
/
+5
*
Quantized version of mistral. (#1009)
Laurent Mazare
2023-09-30
1
-9
/
+27
*
Use yoke to provide a self-referential container for mmaped safetenso… (#939)
Laurent Mazare
2023-09-23
1
-2
/
+1
*
Use the proper block size for quantizing models. (#933)
Laurent Mazare
2023-09-22
1
-2
/
+17
*
T5 quantized example (#922)
Laurent Mazare
2023-09-21
1
-0
/
+53
*
Add a custom softmax implementation. (#744)
Laurent Mazare
2023-09-05
1
-166
/
+0
*
Dilated convolutions (#657)
Laurent Mazare
2023-08-29
3
-6
/
+6
*
Llama quantization. (#625)
Laurent Mazare
2023-08-27
1
-15
/
+75
*
Add the quantize command. (#624)
Laurent Mazare
2023-08-27
1
-1
/
+75
*
More pickle support. (#588)
Laurent Mazare
2023-08-24
1
-1
/
+1
*
Add to the cuda example a reproduction of the issue. (#579)
Laurent Mazare
2023-08-24
1
-2
/
+11
*
Add a test for conv2d with padding + bugfix the random number generation on c...
Laurent Mazare
2023-08-24
1
-0
/
+3
*
Add some group parameter to convolutions. (#566)
Laurent Mazare
2023-08-23
3
-4
/
+4
*
Handle GGUF files in tensor-tools. (#558)
Laurent Mazare
2023-08-23
1
-1
/
+20
*
Small tweaks to tensor-tools. (#517)
Laurent Mazare
2023-08-19
1
-9
/
+15
*
Retrieve tensor data from PyTorch files. (#516)
Laurent Mazare
2023-08-19
1
-5
/
+7
*
Retrieve more information from PyTorch checkpoints. (#515)
Laurent Mazare
2023-08-19
1
-3
/
+9
*
Add ggml support to tensor-tools (#512)
Laurent Mazare
2023-08-19
1
-15
/
+59
*
Preliminary support for importing PyTorch weights. (#511)
Laurent Mazare
2023-08-19
1
-0
/
+16
*
Add the tensor-tools binary. (#510)
Laurent Mazare
2023-08-19
1
-0
/
+72
*
Tensor -> QTensor conversion (#496)
Laurent Mazare
2023-08-18
1
-1
/
+1
*
AVX version of the vecdot for q4_0. (#474)
Laurent Mazare
2023-08-17
1
-0
/
+24
*
Cudnn support (#445)
Laurent Mazare
2023-08-14
1
-5
/
+4
*
Add a softmax bench. (#433)
Laurent Mazare
2023-08-13
1
-1
/
+29
*
Add a matmul benchmark. (#429)
Laurent Mazare
2023-08-13
1
-0
/
+19
*
More accelerate optimizations (#427)
Laurent Mazare
2023-08-13
2
-0
/
+6
*
Small example for benchmarking some cpu ops (#394)
Laurent Mazare
2023-08-10
2
-24
/
+95
*
Add a conv1d benchmark based on the whisper sizes. (#377)
Laurent Mazare
2023-08-09
1
-0
/
+24
*
Add some conv1d test + bugfix using padding. (#349)
Laurent Mazare
2023-08-08
1
-20
/
+6
*
Support the Accelerate BLAS on macOS. (#325)
Laurent Mazare
2023-08-05
1
-0
/
+3
*
Rename the candle crate to candle-core (#301)
Laurent Mazare
2023-08-02
3
-3
/
+3
*
Simplify Tensor::randn. (#255)
Laurent Mazare
2023-07-27
1
-0
/
+5
*
Simplify the parameters used by sum and sum_keepdim. (#165)
Laurent Mazare
2023-07-14
2
-6
/
+6
*
Use the same default as pytorch for sum. (#164)
Laurent Mazare
2023-07-13
2
-10
/
+10
*
Sketch a fast cuda kernel for reduce-sum. (#109)
Laurent Mazare
2023-07-08
1
-0
/
+15
*
Add some very simple sum benchmark. (#108)
Laurent Mazare
2023-07-08
2
-34
/
+51
*
Add mkl support for matrix multiply. (#86)
Laurent Mazare
2023-07-06
2
-0
/
+6
*
Move llama in a cargo-examples directory.
laurent
2023-07-03
4
-912
/
+0
*
Adding a bit more docs around safety.
Nicolas Patry
2023-07-03
1
-1
/
+1
*
Move more safetensors bits to the shared module.
laurent
2023-07-03
1
-16
/
+8
*
Move some safetensors bits in the candle-core crate.
laurent
2023-07-03
1
-31
/
+2
*
Add a flag for custom prompt.
laurent
2023-07-01
1
-2
/
+7
*
Early conversion for the llama weights.
laurent
2023-06-30
2
-45
/
+19
*
Add a const to easily tweak the dtype used for llama internal computations.
laurent
2023-06-30
1
-4
/
+8
*
Tweak the kv-cache flag.
laurent
2023-06-29
1
-4
/
+4
*
Add a flag.
laurent
2023-06-29
1
-6
/
+11
*
Enable the KV cache after fixing the caching length and the rope bits.
laurent
2023-06-29
1
-14
/
+21
*
Only narrow when needed + deactivate the kv cache.
laurent
2023-06-29
1
-2
/
+6
*
Add some KV cache to llama.
laurent
2023-06-29
1
-36
/
+72
[next]