index
:
forks/candle.git
main
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
candle-examples
/
examples
/
llama
Commit message (
Expand
)
Author
Age
Files
Lines
*
Add the SmolLM2 models. (#2595)
Laurent Mazare
2024-11-03
1
-14
/
+43
*
Fix the repo name for llama 3.1. (#2576)
Laurent Mazare
2024-10-26
1
-2
/
+2
*
Add some llama-3.2 examples. (#2508)
Laurent Mazare
2024-09-26
1
-1
/
+13
*
Add support for Llama 3.1 (#2359)
Eric Buehler
2024-07-26
1
-6
/
+24
*
Support top-k in tthe llama example. (#2150)
Laurent Mazare
2024-05-01
1
-3
/
+21
*
Better time measurement for the llama example. (#2106)
Laurent Mazare
2024-04-22
1
-2
/
+5
*
Use llama v3 by default + add to readme. (#2094)
Laurent Mazare
2024-04-20
1
-1
/
+1
*
Also enable llama-v3 8b instruct. (#2088)
Laurent Mazare
2024-04-19
1
-1
/
+3
*
Llama v3. (#2085)
Laurent Mazare
2024-04-18
1
-9
/
+13
*
Make the cache for the llama model explicit too. (#1745)
Laurent Mazare
2024-02-22
1
-3
/
+3
*
Use the tokenizer-output-stream in the llama example. (#1715)
Laurent Mazare
2024-02-15
1
-11
/
+9
*
fix index_pos bug when kv cache is disabled. (#1517)
optman
2024-01-06
1
-4
/
+4
*
Add support for tiny-llama-1.1b. (#1512)
Laurent Mazare
2023-12-31
1
-2
/
+9
*
Rework the llama example config, add the solar model. (#1485)
Laurent Mazare
2023-12-26
1
-72
/
+36
*
Adapt more examples to the updated safetensor api. (#947)
Laurent Mazare
2023-09-23
1
-9
/
+1
*
Implement top_p / nucleus sampling (#819)
Juarez Bochi
2023-09-12
1
-1
/
+5
*
Move some models to candle-transformers so that it's easier to re-use. (#794)
Laurent Mazare
2023-09-10
2
-448
/
+1
*
Add some optional repeat penalty. (#623)
Laurent Mazare
2023-08-27
1
-0
/
+18
*
s/panic/bail/
Nicolas Patry
2023-08-25
1
-2
/
+2
*
Adding support for codellama in examples.
Nicolas Patry
2023-08-25
2
-6
/
+26
*
GQA support in the quantized model. (#555)
Laurent Mazare
2023-08-22
1
-1
/
+1
*
Add a simple Module trait and implement it for the various nn layers (#500)
Laurent Mazare
2023-08-18
1
-1
/
+1
*
Add an abstract type for RmsNorm. (#499)
Laurent Mazare
2023-08-18
1
-1
/
+1
*
Layer norm tweaks (#482)
Laurent Mazare
2023-08-17
1
-19
/
+4
*
Add some tracing to the quantized example. (#473)
Laurent Mazare
2023-08-16
1
-1
/
+0
*
Fixing llamav1
Nicolas Patry
2023-08-16
1
-2
/
+2
*
Get the ggml based llama to generate some text. (#464)
Laurent Mazare
2023-08-16
1
-5
/
+1
*
Clippy.
Nicolas Patry
2023-08-16
1
-5
/
+5
*
Using the real config from the hub when available.
Nicolas Patry
2023-08-16
2
-43
/
+75
*
Tweak the llama example. (#450)
Laurent Mazare
2023-08-15
1
-63
/
+14
*
Support local weights & dynamic outputs (#447)
Guoqing Bao
2023-08-15
1
-15
/
+39
*
Add a cuda kernel for upsampling. (#441)
Laurent Mazare
2023-08-14
1
-2
/
+2
*
Remove the checkpoint conversion script. (#405)
Laurent Mazare
2023-08-11
2
-202
/
+0
*
Support the Accelerate BLAS on macOS. (#325)
Laurent Mazare
2023-08-05
1
-0
/
+3
*
Add some tracing to llama. (#318)
Laurent Mazare
2023-08-03
2
-4
/
+53
*
Use u8 tensors for masks. (#273)
Laurent Mazare
2023-07-29
1
-2
/
+1
*
Support both llama v1 and llama v2. (#272)
Laurent Mazare
2023-07-28
2
-2
/
+20
*
Line-up the llama implementation with the python-transformers one. (#271)
Laurent Mazare
2023-07-28
1
-43
/
+28
*
Softmax numerical stability. (#267)
Laurent Mazare
2023-07-28
1
-1
/
+1
*
Upgrading hf-hub to `0.2.0` (Modified API to not pass the Repo around
Nicolas Patry
2023-07-27
1
-4
/
+4
*
Switch to using llama-v2 by default. (#251)
Laurent Mazare
2023-07-26
1
-4
/
+4
*
Lining up the flash attn version with the non-flash one. (#248)
Laurent Mazare
2023-07-26
1
-11
/
+10
*
Again set a few extra params in flash-attn. (#245)
Laurent Mazare
2023-07-26
1
-1
/
+5
*
Proper flash-attn parameters. (#244)
Laurent Mazare
2023-07-26
1
-4
/
+12
*
Better handling of dtypes in llama. (#243)
Laurent Mazare
2023-07-26
2
-13
/
+12
*
Add flash attention (#241)
Laurent Mazare
2023-07-26
2
-8
/
+30
*
Rename the .r functions to .dims so as to be a bit more explicit. (#220)
Laurent Mazare
2023-07-22
1
-6
/
+6
*
Support for MQA for llama v2. (#205)
Laurent Mazare
2023-07-20
2
-109
/
+122
*
Removing `candle-hub` internal to extract into `hf-hub` standalone.
Nicolas Patry
2023-07-19
1
-1
/
+1
*
Add some 'cuda-if-available' helper function. (#172)
Laurent Mazare
2023-07-15
1
-14
/
+1
[next]