summaryrefslogtreecommitdiff
path: root/candle-examples/examples/llama2-c
Commit message (Expand)AuthorAgeFilesLines
* Explicit caching in llama2.c.laurent2024-02-222-20/+21
* Use the tokenizer-output-stream in the llama example. (#1715)Laurent Mazare2024-02-151-7/+6
* Quantized GGUF style (#1523)Nicolas Patry2024-01-171-4/+4
* Infer the config for llama2-c. (#1208)Laurent Mazare2023-10-282-3/+13
* Move the llama2-c model in transformers. (#1205)Laurent Mazare2023-10-284-712/+3
* Add a quantized variant of llama2.c (#1197)Laurent Mazare2023-10-273-10/+285
* Implement top_p / nucleus sampling (#819)Juarez Bochi2023-09-121-1/+6
* Add a repeat penality to the llama2-c command line example. (#713)Laurent Mazare2023-09-011-0/+18
* Add the optimizer trait. (#702)Laurent Mazare2023-09-011-0/+1
* Add a simple Module trait and implement it for the various nn layers (#500)Laurent Mazare2023-08-181-1/+1
* Add an abstract type for RmsNorm. (#499)Laurent Mazare2023-08-181-5/+5
* Layer norm tweaks (#482)Laurent Mazare2023-08-171-34/+8
* Support the Accelerate BLAS on macOS. (#325)Laurent Mazare2023-08-051-0/+3
* Add the candle-datasets crate (#322)Laurent Mazare2023-08-052-119/+7
* Transpose the weight matrixes for llama2.c. (#321)Laurent Mazare2023-08-041-8/+15
* Support safetensors weights in llama2.c inference. (#317)Laurent Mazare2023-08-032-7/+18
* Use AdamW in the llama2 training. (#308)Laurent Mazare2023-08-021-2/+9
* Llama more training (#297)Laurent Mazare2023-08-012-18/+26
* Add training for the llama2.c example (#296)Laurent Mazare2023-08-013-7/+216
* Move the weight bits in a separate module. (#295)Laurent Mazare2023-08-013-164/+168
* Add some batcher variants that handle errors. (#294)Laurent Mazare2023-08-011-4/+4
* Add the batcher. (#293)Laurent Mazare2023-08-011-18/+14
* Use subcommands in llama2. (#292)Laurent Mazare2023-08-011-100/+90
* Pre-tokenized evaluation mode for llama2.c. (#291)Laurent Mazare2023-08-011-30/+51
* Evaluate on the pre-tokenized file. (#290)Laurent Mazare2023-07-311-1/+58
* Remove the end of text tokens. (#289)Laurent Mazare2023-07-311-1/+2
* Add an eval mode to llama2-c (#288)Laurent Mazare2023-07-312-35/+87
* Add a prompt and support more models in llama2-c. (#285)Laurent Mazare2023-07-312-6/+26
* Use the hub models for llama2.c (#284)Laurent Mazare2023-07-311-25/+37
* Use u8 tensors for masks. (#273)Laurent Mazare2023-07-291-2/+1
* Softmax numerical stability. (#267)Laurent Mazare2023-07-281-1/+1
* Use the binary decoder for llama2.c. (#230)Laurent Mazare2023-07-242-65/+85
* Add llama2.c as an example. (#229)Laurent Mazare2023-07-242-0/+558