summaryrefslogtreecommitdiff
path: root/candle-examples/examples/llama2-c/main.rs
Commit message (Expand)AuthorAgeFilesLines
* Explicit caching in llama2.c.laurent2024-02-221-15/+15
* Use the tokenizer-output-stream in the llama example. (#1715)Laurent Mazare2024-02-151-7/+6
* Quantized GGUF style (#1523)Nicolas Patry2024-01-171-4/+4
* Infer the config for llama2-c. (#1208)Laurent Mazare2023-10-281-2/+12
* Move the llama2-c model in transformers. (#1205)Laurent Mazare2023-10-281-3/+3
* Add a quantized variant of llama2.c (#1197)Laurent Mazare2023-10-271-6/+54
* Implement top_p / nucleus sampling (#819)Juarez Bochi2023-09-121-1/+6
* Add a repeat penality to the llama2-c command line example. (#713)Laurent Mazare2023-09-011-0/+18
* Support the Accelerate BLAS on macOS. (#325)Laurent Mazare2023-08-051-0/+3
* Add the candle-datasets crate (#322)Laurent Mazare2023-08-051-1/+1
* Support safetensors weights in llama2.c inference. (#317)Laurent Mazare2023-08-031-6/+17
* Add training for the llama2.c example (#296)Laurent Mazare2023-08-011-7/+33
* Move the weight bits in a separate module. (#295)Laurent Mazare2023-08-011-161/+6
* Add some batcher variants that handle errors. (#294)Laurent Mazare2023-08-011-4/+4
* Add the batcher. (#293)Laurent Mazare2023-08-011-18/+14
* Use subcommands in llama2. (#292)Laurent Mazare2023-08-011-100/+90
* Pre-tokenized evaluation mode for llama2.c. (#291)Laurent Mazare2023-08-011-30/+51
* Evaluate on the pre-tokenized file. (#290)Laurent Mazare2023-07-311-1/+58
* Remove the end of text tokens. (#289)Laurent Mazare2023-07-311-1/+2
* Add an eval mode to llama2-c (#288)Laurent Mazare2023-07-311-20/+75
* Add a prompt and support more models in llama2-c. (#285)Laurent Mazare2023-07-311-4/+22
* Use the hub models for llama2.c (#284)Laurent Mazare2023-07-311-25/+37
* Use the binary decoder for llama2.c. (#230)Laurent Mazare2023-07-241-57/+74
* Add llama2.c as an example. (#229)Laurent Mazare2023-07-241-0/+240