summaryrefslogtreecommitdiff
path: root/candle-examples/examples/llama_multiprocess
Commit message (Expand)AuthorAgeFilesLines
* Apply rustfmt. (#2421)Laurent Mazare2024-08-161-1/+0
* Fix build issue in EOS Token in llama-multiprocess (#2420)Hadi2024-08-161-2/+11
* Small cleanups to the llama multi-process example. (#2098)Laurent Mazare2024-04-202-68/+42
* Only download the weights in the main process (and not in the child processes...Laurent Mazare2024-04-201-18/+18
* Multiprocess/multi-GPU support for llama 3. (#2092)Laurent Mazare2024-04-202-136/+124
* Use the new hub helper function. (#1484)Laurent Mazare2023-12-261-8/+1
* Self-contained safetensors for the multiprocess llama example. (#950)Laurent Mazare2023-09-241-10/+3
* fix: add missing`top_p` in llama_multiprocess (#905)Mahmoud2023-09-201-1/+5
* VarBuilder cleanup (#627)Laurent Mazare2023-08-272-22/+21
* s/panic/bail/Nicolas Patry2023-08-251-2/+2
* Adding support for codellama in examples.Nicolas Patry2023-08-252-41/+65
* Add a simple Module trait and implement it for the various nn layers (#500)Laurent Mazare2023-08-181-1/+1
* Add an abstract type for RmsNorm. (#499)Laurent Mazare2023-08-182-7/+9
* Relax the requirements on CustomOp. (#486)Laurent Mazare2023-08-171-1/+1
* Layer norm tweaks (#482)Laurent Mazare2023-08-171-39/+6
* Remove the checkpoint conversion script. (#405)Laurent Mazare2023-08-111-3/+0
* Rename the candle crate to candle-core (#301)Laurent Mazare2023-08-021-5/+5
* Remove single function.Nicolas Patry2023-07-281-11/+2
* Making multiprocess require flash-attn.Nicolas Patry2023-07-282-92/+47
* Line-up the llama implementation with the python-transformers one. (#271)Laurent Mazare2023-07-281-1/+1
* Softmax numerical stability. (#267)Laurent Mazare2023-07-281-1/+1
* Putting back Send + SyncNicolas Patry2023-07-271-11/+19
* Removing inner dependency on safetensors.Nicolas Patry2023-07-271-12/+11
* Fixed TP sharded version.Nicolas Patry2023-07-272-34/+60
* TP sharding v2Nicolas Patry2023-07-272-0/+731