summaryrefslogtreecommitdiff
path: root/candle-core/examples/cpu_benchmarks.rs
diff options
context:
space:
mode:
authorLaurent Mazare <laurent.mazare@gmail.com>2023-08-16 12:41:07 +0100
committerGitHub <noreply@github.com>2023-08-16 12:41:07 +0100
commit3071134788334c972d9e356f53887d2b2ff026b7 (patch)
treeadbb58e3babee5d62fa6150bde4f9bb03770607c /candle-core/examples/cpu_benchmarks.rs
parentfec87e86f50da78656a0fb28fc254390435fb3fd (diff)
downloadcandle-3071134788334c972d9e356f53887d2b2ff026b7.tar.gz
candle-3071134788334c972d9e356f53887d2b2ff026b7.tar.bz2
candle-3071134788334c972d9e356f53887d2b2ff026b7.zip
Get the ggml based llama to generate some text. (#464)
* Add more stats to the ggml example. * Build a quantized model from the file content. * Move the tensor retrieval in the main crate. * Start adding the forward pass. * Add more to the forward pass of the quantized llama. * Apply the attention layers. * Add the sampling loop. * Get the sampling loop to work. * Minor tweak. * Add a quantize/dequantize test. * Bugfix. * Add a comment + swap the order. * Bugfixes.
Diffstat (limited to 'candle-core/examples/cpu_benchmarks.rs')
0 files changed, 0 insertions, 0 deletions