summaryrefslogtreecommitdiff
path: root/candle-wasm-examples/llama2-c
diff options
context:
space:
mode:
authorLaurent Mazare <laurent.mazare@gmail.com>2024-09-05 22:46:55 +0100
committerGitHub <noreply@github.com>2024-09-05 23:46:55 +0200
commite3261216b157a7305c18ccdd766b6e2a41afe483 (patch)
treeedec85bf6faae0c6662e833f74d79ccc2579ef05 /candle-wasm-examples/llama2-c
parentc02b7c327297f25fc9cc60b8f39c9aa32e40ff21 (diff)
downloadcandle-e3261216b157a7305c18ccdd766b6e2a41afe483.tar.gz
candle-e3261216b157a7305c18ccdd766b6e2a41afe483.tar.bz2
candle-e3261216b157a7305c18ccdd766b6e2a41afe483.zip
Clippy fixes for 1.81.0. (#2461)
* Clippy fixes for 1.81.0. * Another fix.
Diffstat (limited to 'candle-wasm-examples/llama2-c')
-rw-r--r--candle-wasm-examples/llama2-c/src/model.rs2
1 files changed, 1 insertions, 1 deletions
diff --git a/candle-wasm-examples/llama2-c/src/model.rs b/candle-wasm-examples/llama2-c/src/model.rs
index ab9333d2..01f5910a 100644
--- a/candle-wasm-examples/llama2-c/src/model.rs
+++ b/candle-wasm-examples/llama2-c/src/model.rs
@@ -275,7 +275,7 @@ impl Llama {
let lm_head = linear(cfg.dim, cfg.vocab_size, vb.pp("lm_head"))?;
let norm = rms_norm(cfg.dim, cfg.norm_eps, vb.pp("model.norm"))?;
let blocks: Vec<_> = (0..cfg.n_layers)
- .map(|i| Block::load(vb.pp(&format!("model.layers.{i}")), cache, cfg).unwrap())
+ .map(|i| Block::load(vb.pp(format!("model.layers.{i}")), cache, cfg).unwrap())
.collect();
Ok(Self::new(wte, blocks, norm, lm_head))
}