summaryrefslogtreecommitdiff
path: root/candle-wasm-examples/llama2-c/llama2cWorker.js
diff options
context:
space:
mode:
authorRadamés Ajna <radamajna@gmail.com>2023-09-26 22:01:59 -0700
committerGitHub <noreply@github.com>2023-09-27 06:01:59 +0100
commit9571b200c9a7c835f66a5444e62f8100e99a4102 (patch)
treefb1e9077e3190264aac388b4aee9cc23e76cfbac /candle-wasm-examples/llama2-c/llama2cWorker.js
parentce0a4e3a85c40f2b46ce2ee5f58ab56c30f38d99 (diff)
downloadcandle-9571b200c9a7c835f66a5444e62f8100e99a4102.tar.gz
candle-9571b200c9a7c835f66a5444e62f8100e99a4102.tar.bz2
candle-9571b200c9a7c835f66a5444e62f8100e99a4102.zip
fix firstToken, minor ui changes (#971)
Diffstat (limited to 'candle-wasm-examples/llama2-c/llama2cWorker.js')
-rw-r--r--candle-wasm-examples/llama2-c/llama2cWorker.js11
1 files changed, 9 insertions, 2 deletions
diff --git a/candle-wasm-examples/llama2-c/llama2cWorker.js b/candle-wasm-examples/llama2-c/llama2cWorker.js
index abaf3401..a46b5bc8 100644
--- a/candle-wasm-examples/llama2-c/llama2cWorker.js
+++ b/candle-wasm-examples/llama2-c/llama2cWorker.js
@@ -50,6 +50,7 @@ async function generate(data) {
tokenizerURL,
prompt,
temp,
+ top_p,
repeatPenalty,
seed,
maxSeqLen,
@@ -59,11 +60,17 @@ async function generate(data) {
const model = await Llama2C.getInstance(weightsURL, modelID, tokenizerURL);
self.postMessage({ status: "loading", message: "Initializing model" });
- model.init_with_prompt(prompt, temp, repeatPenalty, seed);
+ const firstToken = model.init_with_prompt(
+ prompt,
+ temp,
+ top_p,
+ repeatPenalty,
+ seed
+ );
const seq_len = model.get_seq_len();
- let sentence = "";
+ let sentence = firstToken;
let maxTokens = maxSeqLen ? maxSeqLen : seq_len - prompt.length - 1;
let startTime = performance.now();
let tokensCount = 0;