diff options
author | Radamés Ajna <radamajna@gmail.com> | 2023-09-26 22:01:59 -0700 |
---|---|---|
committer | GitHub <noreply@github.com> | 2023-09-27 06:01:59 +0100 |
commit | 9571b200c9a7c835f66a5444e62f8100e99a4102 (patch) | |
tree | fb1e9077e3190264aac388b4aee9cc23e76cfbac /candle-wasm-examples/llama2-c/llama2cWorker.js | |
parent | ce0a4e3a85c40f2b46ce2ee5f58ab56c30f38d99 (diff) | |
download | candle-9571b200c9a7c835f66a5444e62f8100e99a4102.tar.gz candle-9571b200c9a7c835f66a5444e62f8100e99a4102.tar.bz2 candle-9571b200c9a7c835f66a5444e62f8100e99a4102.zip |
fix firstToken, minor ui changes (#971)
Diffstat (limited to 'candle-wasm-examples/llama2-c/llama2cWorker.js')
-rw-r--r-- | candle-wasm-examples/llama2-c/llama2cWorker.js | 11 |
1 files changed, 9 insertions, 2 deletions
diff --git a/candle-wasm-examples/llama2-c/llama2cWorker.js b/candle-wasm-examples/llama2-c/llama2cWorker.js index abaf3401..a46b5bc8 100644 --- a/candle-wasm-examples/llama2-c/llama2cWorker.js +++ b/candle-wasm-examples/llama2-c/llama2cWorker.js @@ -50,6 +50,7 @@ async function generate(data) { tokenizerURL, prompt, temp, + top_p, repeatPenalty, seed, maxSeqLen, @@ -59,11 +60,17 @@ async function generate(data) { const model = await Llama2C.getInstance(weightsURL, modelID, tokenizerURL); self.postMessage({ status: "loading", message: "Initializing model" }); - model.init_with_prompt(prompt, temp, repeatPenalty, seed); + const firstToken = model.init_with_prompt( + prompt, + temp, + top_p, + repeatPenalty, + seed + ); const seq_len = model.get_seq_len(); - let sentence = ""; + let sentence = firstToken; let maxTokens = maxSeqLen ? maxSeqLen : seq_len - prompt.length - 1; let startTime = performance.now(); let tokensCount = 0; |