summaryrefslogtreecommitdiff
path: root/candle-wasm-examples/llama2-c
diff options
context:
space:
mode:
authorRadamés Ajna <radamajna@gmail.com>2023-09-05 11:24:43 -0700
committerGitHub <noreply@github.com>2023-09-05 19:24:43 +0100
commit6a40decc76246b23f7f1c6ada539c93afadfffba (patch)
tree3681808348103462949105f5f000a3462bf83c78 /candle-wasm-examples/llama2-c
parenta0d65585db0323747f71b4f33831a165b56a759b (diff)
downloadcandle-6a40decc76246b23f7f1c6ada539c93afadfffba.tar.gz
candle-6a40decc76246b23f7f1c6ada539c93afadfffba.tar.bz2
candle-6a40decc76246b23f7f1c6ada539c93afadfffba.zip
Minor WASM UI improvements (#748)
* add stats * random seed btn * minor ui improvoments
Diffstat (limited to 'candle-wasm-examples/llama2-c')
-rw-r--r--candle-wasm-examples/llama2-c/lib-example.html38
-rw-r--r--candle-wasm-examples/llama2-c/llama2cWorker.js12
2 files changed, 36 insertions, 14 deletions
diff --git a/candle-wasm-examples/llama2-c/lib-example.html b/candle-wasm-examples/llama2-c/lib-example.html
index bc519e4b..5995f003 100644
--- a/candle-wasm-examples/llama2-c/lib-example.html
+++ b/candle-wasm-examples/llama2-c/lib-example.html
@@ -60,23 +60,30 @@
const seed = getValue("seed");
const maxSeqLen = getValue("max-seq");
- function updateStatus({ status, message, prompt, sentence }) {
+ function updateStatus(data) {
const outStatus = document.querySelector("#output-status");
const outGen = document.querySelector("#output-generation");
+ const outCounter = document.querySelector("#output-counter");
- switch (status) {
+ switch (data.status) {
case "loading":
outStatus.hidden = false;
- outStatus.textContent = message;
+ outStatus.textContent = data.message;
outGen.hidden = true;
+ outCounter.hidden = true;
break;
case "generating":
+ const { message, prompt, sentence, tokensSec, totalTime } = data;
outStatus.hidden = true;
+ outCounter.hidden = false;
outGen.hidden = false;
outGen.innerHTML = `<span class="font-semibold">${prompt}</span>${sentence.replace(
/\<s\>|\<\/s\>/g,
""
)}`;
+ outCounter.innerHTML = `${(totalTime / 1000).toFixed(
+ 2
+ )}s (${tokensSec.toFixed(2)} tok/s)`;
break;
case "complete":
outStatus.hidden = true;
@@ -206,8 +213,9 @@
id="prompt"
class="font-light w-full px-3 py-2 mx-1 resize-none outline-none"
placeholder="Add your prompt here..."
+ value="Once upon a time"
/>
- <button class="invisible" id="clear-btn">
+ <button id="clear-btn">
<svg
fill="none"
xmlns="http://www.w3.org/2000/svg"
@@ -225,7 +233,6 @@
</button>
<button
id="run"
- disabled
class="bg-gray-700 hover:bg-gray-800 text-white font-normal py-2 w-16 rounded disabled:bg-gray-300 disabled:cursor-not-allowed"
>
Run
@@ -291,17 +298,26 @@
value="299792458"
class="font-light border border-gray-700 text-right rounded-md p-2"
/>
+ <button
+ id="run"
+ onclick="document.querySelector('#seed').value = BigInt(Math.floor(Math.random() * 2**64-1))"
+ class="bg-gray-700 hover:bg-gray-800 text-white font-normal py-1 w-[50px] rounded disabled:bg-gray-300 disabled:cursor-not-allowed text-sm"
+ >
+ Rand
+ </button>
</div>
<div>
<h3 class="font-medium">Generation:</h3>
-
<div
- class="min-h-[250px] bg-slate-100 text-gray-500 p-4 rounded-md grid"
+ class="min-h-[250px] bg-slate-100 text-gray-500 p-4 rounded-md flex flex-col gap-2"
>
- <p hidden id="output-generation"></p>
- <span
- id="output-status"
- class="justify-self-center self-center font-light"
+ <div
+ id="output-counter"
+ hidden
+ class="ml-auto font-semibold grid-rows-1 text-sm"
+ ></div>
+ <p hidden id="output-generation" class="grid-rows-2"></p>
+ <span id="output-status" class="m-auto font-light"
>No output yet</span
>
</div>
diff --git a/candle-wasm-examples/llama2-c/llama2cWorker.js b/candle-wasm-examples/llama2-c/llama2cWorker.js
index ba303aaa..e4229055 100644
--- a/candle-wasm-examples/llama2-c/llama2cWorker.js
+++ b/candle-wasm-examples/llama2-c/llama2cWorker.js
@@ -60,9 +60,10 @@ async function generate(data) {
const seq_len = model.get_seq_len();
let sentence = "";
- let max_tokens = maxSeqLen ? maxSeqLen : seq_len - prompt.length - 1;
-
- while (max_tokens--) {
+ let maxTokens = maxSeqLen ? maxSeqLen : seq_len - prompt.length - 1;
+ let startTime = performance.now();
+ let tokensCount = 0;
+ while (tokensCount < maxTokens) {
await new Promise(async (resolve) => {
if (controller && controller.signal.aborted) {
self.postMessage({
@@ -73,6 +74,8 @@ async function generate(data) {
return;
}
const token = await model.next_token();
+ const tokensSec =
+ ((tokensCount + 1) / (performance.now() - startTime)) * 1000;
sentence += token;
self.postMessage({
@@ -80,10 +83,13 @@ async function generate(data) {
message: "Generating token",
token: token,
sentence: sentence,
+ totalTime: performance.now() - startTime,
+ tokensSec,
prompt: prompt,
});
setTimeout(resolve, 0);
});
+ tokensCount++;
}
self.postMessage({
status: "complete",