summaryrefslogtreecommitdiff
path: root/README.md
diff options
context:
space:
mode:
authorLaurent Mazare <laurent.mazare@gmail.com>2024-02-11 17:04:57 +0100
committerGitHub <noreply@github.com>2024-02-11 17:04:57 +0100
commit1e26d539d9f9574222e8d049fdbfadfa09e3ce2e (patch)
tree0fee8fdca3dd10f47ebbcdb2724249363dbe53aa /README.md
parent74497e6bf738366d4c599b42826e204fbcb96f37 (diff)
downloadcandle-1e26d539d9f9574222e8d049fdbfadfa09e3ce2e.tar.gz
candle-1e26d539d9f9574222e8d049fdbfadfa09e3ce2e.tar.bz2
candle-1e26d539d9f9574222e8d049fdbfadfa09e3ce2e.zip
Improved mamba model optimized for inference (#1694)
* Sketch the mamba model for inference. * Complete the forward pass. * Add the mamba example. * Optimize the selective-scan part. * Fix a couple shape mismatches and get inference to work. * Tweak the readmes. * More readme tweaks.
Diffstat (limited to 'README.md')
-rw-r--r--README.md4
1 files changed, 2 insertions, 2 deletions
diff --git a/README.md b/README.md
index 90344b34..9bfa30d8 100644
--- a/README.md
+++ b/README.md
@@ -67,7 +67,7 @@ We also provide a some command line based examples using state of the art models
- [StableLM-3B-4E1T](./candle-examples/examples/stable-lm/): a 3b general LLM
pre-trained on 1T tokens of English and code datasets. Also supports
StableLM-2, a 1.6b LLM trained on 2T tokens, as well as the code variants.
-- [Minimal Mamba](./candle-examples/examples/mamba-minimal/): a minimal
+- [Mamba](./candle-examples/examples/mamba/): an inference only
implementation of the Mamba state space model.
- [Mistral7b-v0.1](./candle-examples/examples/mistral/): a 7b general LLM with
better performance than all publicly available 13b models as of 2023-09-28.
@@ -186,7 +186,7 @@ If you have an addition to this list, please submit a pull request.
- Falcon.
- StarCoder.
- Phi 1, 1.5, and 2.
- - Minimal Mamba
+ - Mamba, Minimal Mamba
- Mistral 7b v0.1.
- Mixtral 8x7b v0.1.
- StableLM-3B-4E1T, StableLM-2-1.6B, Stable-Code-3B.