summaryrefslogtreecommitdiff
path: root/README.md
diff options
context:
space:
mode:
authorLaurent Mazare <laurent.mazare@gmail.com>2024-02-28 21:02:41 +0100
committerGitHub <noreply@github.com>2024-02-28 21:02:41 +0100
commit4fd00b890036ef67391a9cc03f896247d0a75711 (patch)
tree19121cce4cab5406e9fda202de71a32427c096bd /README.md
parent57267cd53612ede04090853680125b17956804f3 (diff)
downloadcandle-4fd00b890036ef67391a9cc03f896247d0a75711.tar.gz
candle-4fd00b890036ef67391a9cc03f896247d0a75711.tar.bz2
candle-4fd00b890036ef67391a9cc03f896247d0a75711.zip
Add the StarCoder2 model. (#1779)
* Add the StarCoder2 model. * Add the example code and get things to work. * And also tweak the readme.
Diffstat (limited to 'README.md')
-rw-r--r--README.md5
1 files changed, 3 insertions, 2 deletions
diff --git a/README.md b/README.md
index aea30c5b..e76f9853 100644
--- a/README.md
+++ b/README.md
@@ -76,7 +76,8 @@ We also provide a some command line based examples using state of the art models
- [Mixtral8x7b-v0.1](./candle-examples/examples/mixtral/): a sparse mixture of
experts 8x7b general LLM with better performance than a Llama 2 70B model with
much faster inference.
-- [StarCoder](./candle-examples/examples/bigcode/): LLM specialized to code generation.
+- [StarCoder](./candle-examples/examples/bigcode/) and
+ [StarCoder2](./candle-examples/examples/starcoder2/): LLM specialized to code generation.
- [Qwen1.5](./candle-examples/examples/qwen/): Bilingual (English/Chinese) LLMs.
- [RWKV v5](./candle-examples/examples/rwkv/): An RNN with transformer level LLM
performance.
@@ -191,7 +192,7 @@ If you have an addition to this list, please submit a pull request.
- Language Models.
- LLaMA v1 and v2 with variants such as SOLAR-10.7B.
- Falcon.
- - StarCoder.
+ - StarCoder, StarCoder2.
- Phi 1, 1.5, and 2.
- Mamba, Minimal Mamba
- Gemma 2b and 7b.