summaryrefslogtreecommitdiff
path: root/README.md
diff options
context:
space:
mode:
authorLaurent Mazare <laurent.mazare@gmail.com>2024-04-24 20:54:24 +0200
committerGitHub <noreply@github.com>2024-04-24 20:54:24 +0200
commitcfab6e761696c18b1ce5d3a339ab57ef191ca749 (patch)
tree2c12c2092807e6170adfe934d48591257a456cec /README.md
parent11d4a3c588e9ecb5bcc5a57773973b48036155f0 (diff)
downloadcandle-cfab6e761696c18b1ce5d3a339ab57ef191ca749.tar.gz
candle-cfab6e761696c18b1ce5d3a339ab57ef191ca749.tar.bz2
candle-cfab6e761696c18b1ce5d3a339ab57ef191ca749.zip
Mention phi-v3 in the readmes. (#2122)
Diffstat (limited to 'README.md')
-rw-r--r--README.md5
1 files changed, 3 insertions, 2 deletions
diff --git a/README.md b/README.md
index d7b01276..d730313a 100644
--- a/README.md
+++ b/README.md
@@ -66,7 +66,8 @@ We also provide a some command line based examples using state of the art models
- [Gemma](./candle-examples/examples/gemma/): 2b and 7b general LLMs from Google Deepmind.
- [RecurrentGemma](./candle-examples/examples/recurrent-gemma/): 2b and 7b
Griffin based models from Google that mix attention with a RNN like state.
-- [Phi-1, Phi-1.5, and Phi-2](./candle-examples/examples/phi/): 1.3b and 2.7b general LLMs with performance on par with LLaMA-v2 7b.
+- [Phi-1, Phi-1.5, Phi-2, and Phi-3](./candle-examples/examples/phi/): 1.3b,
+ 2.7b, and 3.8b general LLMs with performance on par with 7b models.
- [StableLM-3B-4E1T](./candle-examples/examples/stable-lm/): a 3b general LLM
pre-trained on 1T tokens of English and code datasets. Also supports
StableLM-2, a 1.6b LLM trained on 2T tokens, as well as the code variants.
@@ -203,7 +204,7 @@ If you have an addition to this list, please submit a pull request.
- LLaMA v1, v2, and v3 with variants such as SOLAR-10.7B.
- Falcon.
- StarCoder, StarCoder2.
- - Phi 1, 1.5, and 2.
+ - Phi 1, 1.5, 2, and 3.
- Mamba, Minimal Mamba
- Gemma 2b and 7b.
- Mistral 7b v0.1.