summaryrefslogtreecommitdiff
path: root/README.md
diff options
context:
space:
mode:
authorLaurent Mazare <laurent.mazare@gmail.com>2023-12-15 19:16:06 -0600
committerGitHub <noreply@github.com>2023-12-15 19:16:06 -0600
commit30a958e5dd6152da0d9e4cf5ce338bd2dd6a0ec4 (patch)
treeaa505d27e9f68f0e37042ff0eca02ea0486ec537 /README.md
parent614842b311a12ac5aba130e165763f997d8ff324 (diff)
downloadcandle-30a958e5dd6152da0d9e4cf5ce338bd2dd6a0ec4.tar.gz
candle-30a958e5dd6152da0d9e4cf5ce338bd2dd6a0ec4.tar.bz2
candle-30a958e5dd6152da0d9e4cf5ce338bd2dd6a0ec4.zip
Quantized mixtral model (#1442)
* Add the Mixtral model. * Add more of the mixtral layers. * Add the final layers for mixtral. * Sketch the expert selection. * Add some expert routing logic. * Hopefully finish the routing logic for mixtral. * Add the mixtral example. * Fix the weight filenames. * Bugfix. * Another fix. * Yet another fix + remove the unused pragma. * Shape fix. * Support for quantized mixtral. * Support mixtral in the quantized example. * Mlp or moe type. * Fix the expert field namings. * Refactor the mlp bit. * More MoE logic. * Add the MoE quantized logic. * Fix the experts length.
Diffstat (limited to 'README.md')
0 files changed, 0 insertions, 0 deletions