summaryrefslogtreecommitdiff
path: root/candle-transformers
Commit message (Expand)AuthorAgeFilesLines
* Fix lints for clippy 1.75. (#1494)Laurent Mazare2023-12-281-1/+1
* add config_amazon_mistral_lite (#1493)Daniel Clough2023-12-281-0/+18
* Bump the crate version to 0.3.3. (#1490)Laurent Mazare2023-12-281-3/+3
* feat: add clear_kv_cache to mistral and qmistral models (#1464)drbh2023-12-212-0/+28
* make fn name generic (#1459)Daniel Clough2023-12-211-1/+2
* add fn config_chat_ml (#1458)Daniel Clough2023-12-201-0/+19
* Merge pull request #1318 from huggingface/metal4Nicolas Patry2023-12-201-0/+1
|\
| * Revert candle-transformers.Nicolas Patry2023-12-151-38/+4
| * Fixing softmax.Nicolas Patry2023-12-151-4/+0
| * Fixing tests + matmul from MFANicolas Patry2023-12-131-1/+0
| * Better version ?Nicolas Patry2023-12-131-4/+5
| * Lots of updates including some stack of command buffers.nicolas2023-12-122-4/+43
* | Bump the crate version to 0.3.2. (#1452)Laurent Mazare2023-12-171-3/+3
* | Fix a couple typos (#1451)Laurent Mazare2023-12-173-7/+7
* | Quantized mixtral model (#1442)Laurent Mazare2023-12-151-21/+150
* | Add the Mixtral model. (#1437)Laurent Mazare2023-12-152-0/+500
* | Quantized version for phi-v2. (#1430)Laurent Mazare2023-12-131-0/+18
* | Support for phi-2. (#1429)Laurent Mazare2023-12-131-0/+34
* | Speed up bert with approx gelu (#1410)Juarez Bochi2023-12-061-2/+4
* | Stable Diffusion Turbo Support (#1395)Edwin Cheng2023-12-034-35/+168
* | Add Euler Ancestral Discrete Scheduler (#1390)Edwin Cheng2023-12-025-5/+312
|/
* Use the llama weight names for the Yi example. (#1381)Laurent Mazare2023-11-271-2/+6
* Distibert (#1366)Odunayo2023-11-242-0/+343
* Fix linspace implementation (#1358)MilkFather2023-11-231-7/+10
* Use the whisper-v3 tokenizer now that it has been added. (#1337)Laurent Mazare2023-11-161-1/+1
* feat: adds reset_kv_cache (#1335)drbh2023-11-161-0/+20
* fix: address clippy 0.1.74 issues (#1336)drbh2023-11-161-2/+1
* Update for 0.3.1. (#1324)Laurent Mazare2023-11-111-3/+3
* Add the Yi-6b and Yi-34b models. (#1320)Laurent Mazare2023-11-112-0/+378
* Add min to buckets in relative_position_bucket (#1312)Andy Braga2023-11-101-1/+1
* Add support to UL2 model family (#1300)Juarez Bochi2023-11-093-11/+53
* Add support for TrOCR Model (#1303)Ogundepo Odunayo2023-11-093-15/+465
* Fix bug introduced in madlad PR (#1298)Juarez Bochi2023-11-082-4/+4
* Preliminary support for whisper v3. (#1294)Laurent Mazare2023-11-082-3/+7
* Add support for MADLAD400 (#1285)Juarez Bochi2023-11-073-4/+33
* Share the layer-norm implementation. (#1248)Laurent Mazare2023-11-032-56/+32
* Remove the unused pragma for marian. (#1236)Laurent Mazare2023-11-011-4/+32
* Consolidate the with-tracing usage. (#1234)Laurent Mazare2023-11-014-102/+8
* Preliminary support for ssd1b. (#1233)Laurent Mazare2023-11-012-0/+73
* Add a KV cache to marian decoding. (#1226)Laurent Mazare2023-10-311-14/+40
* Add support for the marian base model. (#1221)Laurent Mazare2023-10-301-0/+25
* Use the hub files for the marian example. (#1220)Laurent Mazare2023-10-301-10/+29
* Bugfixes for marian-mt. (#1219)Laurent Mazare2023-10-301-9/+18
* Marian MT model (#1210)Laurent Mazare2023-10-293-0/+421
* Allow for different behavior between training and eval (#1213)Laurent Mazare2023-10-291-16/+19
* feat: implement VGG13, VGG16 and VGG19 (#1211)drbh2023-10-292-0/+255
* Infer the config for llama2-c. (#1208)Laurent Mazare2023-10-282-1/+50
* Move the llama2-c model in transformers. (#1205)Laurent Mazare2023-10-285-0/+712
* Make more models cloneable. (#1203)Laurent Mazare2023-10-283-26/+26
* Add the relu2 and relu6 activations. (#1201)Laurent Mazare2023-10-272-0/+57