summaryrefslogtreecommitdiff
path: root/candle-examples/examples/bert
Commit message (Expand)AuthorAgeFilesLines
* bert attention mask (#1934)Zheng Li2024-08-011-2/+10
* Speed up bert with approx gelu (#1410)Juarez Bochi2023-12-062-3/+52
* Handle LongStorage in pytorch checkpoints. (#1152)Laurent Mazare2023-10-221-27/+20
* Adapt more examples to the updated safetensor api. (#947)Laurent Mazare2023-09-231-3/+2
* Add more example readmes. (#828)Laurent Mazare2023-09-121-0/+44
* Move some models to candle-transformers so that it's easier to re-use. (#794)Laurent Mazare2023-09-102-570/+1
* Upgrading hf-hub.Nicolas Patry2023-08-291-4/+4
* Add a yolo-v3 example. (#528)Laurent Mazare2023-08-201-0/+3
* Add a simple Module trait and implement it for the various nn layers (#500)Laurent Mazare2023-08-181-1/+1
* Add a cuda kernel for upsampling. (#441)Laurent Mazare2023-08-141-1/+4
* Normalize embeddings in the bert example. (#390)Laurent Mazare2023-08-101-0/+14
* Softmax numerical stability. (#267)Laurent Mazare2023-07-281-1/+1
* Upgrading hf-hub to `0.2.0` (Modified API to not pass the Repo aroundNicolas Patry2023-07-271-3/+4
* Rename the .r functions to .dims so as to be a bit more explicit. (#220)Laurent Mazare2023-07-222-3/+3
* Removing `candle-hub` internal to extract into `hf-hub` standalone.Nicolas Patry2023-07-191-1/+1
* Use mkl to accelerate binary ops. (#190)Laurent Mazare2023-07-181-2/+5
* Add Shape try into (#189)Laurent Mazare2023-07-181-1/+0
* Add more tracing details to bert. (#188)Laurent Mazare2023-07-181-2/+46
* Bert tracing (#184)Laurent Mazare2023-07-172-461/+545
* Add some 'cuda-if-available' helper function. (#172)Laurent Mazare2023-07-151-14/+1
* Removing cuda default.Nicolas Patry2023-07-141-1/+10
* Simplify the parameters used by sum and sum_keepdim. (#165)Laurent Mazare2023-07-141-1/+1
* Use the same default as pytorch for sum. (#164)Laurent Mazare2023-07-131-4/+4
* Tensor mutability (#154)Laurent Mazare2023-07-131-1/+1
* Remove some dead-code pragmas. (#137)Laurent Mazare2023-07-111-10/+2
* VarBuilder path creation (#131)Laurent Mazare2023-07-101-72/+42
* Move the var-builder in a central place. (#130)Laurent Mazare2023-07-101-61/+3
* [nn] Move the Embedding and Activation parts. (#116)Laurent Mazare2023-07-101-30/+7
* Sketch the candle-nn crate. (#115)Laurent Mazare2023-07-101-76/+33
* Merge pull request #92 from LaurentMazare/sync_hubNicolas Patry2023-07-071-8/+7
|\
| * Creating new sync Api for `candle-hub`.Nicolas Patry2023-07-061-8/+7
* | Enabling `roberta` for the example (it's the same model as Bert, withNicolas Patry2023-07-061-8/+26
|/
* Add mkl support for matrix multiply. (#86)Laurent Mazare2023-07-061-0/+4
* Add a simpler way to specify the dim index for some ops.laurent2023-07-051-2/+2
* Use avg pooling before the cosine similarity.laurent2023-07-051-3/+4
* Add some sentence similarity comparision to the bert example.laurent2023-07-051-63/+122
* Some polish.Nicolas Patry2023-07-051-7/+11
* Adding offline mode.Nicolas Patry2023-07-051-11/+41
* Upgrading bert example to work with `bert-base-uncased`.Nicolas Patry2023-07-041-16/+61
* Add a batch dimension on the bert example.laurent2023-07-041-11/+15
* Allow for batch dimensions in the embedding layer.laurent2023-07-031-6/+14
* Minor tweaks.laurent2023-07-031-1/+2
* Remove the fixed length hack.laurent2023-07-031-1/+1
* Bugfix for transpose.laurent2023-07-031-3/+4
* Fix the layer norm to properly handle bias.laurent2023-07-031-11/+30
* Get some embeddings out.laurent2023-07-031-14/+23
* Get the tensors to be loaded properly.laurent2023-07-031-11/+72
* Add the varbuilder + check shapes.laurent2023-07-031-62/+156
* Model creation.laurent2023-07-031-6/+134
* Complete (?) the forward pass.laurent2023-07-031-2/+28