summaryrefslogtreecommitdiff
path: root/candle-examples/examples/mnist-training/main.rs
Commit message (Collapse)AuthorAgeFilesLines
* Allow for different behavior between training and eval (#1213)Laurent Mazare2023-10-291-2/+2
| | | | | * Forward with training. * Do not use dropout on vgg evaluation.
* Add the optimizer trait. (#702)Laurent Mazare2023-09-011-2/+2
|
* Mnist training dropout (#677)Laurent Mazare2023-08-301-7/+11
| | | | | * Use dropout in the mnist training. * Fix.
* Add some documentation. (#673)Laurent Mazare2023-08-301-1/+1
| | | | | * Add some documentation. * Bump the crate version.
* Simplify usage of the pool functions. (#662)Laurent Mazare2023-08-291-7/+9
| | | | | | | * Simplify usage of the pool functions. * Small tweak. * Attempt at using apply to simplify the convnet definition.
* Add a convnet training example. (#661)Laurent Mazare2023-08-291-1/+104
| | | | | | | * Add a convnet example. * Dataset fix. * Randomize batches.
* Re-enable local dir for mnist.Nicolas Patry2023-08-281-1/+9
|
* Training:Nicolas Patry2023-08-281-1/+1
| | | | | | | - Removed a lot of surface (SerializedFileReader ownership is really painful). - Moved example + vision to hf.co version. - Removed feature gate.
* VarBuilder cleanup (#627)Laurent Mazare2023-08-271-2/+2
| | | | | | | | | * VarBuilder cleanup. * Implement the basic varbuilders. * Add the sharded code. * Proper support for tensor sharding.
* Add a yolo-v3 example. (#528)Laurent Mazare2023-08-201-0/+3
| | | | | | | | | | | | | | | * Add a couple functions required for yolo. * Add the yolo-v3 example. * Add minimum and maximum. * Use the newly introduced maximum. * Cuda support for min/max + add some testing. * Allow for more tests to work with accelerate. * Fix a typo.
* Add a simple Module trait and implement it for the various nn layers (#500)Laurent Mazare2023-08-181-1/+1
| | | | | | | * Start adding the module trait. * Use the module trait. * Implement module for qmatmul.
* Add the candle-datasets crate (#322)Laurent Mazare2023-08-051-2/+2
| | | | | | | | | | | | | * Move the vision datasets to a separate crate. * Move the batcher bits. * Update the readme. * Move the tiny-stories bits. --------- Co-authored-by: Jane Doe <jane.doe@example.org>
* Llama more training (#297)Laurent Mazare2023-08-011-124/+15
| | | | | | | | | | | * Rework the var-builder to handle initializations. * Add some helper functions for layer creation. * Improve the layer initializations. * Get initialized variables. * Precompute the rot embeddings when training lamas.
* Make the nll op closer to the pytorch version + add a test. (#286)Laurent Mazare2023-07-311-4/+1
|
* Add a flag to set the number of epochs in the mnist training (#283)Laurent Mazare2023-07-311-14/+28
| | | | | * Add a flag to change the number of epochs for the mnist training. * Increase the learning rate for the MLP.
* Load a trained checkpoint in the mnist example. (#280)Laurent Mazare2023-07-301-3/+36
|
* Add a flag to save the trained weights. (#279)Laurent Mazare2023-07-301-0/+228