| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
| |
* Forward with training.
* Do not use dropout on vgg evaluation.
|
| |
|
|
|
|
|
| |
* Use dropout in the mnist training.
* Fix.
|
|
|
|
|
| |
* Add some documentation.
* Bump the crate version.
|
|
|
|
|
|
|
| |
* Simplify usage of the pool functions.
* Small tweak.
* Attempt at using apply to simplify the convnet definition.
|
|
|
|
|
|
|
| |
* Add a convnet example.
* Dataset fix.
* Randomize batches.
|
| |
|
|
|
|
|
|
|
| |
- Removed a lot of surface (SerializedFileReader ownership is really
painful).
- Moved example + vision to hf.co version.
- Removed feature gate.
|
|
|
|
|
|
|
|
|
| |
* VarBuilder cleanup.
* Implement the basic varbuilders.
* Add the sharded code.
* Proper support for tensor sharding.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Add a couple functions required for yolo.
* Add the yolo-v3 example.
* Add minimum and maximum.
* Use the newly introduced maximum.
* Cuda support for min/max + add some testing.
* Allow for more tests to work with accelerate.
* Fix a typo.
|
|
|
|
|
|
|
| |
* Start adding the module trait.
* Use the module trait.
* Implement module for qmatmul.
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* Move the vision datasets to a separate crate.
* Move the batcher bits.
* Update the readme.
* Move the tiny-stories bits.
---------
Co-authored-by: Jane Doe <jane.doe@example.org>
|
|
|
|
|
|
|
|
|
|
|
| |
* Rework the var-builder to handle initializations.
* Add some helper functions for layer creation.
* Improve the layer initializations.
* Get initialized variables.
* Precompute the rot embeddings when training lamas.
|
| |
|
|
|
|
|
| |
* Add a flag to change the number of epochs for the mnist training.
* Increase the learning rate for the MLP.
|
| |
|
|
|