| Commit message (Expand) | Author | Age | Files | Lines |
* | Expose AdamW parameters (#1449) | Dave Lage | 2023-12-16 | 1 | -0/+8 |
* | Only optimize float tensors. (#1069) | Laurent Mazare | 2023-10-10 | 1 | -0/+5 |
* | Add the optimizer trait. (#702) | Laurent Mazare | 2023-09-01 | 1 | -46/+60 |
* | Set the learning rate. (#687) | Laurent Mazare | 2023-08-31 | 1 | -0/+8 |
* | Add the AdamW optimizer. (#307) | Laurent Mazare | 2023-08-02 | 1 | -2/+114 |
* | Llama more training (#297) | Laurent Mazare | 2023-08-01 | 1 | -1/+8 |
* | Add the gradient for reduce-sum. (#162) | Laurent Mazare | 2023-07-13 | 1 | -1/+1 |
* | Add the SGD optimizer (#160) | Laurent Mazare | 2023-07-13 | 1 | -0/+47 |