# Best-Deep-Learning-Optimizers **Repository Path**: frontxiang/Best-Deep-Learning-Optimizers ## Basic Information - **Project Name**: Best-Deep-Learning-Optimizers - **Description**: Collection of the latest, greatest, deep learning optimizers (for Pytorch) - CNN, NLP suitable - **Primary Language**: Unknown - **License**: Apache-2.0 - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2020-06-07 - **Last Updated**: 2020-12-19 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # Best-Deep-Learning-Optimizers
Collection of the latest, greatest, deep learning optimizers (for Pytorch) - CNN, NLP suitable

Current top performer = DiffMod (AdaMod + DiffGrad combined)..but this is only on initial testing.

12/27 - added DiffGrad, and unofficial version 1 support (coded from the paper).
12/28 - added Diff_RGrad = diffGrad + Rectified Adam to start off....seems to work quite well. Medium article (summary and FastAI example usage): https://medium.com/@lessw/meet-diffgrad-new-deep-learning-optimizer-that-solves-adams-overshoot-issue-ec63e28e01b2 Official diffGrad paper: https://arxiv.org/abs/1909.11015v2 12/31 - AdaMod and DiffMod added. Initial SLS files added (but more work needed). In Progress:

A - Parabolic Approximation Line Search: https://arxiv.org/abs/1903.11991v2 B - Stochastic Line Search (SLS): pending (needs param group support) c - AvaGrad General papers of relevance: Does Adam stick close to the optimal point? https://arxiv.org/abs/1911.00289v1 Probabalistic line searches for stochastic optimization (2017, matlab only but good theory work): https://arxiv.org/abs/1703.10034v2