# keras-adabound **Repository Path**: frontxiang/keras-adabound ## Basic Information - **Project Name**: keras-adabound - **Description**: Keras implementation of AdaBound - **Primary Language**: Unknown - **License**: MIT - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2020-06-07 - **Last Updated**: 2021-03-11 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README # AdaBound for Keras Keras port of [AdaBound Optimizer for PyTorch](https://github.com/Luolc/AdaBound), from the paper [Adaptive Gradient Methods with Dynamic Bound of Learning Rate.](https://openreview.net/forum?id=Bkg3g2R9FX) ## Usage Add the `adabound.py` script to your project, and import it. Can be a dropin replacement for `Adam` Optimizer. Also supports `AMSBound` variant of the above, equivalent to `AMSGrad` from Adam. ```python from adabound import AdaBound optm = AdaBound(lr=1e-03, final_lr=0.1, gamma=1e-03, weight_decay=0., amsbound=False) ``` ## Results With a wide ResNet 34 and horizontal flips data augmentation, and 100 epochs of training with batchsize 128, it hits 92.16% (called v1). Weights are available inside the [Releases tab](https://github.com/titu1994/keras-adabound/releases/tag/0.1) #### NOTE - The smaller ResNet 20 models have been removed as they did not perform as expected and were depending on a flaw during the initial implementation. The ResNet 32 shows the actual performance of this optimizer. > With a small ResNet 20 and width + height data + horizontal flips data augmentation, and 100 epochs of training with batchsize 1024, it hits 89.5% (called v1). > On a small ResNet 20 with only width and height data augmentations, with batchsize 1024 trained for 100 epochs, the model gets close to 86% on the test set (called v3 below). ### Train Set Accuracy ### Train Set Loss ### Test Set Accuracy ### Test Set Loss # Requirements - Keras 2.2.4+ & Tensorflow 1.12+ (Only supports TF backend for now). - Numpy