About 54 results
Open links in new tab
  1. Optimizers - Keras

    Keras documentation: Optimizers Abstract optimizer base class. If you intend to create your own optimization algorithm, please inherit from this class and override the following methods: build: …

  2. Optimizers - Keras

    Keras documentation: Optimizers Apply gradients to variables. Arguments grads_and_vars: List of (gradient, variable) pairs. name: string, defaults to None. The name of the namescope to use when …

  3. Adam - Keras

    Keras documentation: Adam Optimizer that implements the Adam algorithm. Adam optimization is a stochastic gradient descent method that is based on adaptive estimation of first-order and second …

  4. Model training APIs - Keras

    Arguments optimizer: String (name of optimizer) or optimizer instance. See keras.optimizers. loss: Loss function. May be a string (name of loss function), or a keras.losses.Loss instance. See keras.losses. …

  5. Adagrad - Keras

    Keras documentation: Adagrad Optimizer that implements the Adagrad algorithm. Adagrad is an optimizer with parameter-specific learning rates, which are adapted relative to how frequently a …

  6. AdamW - Keras

    Keras documentation: AdamW Optimizer that implements the AdamW algorithm. AdamW optimization is a stochastic gradient descent method that is based on adaptive estimation of first-order and second …

  7. SGD - Keras

    Keras documentation: SGD Gradient descent (with momentum) optimizer. Update rule for parameter w with gradient g when momentum is 0:

  8. Lion - Keras

    Keras documentation: Lion Optimizer that implements the Lion algorithm. The Lion optimizer is a stochastic-gradient-descent method that uses the sign operator to control the magnitude of the …

  9. Muon - Keras

    Keras documentation: Muon Optimizer that implements the Muon algorithm. Note that this optimizer should not be used in the following layers: Embedding layer Final output fully connected layer Any …

  10. Adadelta - Keras

    Keras documentation: Adadelta Optimizer that implements the Adadelta algorithm. Adadelta optimization is a stochastic gradient descent method that is based on adaptive learning rate per dimension to …