• Docs >
  • Optimizers & Schedulers
Shortcuts

Optimizers & Schedulers

Optimizers

Asteroid relies on torch_optimizer and torch for optimizers. We provide a simple get method that retrieves optimizers from string, which makes it easy to specify optimizers from the command line.

Here is a list of supported optimizers, retrievable from string:

  • AccSGD

  • AdaBound

  • AdaMod

  • DiffGrad

  • Lamb

  • NovoGrad

  • PID

  • QHAdam

  • QHM

  • RAdam

  • SGDW

  • Yogi

  • Ranger

  • RangerQH

  • RangerVA

  • Adam

  • RMSprop

  • SGD

  • Adadelta

  • Adagrad

  • Adamax

  • AdamW

  • ASG

Schedulers

Asteroid provides step-wise learning schedulers, integrable to pytorch-lightning via System.