# MultiStepScheduler#

class composer.optim.MultiStepScheduler(milestones, gamma=0.1)[source]#

Decays the learning rate discretely at fixed milestones.

This scheduler is based on MultiStepLR from PyTorch.

Decays the learning rate by a factor of gamma whenever a time milestone in milestones is reached.

Specifically, the learning rate multiplier $$\alpha$$ can be expressed as:

$\alpha(t) = \gamma ^ x$

Where $$x$$ represents the amount of milestones that have been reached, and $$\gamma$$ represents the multiplicative decay factor.

Parameters
• milestones (List[str | Time]) – Times at which the learning rate should change.

• gamma (float) – Multiplicative decay factor. Default = 0.1.