Share via


KnownStochasticOptimizer enum

Known values of StochasticOptimizer that the service accepts.

Fields

Adam

Adam is algorithm the optimizes stochastic objective functions based on adaptive estimates of moments

Adamw

AdamW is a variant of the optimizer Adam that has an improved implementation of weight decay.

None

No optimizer selected.

Sgd

Stochastic Gradient Descent optimizer.