Optim¶
optim¶
-
class
seq2seq.optim.optim.
Optimizer
(optim, max_grad_norm=0)¶ The Optimizer class encapsulates torch.optim package and provides functionalities for learning rate scheduling and gradient norm clipping.
Parameters: - optim (torch.optim.Optimizer) – optimizer object, the parameters to be optimized should be given when instantiating the object, e.g. torch.optim.SGD(params)
- max_grad_norm (float, optional) – value used for gradient norm clipping, set 0 to disable (default 0)
-
set_scheduler
(scheduler)¶ Set the learning rate scheduler.
Parameters: scheduler (torch.optim.lr_scheduler.*) – object of learning rate scheduler, e.g. torch.optim.lr_scheduler.StepLR
-
step
()¶ Performs a single optimization step, including gradient norm clipping if necessary.