mirror of
https://github.com/zebrajr/pytorch.git
synced 2025-12-07 12:21:27 +01:00
Clarify cycliclr param docs (#20880)
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/20880 This clarifies how the momentum parameters should be used. Reviewed By: soumith Differential Revision: D15482450 fbshipit-source-id: e3649a38876c5912cb101d8e404abca7c3431766
This commit is contained in:
parent
68c3ef72b5
commit
082936f033
|
|
@ -498,8 +498,10 @@ class CyclicLR(_LRScheduler):
|
|||
cycle_momentum (bool): If ``True``, momentum is cycled inversely
|
||||
to learning rate between 'base_momentum' and 'max_momentum'.
|
||||
Default: True
|
||||
base_momentum (float or list): Initial momentum which is the
|
||||
lower boundary in the cycle for each parameter group.
|
||||
base_momentum (float or list): Lower momentum boundaries in the cycle
|
||||
for each parameter group. Note that momentum is cycled inversely
|
||||
to learning rate; at the peak of a cycle, momentum is
|
||||
'base_momentum' and learning rate is 'max_lr'.
|
||||
Default: 0.8
|
||||
max_momentum (float or list): Upper momentum boundaries in the cycle
|
||||
for each parameter group. Functionally,
|
||||
|
|
@ -507,7 +509,10 @@ class CyclicLR(_LRScheduler):
|
|||
The momentum at any cycle is the difference of max_momentum
|
||||
and some scaling of the amplitude; therefore
|
||||
base_momentum may not actually be reached depending on
|
||||
scaling function. Default: 0.9
|
||||
scaling function. Note that momentum is cycled inversely
|
||||
to learning rate; at the start of a cycle, momentum is 'max_momentum'
|
||||
and learning rate is 'base_lr'
|
||||
Default: 0.9
|
||||
last_epoch (int): The index of the last batch. This parameter is used when
|
||||
resuming a training job. Since `step()` should be invoked after each
|
||||
batch instead of after each epoch, this number represents the total
|
||||
|
|
|
|||
Loading…
Reference in New Issue
Block a user