You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@mxnet.apache.org by aa...@apache.org on 2018/12/05 02:01:17 UTC
[incubator-mxnet] branch master updated: doc fix (#13465)
This is an automated email from the ASF dual-hosted git repository.
aaronmarkham pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-mxnet.git
The following commit(s) were added to refs/heads/master by this push:
new 0f85f5d doc fix (#13465)
0f85f5d is described below
commit 0f85f5de9eaa3c1b31bff9337268c071f634073d
Author: Anirudh <an...@gmail.com>
AuthorDate: Tue Dec 4 18:00:58 2018 -0800
doc fix (#13465)
---
python/mxnet/optimizer/optimizer.py | 7 +++++--
1 file changed, 5 insertions(+), 2 deletions(-)
diff --git a/python/mxnet/optimizer/optimizer.py b/python/mxnet/optimizer/optimizer.py
index d7b6821..a085b6f 100644
--- a/python/mxnet/optimizer/optimizer.py
+++ b/python/mxnet/optimizer/optimizer.py
@@ -688,8 +688,11 @@ class LBSGD(Optimizer):
state = momentum * state + lr * rescale_grad * clip(grad, clip_gradient) + wd * weight
weight = weight - state
- For details of the update algorithm see :class:`~mxnet.ndarray.lbsgd_update` and
- :class:`~mxnet.ndarray.lbsgd_mom_update`.
+ For details of the update algorithm see :class:`~mxnet.ndarray.sgd_update`
+ and :class:`~mxnet.ndarray.sgd_mom_update`.
+ In addition to the SGD updates the LBSGD optimizer uses the LARS, Layer-wise
+ Adaptive Rate Scaling, algorithm to have a separate learning rate for each
+ layer of the network, which leads to better stability over large batch sizes.
This optimizer accepts the following parameters in addition to those accepted
by :class:`.Optimizer`.