You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@mxnet.apache.org by aa...@apache.org on 2018/12/05 02:01:17 UTC

[incubator-mxnet] branch master updated: doc fix (#13465)

This is an automated email from the ASF dual-hosted git repository.

aaronmarkham pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-mxnet.git


The following commit(s) were added to refs/heads/master by this push:
     new 0f85f5d  doc fix (#13465)
0f85f5d is described below

commit 0f85f5de9eaa3c1b31bff9337268c071f634073d
Author: Anirudh <an...@gmail.com>
AuthorDate: Tue Dec 4 18:00:58 2018 -0800

    doc fix (#13465)
---
 python/mxnet/optimizer/optimizer.py | 7 +++++--
 1 file changed, 5 insertions(+), 2 deletions(-)

diff --git a/python/mxnet/optimizer/optimizer.py b/python/mxnet/optimizer/optimizer.py
index d7b6821..a085b6f 100644
--- a/python/mxnet/optimizer/optimizer.py
+++ b/python/mxnet/optimizer/optimizer.py
@@ -688,8 +688,11 @@ class LBSGD(Optimizer):
         state = momentum * state + lr * rescale_grad * clip(grad, clip_gradient) + wd * weight
         weight = weight - state
 
-    For details of the update algorithm see :class:`~mxnet.ndarray.lbsgd_update` and
-    :class:`~mxnet.ndarray.lbsgd_mom_update`.
+    For details of the update algorithm see :class:`~mxnet.ndarray.sgd_update`
+    and :class:`~mxnet.ndarray.sgd_mom_update`.
+    In addition to the SGD updates the LBSGD optimizer uses the LARS, Layer-wise
+    Adaptive Rate Scaling, algorithm to have a separate learning rate for each
+    layer of the network, which leads to better stability over large batch sizes.
 
     This optimizer accepts the following parameters in addition to those accepted
     by :class:`.Optimizer`.