You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@mxnet.apache.org by GitBox <gi...@apache.org> on 2018/11/24 11:39:22 UTC

[GitHub] JWarlock opened a new issue #13391: Gluon: Per-layer learning rate for fine tuning a pretrained network

JWarlock opened a new issue #13391: Gluon: Per-layer learning rate for fine tuning a pretrained network
URL: https://github.com/apache/incubator-mxnet/issues/13391
 
 
   Hi!
   I want to train a pretrained model, for example, ResNet, where I want to limit the first two layers’ learning rate to lr/100, middle layers ’ learning rate to lr/10, and fully connected layers ’ learning rate to lr.
   I’ve only found implementation using native mxnet: https://github.com/apache/incubator-mxnet/issues/2242.
   But How to implement it using gluon and gluon.Trainer?

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
users@infra.apache.org


With regards,
Apache Git Services