You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@mxnet.apache.org by "Ray Zhang (JIRA)" <ji...@apache.org> on 2018/07/14 02:40:00 UTC

[jira] [Created] (MXNET-677) int8 quantization does not work on toy mnist dataset

Ray Zhang created MXNET-677:
-------------------------------

             Summary: int8 quantization does not work on toy mnist dataset
                 Key: MXNET-677
                 URL: https://issues.apache.org/jira/browse/MXNET-677
             Project: Apache MXNet
          Issue Type: Bug
            Reporter: Ray Zhang


The reproducible repository is linked here: 

 

Currently, airbnb is using the quantization extensions of mxnet to boost inference time on several convolutional neural network models. However, it has been difficult to achieve. The most complicated bugs lie in the intersection between the python and C++ interface, like the ones that crash jupyter kernels and are hard to run pdb on.

Airbnb currently extensively uses gluon models and are not planning to move to Module models any time soon for training, but it seems that creating a quantized Module model solely for inference is useful. Please refer to the repository for a minimum reproducible example.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@mxnet.apache.org
For additional commands, e-mail: issues-help@mxnet.apache.org