You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@tvm.apache.org by GitBox <gi...@apache.org> on 2022/11/19 03:32:19 UTC

[GitHub] [tvm] honghuichao commented on issue #12707: [Bug] a bug about fake quantize to intege and split op.

honghuichao commented on issue #12707:
URL: https://github.com/apache/tvm/issues/12707#issuecomment-1320778719

   @mbrookhart Sorry to trouble you,I've got a question to consult you. why the output type of  qnn.add must be int8 when the input type  of qnn.add is int8. I thik the output type of qnn.add / qnn.mul/..  can set by user via api.
   because I met such case:
   quantize+dequantize+add+add+quantize:
   I run FQ2i Pass. it will transform to quantize+qnn.add(out_type="int8")+qnn.add.
   but I think the right expr is quantize+qnn.add(output_type="int32")+qnn.add


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@tvm.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org