You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@tvm.apache.org by Wuwei Lin via TVM Discuss <no...@discuss.tvm.ai> on 2019/05/08 17:38:11 UTC
[TVM Discuss] [Development] Improving quantization accuracy with more
precise bias
The above example after annotation:
```
data
| |
sim_quantize(QINPUT) sim_quantize(QINPUT)
| |
add(bn_bias)
|
... /
|
add
```
data is usually output of previous conv2d. There are duplicated simulated_quantize. Followed add in both branches will convert the int8 to int32. So simulated_quantize + add in both branches which will be translated to `right_shift + cast(i8) + cast(i32)`
We use stop_fusion to ensure that previous conv2d result will be casted to int8 before saving in global memory.
You will see the difference running quantized ResNet-50 v2.
---
[Visit Topic](https://discuss.tvm.ai/t/improving-quantization-accuracy-with-more-precise-bias/2444/9) to respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails, [click here](https://discuss.tvm.ai/email/unsubscribe/9c1b4289f3c2f4ab9d9ff2f427d8e04f42aac1dc4c8734fd7507d9c1cb99825e).
Tianqi Chen, UW, Seattle, WA, 98105, United States
http://tracking.discuss.tvm.ai/tracking/unsubscribe?msgid=c2p7vXW3rOvqHpEkdLxa0w2