[Relay][Quantization] KL-divergence-based per-layer calibration (#3538)
authorWuwei Lin <vincentl13x@gmail.com>
Fri, 2 Aug 2019 03:55:27 +0000 (20:55 -0700)
committerziheng <ziheng@apache.org>
Fri, 2 Aug 2019 03:55:27 +0000 (20:55 -0700)
commit33ab3c602874d128d600f9cfa5c23da37a9f1ab2
treefb53a8102521a4166fba8291793be8b4872bafe6
parent5357f49b0ed02e33faa590aa104760ee370a00ec
[Relay][Quantization] KL-divergence-based per-layer calibration (#3538)

* [Relay][Quantization] Support floating-point scale

* [Relay][Quantization] KL-divergence calibration on dataset

* Fix unhandled LeftShift case in QuantizeRealize

* Fix lint

* drop QBias

* fix lint

* address comments

* address comments

* Update comments

* address comments

* lint

* kQIdentity = 0
python/tvm/relay/quantize/__init__.py
python/tvm/relay/quantize/_annotate.py
python/tvm/relay/quantize/kl_divergence.py [new file with mode: 0644]
python/tvm/relay/quantize/quantize.py
src/relay/pass/quantize/calibrate.cc [new file with mode: 0644]
src/relay/pass/quantize/quantize.cc [moved from src/relay/pass/quantize.cc with 93% similarity]
src/relay/pass/quantize/quantize.h [moved from src/relay/pass/quantize.h with 89% similarity]