We have not fully aligned and not consistent behaviour in regard of float
rounding to int values for quantization.
* qnn.op.quantize does not have round param and uses tir.round that rounds to
nearest
* qnn.op.requantize has round param which might accept two values "UPWARD"
(default) and "TONEAREST.
UPWARD that TVM has is something very artificial that might have sense for
quantization and improving accuracy but is not described in standard, is not
supported by hardware and is not used by any deep learning frameworks.
Having this TVM "UPWARD" currently as default parameter, requires to implement
more actions and leads to performance implications.
**Proposal:**
1. Remove round parameter from qnn.op.requantize at all
2. Or standardize the rounding in TVM, add it to qnn.op.quantize, make default
value that something is supported by hardware natively like "TONEAREST" or
"FLOOR"
*More details:*
The current TVM requantize UPWARD mode is defined follow way:
```
rounding : string, optional
Defines the rounding direction when the value is midway between two
representable values.
```
I.e. it affect only numbers having x.5 values, especial for negative numbers.
It is not matched to any round standard in spite of
https://en.wikipedia.org/wiki/Rounding describes many approaches. TVM has new
special one.
How requantize is called from TVM importers:
- mxnet: calls 4 times, no passing of rounding param, i.e. default "UPWARD" is
used
- onnx: calls twice, once without rounding, second time with rounding ==
"TONEAREST"
- qnn_torch: one call without rounding param passing
- tflite: 9 calls without rounding param passing
If we take a look how rounding happens/defined in frameworks, then in most
cases it is not standardized, only onnx defines explicitly. Other can use
different round strategy for different implementations depending on hardware
* onnx
from https://github.com/onnx/onnx/blob/master/docs/Operators.md#QuantizeLinear
For (x / y_scale), it's rounding to nearest ties to even.
Refer to https://en.wikipedia.org/wiki/Rounding for
details.
* mxnet
no mention of the rounding in API, mxnet requantize op does not have round
param
cpu - intgemm is used (https://github.com/kpu/intgemm), for rounding -
_mm256_floor_ps
* qnn-torch
std::lrintf is used in
fused_nbit_rowwise_conversion.cc/FloatToFused8BitRowwiseQuantized__base
the default value of rounding returned by fegetround is 0 that stands for
FE_TONEAREST
floorf is used pytorch/caffe2/quantization/server/norm_minimization_avx2.cc
* tflite
- cl, metal - uses round, i.e. to nearest
- cpu - gemmlowp::RoundingDivideByPOT, rounding-to-nearest
https://github.com/google/gemmlowp/blob/master/fixedpoint/fixedpoint.h#L365
---
[Visit
Topic](https://discuss.tvm.apache.org/t/rounding-parameter-from-float-to-int-for-quantization/11479/1)
to respond.
You are receiving this because you enabled mailing list mode.
To unsubscribe from these emails, [click
here](https://discuss.tvm.apache.org/email/unsubscribe/bb5958ae6456d88ee585a15d12ee55b0bd80ddf37a0fe7922b3fa75c98548cdc).