Re: [apache/incubator-tvm] [DEV] TVM v0.7 Roadmap (#4845)

2020-04-12 Thread shoubhik
What is the expected time of release for this release? what are the chances of it happening in May? -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/apache/incubator-tvm/issues/4845#issuecomment-612762262

Re: [dmlc/tvm] [QNN] [RFC] QNN Dialect -- Prequantize Models (#3591)

2019-07-25 Thread shoubhik
f quantize/dequantize ops > > being int32? Because, the current implementation for > > > > 1. Quantize - limits the inputs to be float32 and output to be (u)i8 > > 2. Dequantize - The input to be (u)int8 and output to be float32 > > > > Or are you suggesting we shoul

Re: [dmlc/tvm] [QNN] [RFC] QNN Dialect -- Prequantize Models (#3591)

2019-07-25 Thread shoubhik
@jackwish, i want to get my understanding correct, when you say > I was looking into PR #3531 and #3512 , and noticed that the PRs are going to > support 32 bits quantization. are you talking about the inputs or outputs of quantize/dequantize ops being int32? Because, the current implementation f

Re: [dmlc/tvm] [RFC][Quantization] Designing and lowering of quantized ops (#3512)

2019-07-25 Thread shoubhik
There are quite a lot of changes here that are depndent on #3531 . I am closing the PR for now. I will reopen this once #3531 is pushed. -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/pull/3512#i

Re: [dmlc/tvm] [RFC][Quantization] Designing and lowering of quantized ops (#3512)

2019-07-25 Thread shoubhik
Closed #3512. -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/pull/3512#event-2510690970

Re: [dmlc/tvm] [RFC][Quantization] Designing and lowering of quantized ops (#3512)

2019-07-17 Thread shoubhik
@liangfu made the changes you suggested. -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/pull/3512#issuecomment-512443841

Re: [dmlc/tvm] [RFC][Quantization] Designing and lowering of quantized ops (#3512)

2019-07-12 Thread shoubhik
> Mainly organizational issues, please make things consistent with what was > discussed in #3531 I have addressed the namespace issues and have followed the same convetion as #3531 in the new commit. -- You are receiving this because you are subscribed to this thread. Reply to this email direc

Re: [dmlc/tvm] [RFC][Quantization] Designing and lowering of quantized ops (#3512)

2019-07-11 Thread shoubhik
@FrozenGene and @tqchen, any other major comments for the PR? -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/pull/3512#issuecomment-510561960

Re: [dmlc/tvm] [RFC][Quantization] Designing and lowering of quantized ops (#3512)

2019-07-08 Thread shoubhik
@tqchen @FrozenGene @ZihengJiang @zhiics @wweic @eqy -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/pull/3512#issuecomment-509422639

Re: [dmlc/tvm] [RFC][Quantization] Designing and lowering of quantized ops (#3457)

2019-07-08 Thread shoubhik
Rebased to new PR #3512 -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/pull/3457#issuecomment-509421969

Re: [dmlc/tvm] [RFC][Quantization] Designing and lowering of quantized ops (#3457)

2019-07-08 Thread shoubhik
Closed #3457. -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: https://github.com/dmlc/tvm/pull/3457#event-2467657739

[dmlc/tvm] [RFC][Quantization] Designing and lowering of quantized ops (#3457)

2019-06-28 Thread shoubhik
The purpose of this PR is to dive deep into the desing of the quantized ops. To start the discussion I have implemented the Quantize and dequantize op which are easy to implement. There is one more such [PR](https://github.com/dmlc/tvm/issues/2351) but there the conversation has meandered towar

Re: [dmlc/tvm] [RFC][Quantization] Support quantized models from TensorflowLite (#2351)

2019-06-24 Thread shoubhik
> > We need to add `in_dtype` in the dequantize op as the calculations will be > > different, especially the range to use. > > Guess the input tensor has such information already? @jackwish, the input data is generally an `Expr` can be `Var` or `IntImm` or some other type of `Expr`. How will i

Re: [dmlc/tvm] [RFC][Quantization] Support quantized models from TensorflowLite (#2351)

2019-06-18 Thread shoubhik
> Thanks. Let's lay down the high-level API design for some of the quantized > operators. A large portion of this is coming from the following relevant > discussions. Thanks to @jackwish, @FrozenGene and @jnorwood for sharing their > experiences with quantization, and also @sho

Re: [dmlc/tvm] [RFC][Quantization] Support quantized models from TensorflowLite (#2351)

2019-06-15 Thread shoubhik
@FrozenGene a clarifying question to your above comment. If we pass in the output scale and shift can we not compute int32-> int8 by simply adding more nodes in the graph. -- You are receiving this because you are subscribed to this thread. Reply to this email directly or view it on GitHub: htt