Hi, everyone.
I'm waiting for the videos from TVMConf '21.
I see '20's was uploaded only a few days after the conference.
Could you share any information regarding the possible dates for '21's?
I'd be thrilled to watch them.
Thanks,
Jason
---
[Visit
Topic](https://discuss.tvm.apache.org/
Hi,
I'm trying to use TVM's stack to deploy INT8-quantized Transformer-based models.
I tried Relay + Ansor(AutoScheduler) for a Transformer (# layers = 1) and the
results weren't so neat.
|Time (ms)|Original|Quantized|
| --- | --- | --- |
|PyTorch|20|--|
|TVM (Relay, optimized)|130|120|
Thanks for the reply.
* PyTorch -> Relay -> Ansor -> TVM's low-level code -> LLVM/NVCC (LLVM was used
above)
* Both CPU and GPU (in particular, NVIDIA T4)
---
[Visit Topic](https://discuss.tvm.apache.org/t/quantized-transformer/11850/3)
to respond.
You are receiving this because you enab