Hi, I've been working on generating CUDA code for sparse matrix multiplication 
using TVM in order to save me from the laborious work of designing and tuning 
the kernel. Since I'm new to TVM , I can't figure out why topi cannot express 
the computation. In this 
[code](https://github.com/amazon-research/FeatGraph/blob/master/python/featgraph/op/vanilla_spmm.py),
 it seems that the compute and schedule method can also be used in the SpMM. I 
suppose that maybe I can also use cache_read or other techniques described in 
this 
[tutorial](https://tvm.apache.org/docs/tutorials/optimize/opt_conv_cuda.html#sphx-glr-tutorials-optimize-opt-conv-cuda-py)
 to achieve higher performance? I've also read the 
[code](https://github.com/apache/tvm/blob/main/python/tvm/topi/cuda/sparse.py#L134)
 you mentioned above, and I'm wondering if there is a shorted way to express 
such operation. Thanks!





---
[Visit 
Topic](https://discuss.tvm.apache.org/t/sparse-opencl-error-scheduling-sparse-computations-that-use-tir-ir-builder/8585/8)
 to respond.

You are receiving this because you enabled mailing list mode.

To unsubscribe from these emails, [click 
here](https://discuss.tvm.apache.org/email/unsubscribe/48c2d306f825ad308695b3cd7c36a681756201ce077f8f898ffd8d447a9b842a).

Reply via email to