On Mon, Sep 13, 2021 at 3:20 PM Arnd Bergmann <a...@arndb.de> wrote:

> One straightforward hardware independent low-level API would
> be the traditional BLAS GEMM call[1] for matrix multiplication
> and its variants (integer, float, bfloat16, ...).

What this (and subsequent posts from Dave and Daniel) show, is
that the general pattern is that what we are accelerating is no longer
the specialized use cases of linear algebra such as 3D "shaders"
or whatever inference linear algebra NPUs are doing, which
appear to include regression, bayesian stuff, gaussian quadrature...
name it.

What we are talking about here is acceleration, using an efficient
data path, of numerical analysis, using tailored hardware.
I'm not even sure we are limited to linear algebra anymore.

Is this what is happening, and should we be thinking numerical
analysis accelerators and their different shapes and sizes
rather than usecase-foo-accelerators, so we don't end up with
this situation again the next time applied math comes knocking
on the door with their next usecase?

Yours,
Linus Walleij

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@tvm.apache.org
For additional commands, e-mail: dev-h...@tvm.apache.org

Reply via email to