Juergen
This is useful -- I was looking at LApack.cc already. It is in line with
what I need (as a template).
I am not worried about saving these things, but I have a 300x300
array of C float,and do a 300 element vector by 300 element multiply on
each of the 3 million rows in a "typical"process
topics below.
> Aim:Data Descriptions/Quantities:Vetting and Filtering:Processing
> speed:
> Frequency of use.
>
> Since you propose to do the work, who can estimate the cost.
>
> From: Xiao-Yong Jin To: fwei...@crisys.com
> Cc: GNU APL
> Sent: Friday, April 28, 2017
Thanks!
I'll probably go with SHMEM for future cuda/opencl use (I was thinking
along those lines). I don't yet need typical size -- the model I am
working with this weekend is vector8.bin, which is 71000 x 200 floats
(71000 words, each with 200 floats = 57MB) in size, but the *big* one is
much lar
Hi Fred,
I have not fully understood what you want to do exactly, but is
looks to me as if you want to go for
native GNU APL functions. Native functions provide the means to
bypass the GNU APL interpreter
itself to the extent desired. For example you
tions/Quantities:Vetting and Filtering:Processing speed:
Frequency of use.
Since you propose to do the work, who can estimate the cost.
From: Xiao-Yong Jin To: fwei...@crisys.com
Cc: GNU APL
Sent: Friday, April 28, 2017 9:32 PM
Subject: Re: [Bug-apl] Use with word2vec
If shared variables c
If shared variables can go through SHMEM, you can probably interface
cuda that way without much bottle neck.
But with the way GNU APL is implemented now, there are just too many
other limitations on performance with arrays of such size.
> On Apr 28, 2017, at 9:19 PM, Fred Weigel wrote:
>
> Jeurg