Chung Hwan Kim writes:

> I and two other students have formed up a team for a project called
> "Accelerating Dynamic Binary Translation with the GPUs". As the name of
> the project suggests our main idea is to parallelize Dynamic Binary
> Translation (DBT) process and speed it up with GPUs using the NVIDIA
> CUDA library.

AFAIK, DBT is a fairly control flow intensive code, so you'll probably
run into lots of branch divergence problems, so that performance will
suffer a lot, even if you use instruction template tables (like in
qemu's PPC target).

Nonetheless, I think new fermi models have less problems with that, but
it's still an architecture thought for control-flow-homogeneous parallel
code.

In any case, I'm not sure what is the real cost of translation related
to execution, it all depends on the kind of applications you're running;
but the computation on the GPU better have a huge speedup compared to
the current approach, or otherwise the data transfers to/from the GPU
will dominate the cost, specially if they're small transfers.


Lluis

--
 "And it's much the same thing with knowledge, for whenever you learn
 something new, the whole world becomes that much richer."
 -- The Princess of Pure Reason, as told by Norton Juster in The Phantom
 Tollbooth

Reply via email to