On Wed, 2008-02-13 at 14:21 +0000, [EMAIL PROTECTED] wrote:
> >
> > On a more rational note, has any thought been given to what "good enough
> > performance for release" will be?
> 
> Should we perhaps add a performance benchmark to the tests?
> 
> Normalising it to account for hardware variations might be a problem.

There are existing benchmarks in examples/benchmarks/ , which have
mostly solved the problem of hardware variations [1] by comparing
instead against the performance of other dynamic language
implementations.  Certainly different versions of those other language
implementations will have varying performance, but we can standardize
our comparison on, say, the most recent release of each.

There is also the language shootout, which has quite a few more
comparisons available.

My personal feeling is that for each comparison benchmark, we should be
comparing PIR and Rakudo performance against the fastest other language
implementation for that benchmark.  Then we should be looking for two
problems -- relatively consistent poor performance (indicating a general
implementation design problem), and particular benchmarks that do
unusually poorly (indicating some particular axis of pain).

Some of these problems will be solveable, others will be amenable to
workarounds, and others will probably just need to be documented.  But
we should really make sure we know as much as possible what the problems
*are* before the first production release.


-'f

[1] Mind you, we still need to know if for example PPC handles method
calls way slower than Intel, so we can determine if there is something
we can do about that.  But that is another story.


Reply via email to