he code stack in dbx looks the same on X64 platforms as the code on
SPARC except the address is smaller on the former.
Greg, I would be interested in knowing if you are still seeing the
problem on Leopard and whether the above setting helps any.
--td
*
*Subject:* Re: [OMPI users] ScaLapack and
All,
I really didn't want to start a new thread discussing the virtues and
vices of every compiler, since this is hardly my forte and the
opportunity to offend someone is fairly high, whilst making myself
look clownish
What I should have said was that "for my organization one cannot
ju
> Intel ifort is 7.7x faster then Linux g95 on MacPro 3.0 GHz
> Intel ifort is 2.9x faster then Linux g95 on Dual Opteron 1.4 GHz
> Intel ifort is 1.8x faster then Linux g95 on SGI Altix 350 dual
> Itanium2 1.4 GHz
> OS X g95 is 2.7x faster then Linux g95 on a MacPro 2.66 GHz (same
> hardwa
On Mar 6, 2008, at 12:49 PM, Doug Reeder wrote:
Greg,
I would disagree with your statement that the available fortran
options can't pass a cost-benefit analysis. I have found that for
scientific programming (e.g., Livermore Fortran Kernels and actual
PDE solvers) that code produced by the
.
Note, I have not confirmed this failure with a test case but the code
stack in dbx looks the same on X64 platforms as the code on SPARC
except
the address is smaller on the former.
Greg, I would be interested in knowing if you are still seeing the
problem on Leopard and whether the above setting
we haven't seen this fail in user codes.
Note, I have not confirmed this failure with a test case but the code
stack in dbx looks the same on X64 platforms as the code on SPARC
except
the address is smaller on the former.
Greg, I would be interested in knowing if you are still seeing the
pr
e on SPARC except
the address is smaller on the former.
Greg, I would be interested in knowing if you are still seeing the
problem on Leopard and whether the above setting helps any.
--td
*
*Subject:* Re: [OMPI users] ScaLapack and BLACS on Leopard
*From:* Terry Dontje (/Terry.Dontje_at_[h
What kind of system lib errors are you seeing and do you have a stack
trace? Note, I was trying something similar with Solaris and 64-bit on
a SPARC machine and was seeing segv's inside the MPI Library due to a
pointer being passed through an integer (thus dropping the upper 32
bits). Funny t