Hi Vivek,
I think I'm a bit lost now. We were originally talking about differences
in trajectories but from the mail you just sent I can see that you have
a segmentation fault, which is another problem.
I can only suggest that if you want to make use of 128 processors you
should download the CVS version of gromacs or wait until the 4.0
is out. Since in gromacs 3.3.x the protein has to reside as a whole
on one of the processors, this very likely limits your scaling.
Also, on 128 processors you will get a PME grid of 128x128xSomething
(since nx and ny have to be divisible by the number of CPUs) which is
probably way bigger than it needs to be (how big is it for a single
CPU?). Together with a PME order of 6 this leads to a large overlap
in the charge grid, which has to be communicated among the processors.
PME order 4 will be better suited for such a high parallelization, but
in general for Gromacs 3.x you should have at least a few thousand atoms
per processor, less than 1000 won't give you decent scaling at all.
Carsten
vivek sharma wrote:
Hi Carsten,
Thanks for your reply. Actually I am running MD simulation on a protein
molecule with 270 residues(2687 atoms), after adding water it is having
45599 atoms, and using the recent version of gromacs test available from
gromacs.org <http://gromacs.org> (gmxtest-3.3.3.tgz)
Following are the entries from the .mdp file I am using.....
**********md.mdp
title = trp_drg MD
cpp = /lib/cpp ; location of cpp on SGI
constraints = all-bonds
integrator = md
dt = 0.002 ; ps !
nsteps = 25000 ; total 50 ps.
nstcomm = 1
nstxout = 2500 ; output coordinates every 5.0 ps
nstvout = 0
nstfout = 0
nstlist = 5
ns_type = grid
rlist = 0.9
coulombtype = PME
rcoulomb = 0.9
rvdw = 1.4
fourierspacing = 0.12
fourier_nx = 0
fourier_ny = 0
fourier_nz = 0
pme_order = 6
ewald_rtol = 1e-5
optimize_fft = yes
; Berendsen temperature coupling is on in four groups
Tcoupl = berendsen
tau_t = 0.1 0.1 0.1
tc-grps = protein NDP sol
ref_t = 300 300 300
; Pressure coupling is on
Pcoupl = berendsen
pcoupltype = isotropic
tau_p = 0.5
compressibility = 4.5e-5
ref_p = 1.0
; Generate velocites is on at 300 K.
gen_vel = yes
gen_temp = 300.0
gen_seed = 173529
****************and Following are the commands I am using
grompp_d -np 128 -f md1.mdp -c 1XU9_A_b4em.gro -p 1XU9_A.top -o
1XU9_A_md1_np128.tpr
submit
mdrun_d
/////arguement for mdrun_d
-s 1XU9_A_md1_np128.tpr -o 1XU9_A_md1_np128.trr -c 1XU9_A_pmd1_np128.gro
-g md_np128.log -e md_np128.edr -np 128
***********Following is the error I am getting
Reading file 1XU9_A_md1_np128.tpr, VERSION 3.3.3 (double precision)
starting mdrun 'CORTICOSTEROID 11-BETA-DEHYDROGENASE, ISOZYME 1'
25000 steps, 50.0 ps.
srun: error: n141: task1: Segmentation fault
srun: Terminating job
****************************************************************
Is this information is helpfull in figuring out the problem.
Please, advice
With Thanks,
Vivek
2008/9/11 Carsten Kutzner <[EMAIL PROTECTED] <mailto:[EMAIL PROTECTED]>>
vivek sharma wrote:
Hi There,
I am running gromacs parellal version on cluster, with different
-np options.
Hi,
which version of gromacs exactly are you using?
On analyzing the 5 nsec trajectory using ngmx, I am finding
difference in the trajectory of two similar runs (only thing
varying in two runs in -np i.e 20 and 64 ), where mdp file and
input files are same in two cases.
I am wondering why I am getting this difference in two
trajectories ?
I am looking for the advice whether I did something wrong or
what may be the probable reason for this difference.
There are many reasons why a parallel run does not yield binary
identical
results to a run with another number of processors, even if you
start from
the same tpr file. If you use PME, then the FFTW could pick a slightly
different algorithm (it will select the fastest for that number of
processors.
This feature you can turn off by passing --disable-fftw-measure to the
gromacs configure script). But still you can get results that are not
binary identical if you do FFTs on a varying number of CPUs. Also, for
limited accuracy which is inherent to any computer, additions need not
be associative, which can show up in parallel additions.
Generally, if you run in double precision, these effects will be way
smaller,
but nevertheless you won't get binary identical results. This will
in all
cases lead to trajectories which slowly diverge from each other.
However,
in the fist few hundred time steps, you should not see any difference in
the first couple of decimals of the variables (positions, velocities,
energies ...)
Also, I am not able to run gromacs faster by increasing the -np
issue,
Please provide the exact command line you used.
Is there any max limit for scaling gromacs on parellal cluster ?
Yes, depending on your MD system and on the cluster you use :)
Carsten
With Thanks,
Vivek
------------------------------------------------------------------------
_______________________________________________
gmx-users mailing list gmx-users@gromacs.org
<mailto:gmx-users@gromacs.org>
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search
before posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
<mailto:[EMAIL PROTECTED]>.
Can't post? Read http://www.gromacs.org/mailing_lists/users.php
--
Dr. Carsten Kutzner
Max Planck Institute for Biophysical Chemistry
Theoretical and Computational Biophysics Department
Am Fassberg 11
37077 Goettingen, Germany
Tel. +49-551-2012313, Fax: +49-551-2012302
http://www.mpibpc.mpg.de/research/dep/grubmueller/
http://www.gwdg.de/~ckutzne <http://www.gwdg.de/%7Eckutzne>
_______________________________________________
gmx-users mailing list gmx-users@gromacs.org
<mailto:gmx-users@gromacs.org>
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before
posting!
Please don't post (un)subscribe requests to the list. Use the www
interface or send it to [EMAIL PROTECTED]
<mailto:[EMAIL PROTECTED]>.
Can't post? Read http://www.gromacs.org/mailing_lists/users.php
------------------------------------------------------------------------
_______________________________________________
gmx-users mailing list gmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php
--
Dr. Carsten Kutzner
Max Planck Institute for Biophysical Chemistry
Theoretical and Computational Biophysics Department
Am Fassberg 11
37077 Goettingen, Germany
Tel. +49-551-2012313, Fax: +49-551-2012302
http://www.mpibpc.mpg.de/research/dep/grubmueller/
http://www.gwdg.de/~ckutzne
_______________________________________________
gmx-users mailing list gmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php