naga raju wrote:
Dear gmx users,
I have some problem regarding scalling factor of cluster. Here I have
given cluster specifications, installation procedure, i request you go
through query and suggest me.
Here is the system specifications:
Master node: Intel Pentium 4, 3.0 GHz with 64HT 800 MHz FSB/2 MB L2
Slave Nodes: Intel Xeon, 3.0 Ghz with HT 800 MHz FSB/2 MB L2
OS : Redhat Linux Enterprizes 4.0, 64 bit
I downloaded fftw-3.0.1.tar.gz. and gromacs-3.3.1tar.gz from gromacs
website.
For parallel installation, I used the fallowing commands(in master node,
in root),
For fftw instalation..
./configure --enable-float --enable-sse --enable-mpi
make
make install
For gromacs installation..
./configure --enable-mpi
make
make mdrun
make install
make install-mdrun
make links
It was installed in /opt/gromacs/
I didn't get any error messages
To the gromacs in parallel, I have used the fallowing commands...
grompp -f equilibrium.mdp -n index.ndx -p dopc.top -c dopc.pdb -np 5 -o
dopc-eq.tpr
mpirun -np 5 -machinefile gmxguest -nolocal /opt/gromacs/bin/mdrunmpi
-s dopc-eq.tpr -c dopc-eq.pdb -o dopc-eq.trr
(Note: gmxguest file contains five node names) I checked in 5 nodes, job
is going in all nodes.
The cluster took 22 hours to finish the job but same job with same time
scale in with one Intel Pentium 4, 3.0 GHz , 64bit, rehat linux OS, it
took 26 hours.
My question is why it is happening, is there any problem in installing
or gromacs not support this type of clusters.
To improve scalling factor of cluster, what i have to do?
If you look at the bottom section of your log file, you'll see GROMACS
report where it spent fractions of its time. It's also going to be less
painful to test this stuff over two processors, rather than five, and
over a short simulation, not a long one.
IIRC "./configure --enable-mpi" won't install
"/opt/gromacs/bin/mdrunmpi" either in that location or with that name so
I suggest you look at config.log in the installation directory and
report your actual configure line, not what you think you did :-)
Otherwise, thanks for providing the details you did.
You should test your cluster with some simple parallel code in which you
can see the speedup, and report what hardware is doing your interconnects.
Mark
_______________________________________________
gmx-users mailing list gmx-users@gromacs.org
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php