Jianhui Tian wrote:
Hi,
I have a small system of fullerene ball with waters. I want to
simulation the system without pbc and pme, treating the coulomb and vdw
without cutoff. The system can be run on single processor. But when I
try to run it with multiple processors, it can't proceed. I am including
the error message at the end of this mail.
Those messages are generic failure messages from the MPI system. The
stack trace therein suggests you should read the stderr and/or log file
to get some diagnostic information from GROMACS. Possibly your system is
blowing up because you've started from an unsuitable configuration.
There are some of my guesses for the possible reason of the error:
1. Do we have to use pbc and pme to use multiple processors for simulation?
Wouldn't think so. However because you have no cut-offs, there will be
no advantage to DD because there is no data locality - every processor
needs the position of every atom. mdrun -pd may work. It strikes me as
possible that this scenario doesn't work in parallel on GROMACS.
2. Can wen use restraints when use multiple processors?
Yes.
3. If a molecule is "divided" into two parts in domain decomposition,
will this be a problem for simulation?
No, that's routine.
Mark
Thanks for any suggestion about this error message.
logfile:--------------------------------------------------------------------------
[0,1,1]: OpenIB on host compute-1-12.local was unable to find any HCAs.
Another transport will be used instead, although this may result in
lower performance.
--------------------------------------------------------------------------
--------------------------------------------------------------------------
[0,1,0]: OpenIB on host compute-1-12.local was unable to find any HCAs.
Another transport will be used instead, although this may result in
lower performance.
--------------------------------------------------------------------------
NNODES=2, MYRANK=0, HOSTNAME=compute-1-12.local
NODEID=0 argc=13
NNODES=2, MYRANK=1, HOSTNAME=compute-1-12.local
NODEID=1 argc=13
.
.
.
.
Reading file
/data/disk04/tianj/trp_remd_091012/equilibrium/test/trp_full_run1.tpr,
VERSION 4.0 (single precision)
[compute-1-12:08033] *** Process received signal ***
[compute-1-12:08033] Signal: Segmentation fault (11)
[compute-1-12:08033] Signal code: Address not mapped (1)
[compute-1-12:08033] Failing at address: 0xc0
[compute-1-12:08033] [ 0] /lib64/libpthread.so.0 [0x378fe0de80]
[compute-1-12:08033] [ 1] /lib64/libc.so.6(_IO_vfprintf+0x39) [0x378f642309]
[compute-1-12:08033] [ 2] /lib64/libc.so.6(_IO_fprintf+0x88) [0x378f64cf68]
[compute-1-12:08033] [ 3]
/usr/local/gromacs-4.0.4/bin/mdrun_sm(mk_mshift+0x315) [0x516593]
[compute-1-12:08033] [ 4] /usr/local/gromacs-4.0.4/bin/mdrun_sm [0x45fa97]
[compute-1-12:08033] [ 5]
/usr/local/gromacs-4.0.4/bin/mdrun_sm(dd_bonded_cg_distance+0x36c)
[0x45ff53]
[compute-1-12:08033] [ 6]
/usr/local/gromacs-4.0.4/bin/mdrun_sm(init_domain_decomposition+0x780)
[0x44d10c]
[compute-1-12:08033] [ 7]
/usr/local/gromacs-4.0.4/bin/mdrun_sm(mdrunner+0x89c) [0x429f6e]
[compute-1-12:08033] [ 8]
/usr/local/gromacs-4.0.4/bin/mdrun_sm(main+0x7ba) [0x4306b6]
[compute-1-12:08033] [ 9] /lib64/libc.so.6(__libc_start_main+0xf4)
[0x378f61d8b4]
[compute-1-12:08033] [10] /usr/local/gromacs-4.0.4/bin/mdrun_sm [0x4199a9]
[compute-1-12:08033] *** End of error message ***
mpirun noticed that job rank 0 with PID 8033 on node compute-1-12.local
exited on signal 11 (Segmentation fault).
1 additional process aborted (not shown)
------------------------------------------------------------------------
_______________________________________________
gmx-users mailing list gmx-users@gromacs.org
http://lists.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to gmx-users-requ...@gromacs.org.
Can't post? Read http://www.gromacs.org/mailing_lists/users.php
_______________________________________________
gmx-users mailing list gmx-users@gromacs.org
http://lists.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to gmx-users-requ...@gromacs.org.
Can't post? Read http://www.gromacs.org/mailing_lists/users.php