Sigh...look at the output of mpicc --showme. It tells you where the OMPI libs 
were installed:

-I/opt/SUNWhpc/HPC8.2.1c/sun/include/64 
-I/opt/SUNWhpc/HPC8.2.1c/sun/include/64/openmpi -R/opt/mx/lib/lib64
-R/opt/SUNWhpc/HPC8.2.1c/sun/lib/lib64 -L/opt/SUNWhpc/HPC8.2.1c/sun/lib/lib64 
-lmpi -lopen-rte -lopen-pal -lnsl -lrt -lm -ldl -lutil -lpthread

Look around a little in those areas - I can't pretend to understand where you 
put them, or if there are copy/paste errors into this thread. But obviously 
OMPI -thinks- the libs are somewhere in there.


On Apr 6, 2011, at 2:18 PM, Nehemiah Dacres wrote:

> [jian@therock lib]$ ls lib64/*.a
> lib64/libotf.a  lib64/libvt.fmpi.a  lib64/libvt.omp.a
> lib64/libvt.a   lib64/libvt.mpi.a   lib64/libvt.ompi.a
> last time i linked one of those files it told me they were in the wrong 
> format. these are in archive format, what format should they be in? 
> 
> 
> On Wed, Apr 6, 2011 at 2:44 PM, Ralph Castain <r...@open-mpi.org> wrote:
> Look at your output from mpicc --showme. It indicates that the OMPI libs were 
> put in the lib64 directory, not lib.
> 
> 
> On Apr 6, 2011, at 1:38 PM, Nehemiah Dacres wrote:
> 
>> I am also trying to get netlib's hpl to run via sun cluster tools so i am 
>> trying to compile it and am having trouble. Which is the proper mpi library 
>> to give? 
>> naturally this isn't going to work 
>> 
>> MPdir        = /opt/SUNWhpc/HPC8.2.1c/sun/
>> MPinc        = -I$(MPdir)/include
>> MPlib        = $(MPdir)/lib/libmpi.a
>> 
>> because that doesn't exist 
>> /opt/SUNWhpc-O/HPC8.2.1c/sun/lib/libotf.a  
>> /opt/SUNWhpc-O/HPC8.2.1c/sun/lib/libvt.fmpi.a  
>> /opt/SUNWhpc-O/HPC8.2.1c/sun/lib/libvt.omp.a
>> /opt/SUNWhpc-O/HPC8.2.1c/sun/lib/libvt.a   
>> /opt/SUNWhpc-O/HPC8.2.1c/sun/lib/libvt.mpi.a   
>> /opt/SUNWhpc-O/HPC8.2.1c/sun/lib/libvt.ompi.a
>> 
>> is what I have for listing *.a  in the lib directory. none of those are 
>> equivilant becasue they are all linked with vampire trace if I am reading 
>> the names right. I've already tried putting  
>> /opt/SUNWhpc-O/HPC8.2.1c/sun/lib/libvt.mpi.a for this and it didnt work 
>> giving errors like 
>> 
>> On Wed, Apr 6, 2011 at 12:42 PM, Terry Dontje <terry.don...@oracle.com> 
>> wrote:
>> Something looks fishy about your numbers.  The first two sets of numbers 
>> look the same and the last set do look better for the most part.  Your 
>> mpirun command line looks weird to me with the "-mca 
>> orte_base_help_aggregate btl,openib,self," did something get chopped off 
>> with the text copy?  You should have had a "-mca btl openib,self".  Can you 
>> do a run with "-mca btl tcp,self", it should be slower.
>> 
>> I really wouldn't have expected another compiler over IB to be that 
>> dramatically lower performing.
>> 
>> --td
>> 
>> 
>> 
>> On 04/06/2011 12:40 PM, Nehemiah Dacres wrote:
>>> also, I'm not sure if I'm reading the results right. According to the last 
>>> run, did using the sun compilers (update 1 )  result in higher performance 
>>> with sunct? 
>>> 
>>> On Wed, Apr 6, 2011 at 11:38 AM, Nehemiah Dacres <dacre...@slu.edu> wrote:
>>> some tests I did. I hope this isn't an abuse of the list. please tell me if 
>>> it is but thanks to all those who helped me. 
>>> 
>>> this  goes to say that the sun MPI works with programs not compiled with 
>>> sun’s compilers. 
>>> this first test was run as a base case to see if MPI works., the sedcond 
>>> run is to see the speed up using OpenIB provides
>>> jian@therock ~]$ mpirun -machinefile list 
>>> /opt/iba/src/mpi_apps/mpi_stress/mpi_stress
>>> Start mpi_stress at Wed Apr  6 10:56:29 2011
>>> 
>>>   Size (bytes)         TxMessages      TxMillionBytes/s       TxMessages/s
>>>             32              10000                  2.77           86485.67
>>>             64              10000                  5.76           90049.42
>>>            128              10000                 11.00           85923.85
>>>            256              10000                 18.78           73344.43
>>>            512              10000                 34.47           67331.98
>>>           1024              10000                 34.81           33998.09
>>>           2048              10000                 17.31            8454.27
>>>           4096              10000                 18.34            4476.61
>>>           8192              10000                 25.43            3104.28
>>>          16384              10000                 15.56             949.50
>>>          32768              10000                 13.95             425.74
>>> 
>>>          65536              10000                  9.88             150.79
>>>         131072               8192                 11.05              84.31
>>>         262144               4096                 13.12              50.04
>>>         524288               2048                 16.54              31.55
>>>        1048576               1024                 19.92              18.99
>>>        2097152                512                 22.54              10.75
>>>        4194304                256                 25.46               6.07
>>> 
>>> Iteration 0 : errors = 0, total = 0 (495 secs, Wed Apr  6 11:04:44 2011)
>>> After 1 iteration(s), 8 mins and 15 secs, total errors = 0
>>> 
>>> here is the infiniband run
>>> 
>>> [jian@therock ~]$ mpirun -mca orte_base_help_aggregate btl,openib,self, 
>>> -machinefile list /opt/iba/src/mpi_apps/mpi_stress/mpi_stress
>>> Start mpi_stress at Wed Apr  6 11:07:06 2011
>>> 
>>>   Size (bytes)         TxMessages      TxMillionBytes/s       TxMessages/s
>>>         32              10000                  2.72           84907.69
>>>         64              10000                  5.83           91097.94
>>>        128              10000                 10.75           83959.63
>>>        256              10000                 18.53           72384.48
>>>        512              10000                 34.96           68285.00
>>>       1024              10000                 11.40           11133.10
>>>       2048              10000                 20.88           10196.34
>>>       4096              10000                 10.13            2472.13
>>>       8192              10000                 19.32            2358.25
>>>      16384              10000                 14.58             890.10
>>>      32768              10000                 15.85             483.61
>>>      65536              10000                  9.04             137.95
>>>      131072        8192                 10.90              83.12
>>>     262144                   4096                 13.57              51.76
>>>     524288                  2048                 16.82              32.08
>>>    1048576        1024                 19.10              18.21
>>>    2097152                512                 22.13              10.55
>>>    4194304                256                 21.66               5.16
>>> 
>>> Iteration 0 : errors = 0, total = 0 (511 secs, Wed Apr  6 11:15:37 2011)
>>> After 1 iteration(s), 8 mins and 31 secs, total errors = 0
>>> compiled with the sun compilers i think 
>>> [jian@therock ~]$ mpirun -mca orte_base_help_aggregate btl,openib,self, 
>>> -machinefile list sunMpiStress
>>> Start mpi_stress at Wed Apr  6 11:23:18 2011
>>> 
>>>   Size (bytes)         TxMessages      TxMillionBytes/s       TxMessages/s
>>>             32              10000                  2.60           81159.60
>>>             64              10000                  5.19           81016.95
>>>            128              10000                 10.23           79953.34
>>>            256              10000                 16.74           65406.52
>>>            512              10000                 23.71           46304.92
>>>           1024              10000                 54.62           53340.73
>>>           2048              10000                 45.75           22340.58
>>>           4096              10000                 29.32            7158.87
>>>           8192              10000                 28.61            3492.77
>>>          16384              10000                184.03           11232.26
>>>          32768              10000                215.69            6582.21
>>>          65536              10000                229.88            3507.64
>>>         131072               8192                231.64            1767.25
>>>         262144               4096                220.73             842.00
>>>         524288               2048                121.61             231.95
>>>        1048576               1024                 66.54              63.46
>>>        2097152                512                 44.20              21.08
>>>        4194304                256                 45.17              10.77
>>> 
>>> Iteration 0 : errors = 0, total = 0 (93 secs, Wed Apr  6 11:24:52 2011)
>>> After 1 iteration(s), 1 mins and 33 secs, total errors = 0
>>> 
>>> sanity check: was sunMpiStress compiled using the sun compilers or oracle 
>>> compilerrs ? 
>>> [jian@therock ~]$ which mpirun
>>> 
>>> /opt/SUNWhpc/HPC8.2.1c/sun/bin/mpirun
>>> [jian@therock ~]$ ldd sunMpiStress
>>>    libmpi.so.0 => /opt/SUNWhpc/HPC8.2.1c/sun/lib/lib64/libmpi.so.0 
>>> (0x00002b5d2c6c3000)
>>>    libopen-rte.so.0 => 
>>> /opt/SUNWhpc/HPC8.2.1c/sun/lib/lib64/libopen-rte.so.0 (0x00002b5d2c8c1000)
>>>    libopen-pal.so.0 => 
>>> /opt/SUNWhpc/HPC8.2.1c/sun/lib/lib64/libopen-pal.so.0 (0x00002b5d2ca19000)
>>>    libnsl.so.1 => /lib64/libnsl.so.1 (0x0000003361400000)
>>>    librt.so.1 => /lib64/librt.so.1 (0x000000335f400000)
>>>    libm.so.6 => /lib64/libm.so.6 (0x000000335e400000)
>>>    libdl.so.2 => /lib64/libdl.so.2 (0x000000335e800000)
>>>    libutil.so.1 => /lib64/libutil.so.1 (0x000000336ba00000)
>>>    libpthread.so.0 => /lib64/libpthread.so.0 (0x000000335ec00000)
>>>    libc.so.6 => /lib64/libc.so.6 (0x000000335e000000)
>>>    /lib64/ld-linux-x86-64.so.2 (0x000000335dc00000)
>>> [jian@therock ~]$ which mpicc
>>> /opt/SUNWhpc/HPC8.2.1c/sun/bin/mpicc
>>> [jian@therock ~]$ mpicc /opt/iba/src/mpi_apps/mpi_stress/mpi_stress.c -o 
>>> sunMpiStress --showme
>>> cc /opt/iba/src/mpi_apps/mpi_stress/mpi_stress.c -o sunMpiStress 
>>> -I/opt/SUNWhpc/HPC8.2.1c/sun/include/64 
>>> -I/opt/SUNWhpc/HPC8.2.1c/sun/include/64/openmpi -R/opt/mx/lib/lib64 
>>> -R/opt/SUNWhpc/HPC8.2.1c/sun/lib/lib64             
>>> -L/opt/SUNWhpc/HPC8.2.1c/sun/lib/lib64 -lmpi -lopen-rte -lopen-pal -lnsl 
>>> -lrt -lm -ldl -lutil -lpthread
>>> [jian@therock ~]$ which cc
>>> /opt/sun/sunstudio12.1/bin/cc
>>> looks like it!
>>> 
>>> 
>>> 
>>> 
>>> 
>>> -- 
>>> Nehemiah I. Dacres
>>> System Administrator 
>>> Advanced Technology Group Saint Louis University
>>> 
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> -- 
>> 
>> Terry D. Dontje | Principal Software Engineer
>> Developer Tools Engineering | +1.781.442.2631
>> Oracle - Performance Technologies
>> 95 Network Drive, Burlington, MA 01803
>> Email terry.don...@oracle.com
>> 
>> 
>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> 
>> -- 
>> Nehemiah I. Dacres
>> System Administrator 
>> Advanced Technology Group Saint Louis University
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> 
> -- 
> Nehemiah I. Dacres
> System Administrator 
> Advanced Technology Group Saint Louis University
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to