Thank you, Cristobal.
That is good news.

Gus Correa

Cristobal Navarro wrote:
i have good news.

after updating to a newer kernel on ubuntu server nodes, sm is not a problem anymore for the nehalem CPUs!!!
my older kernel, was
Linux 2.6.32-22-server #36-Ubuntu SMP Thu Jun 3 20:38:33 UTC 2010 x86_64 GNU/Linux

and i upgraded to
Linux agua 2.6.32-24-server #39-Ubuntu SMP Wed Jul 28 06:21:40 UTC 2010 x86_64 GNU/Linux

that solved everything.
Gus, maybe the problem you had with fedora can be solved in a similar way.

we should keep this for the records.

regards
Cristobal






On Wed, Jul 28, 2010 at 6:45 PM, Gus Correa <g...@ldeo.columbia.edu <mailto:g...@ldeo.columbia.edu>> wrote:

    Cristobal Navarro wrote:

        Gus
        my kernel for all nodes is this one:
        Linux 2.6.32-22-server #36-Ubuntu SMP Thu Jun 3 20:38:33 UTC
        2010 x86_64 GNU/Linux


    Kernel is not my league.

    However, it would be great if somebody clarified
    for good these issues with Nehalem/Westmere, HT,
    shared memory and what the kernel is doing,
    or how to make the kernel do the right thing.
    Maybe Intel could tell.


        at least for the moment i will use this configuration, at least
        for deveplopment/testing  of the parallel programs.
        lag is minimum :)

        whenever i get another kernel update, i will test again to check
        if sm works, would be good to know that suddenly another
        distribution supports nehalem sm.

        best regards and thanks again
        Cristobal
        ps: guess what are the names of the other 2 nodes lol


    Acatenango (I said that before), and Pacaya.

    Maybe: Santa Maria, Santiaguito, Atitlan, Toliman, San Pedro,
    Cerro de Oro ... too many volcanoes, and some are multithreaded ...
    You need to buy more nodes!

    Gus




        On Wed, Jul 28, 2010 at 5:50 PM, Gus Correa
        <g...@ldeo.columbia.edu <mailto:g...@ldeo.columbia.edu>
        <mailto:g...@ldeo.columbia.edu <mailto:g...@ldeo.columbia.edu>>>
        wrote:

           Hi Cristobal

           Please, read my answer (way down the message) below.

           Cristobal Navarro wrote:



               On Wed, Jul 28, 2010 at 3:28 PM, Gus Correa
               <g...@ldeo.columbia.edu <mailto:g...@ldeo.columbia.edu>
        <mailto:g...@ldeo.columbia.edu <mailto:g...@ldeo.columbia.edu>>
               <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu> <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>>>>
               wrote:

                  Hi Cristobal

                  Cristobal Navarro wrote:



                      On Wed, Jul 28, 2010 at 11:09 AM, Gus Correa
                      <g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu> <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>>
               <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu> <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>>>
                      <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>
               <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>> <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>
               <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>>>>>
                      wrote:

                         Hi Cristobal

                         In case you are not using full path name for
               mpiexec/mpirun,
                         what does "which mpirun" say?


                      --> $which mpirun
                           /opt/openmpi-1.4.2


                         Often times this is a source of confusion, old
               versions may
                         be first on the PATH.

                         Gus


                      openMPI version problem is now gone, i can confirm
        that the
                      version is consistent now :), thanks.


                  This is good news.


                      however, i keep getting this kernel crash
        randomnly when i
                      execute with -np higher than 5
                      these are Xeons, with Hyperthreading On, is that a
        problem??


                  The problem may be with Hyperthreading, maybe not.
                  Which Xeons?


               --> they are not so old, not so new either
               fcluster@agua:~$ cat /proc/cpuinfo | more
               processor : 0
               vendor_id : GenuineIntel
               cpu family : 6
               model : 26
               model name : Intel(R) Xeon(R) CPU           E5520  @ 2.27GHz
               stepping : 5
               cpu MHz : 1596.000
               cache size : 8192 KB
               physical id : 0
               siblings : 8
               core id : 0
               cpu cores : 4
               apicid : 0
               initial apicid : 0
               fpu : yes
               fpu_exception : yes
               cpuid level : 11
               wp : yes
               flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr
        pge mca
               cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss h
               t tm pbe syscall nx rdtscp lm constant_tsc arch_perfmon
        pebs bts
               rep_good xtopology nonstop_tsc aperfmperf pni dtes64
        monitor ds_
               cpl vmx est tm2 ssse3 cx16 xtpr pdcm dca sse4_1 sse4_2 popcnt
               lahf_lm ida tpr_shadow vnmi flexpriority ept vpid
               bogomips : 4522.21
               clflush size : 64
               cache_alignment : 64
               address sizes : 40 bits physical, 48 bits virtual
               power management:
               ...same for cpu1, 2, 3, ..., 15.


           AHA! Nehalems!

           Here they are E5540, just a different clock speed, I suppose.


               information on how the cpu is distributed
               fcluster@agua:~$ lstopo
               System(7992MB)
                Socket#0 + L3(8192KB)
                  L2(256KB) + L1(32KB) + Core#0
                    P#0
                    P#8
                  L2(256KB) + L1(32KB) + Core#1
                    P#2
                    P#10
                  L2(256KB) + L1(32KB) + Core#2
                    P#4
                    P#12
                  L2(256KB) + L1(32KB) + Core#3
                    P#6
                    P#14
                Socket#1 + L3(8192KB)
                  L2(256KB) + L1(32KB) + Core#0
                    P#1
                    P#9
                  L2(256KB) + L1(32KB) + Core#1
                    P#3
                    P#11
                  L2(256KB) + L1(32KB) + Core#2
                    P#5
                    P#13
                  L2(256KB) + L1(32KB) + Core#3
                    P#7
                    P#15



                           If I remember right, the old hyperthreading
        on old Xeons was
                  problematic.

                  OTOH, about 1-2 months ago I had trouble with OpenMPI on a
                  relatively new Xeon Nehalem machine with (the new)
        Hyperthreading
                  turned on,
                  and Fedora Core 13.
                  The machine would hang with the OpenMPI connectivity
        example.
                  I reported this to the list, you may find in the archives.


               --i foudn the archives recently about an hour ago, was
        not sure
               if it was the same problem but i removed HT for testing with
               setting the online flag to 0 on the extra cpus showed with
               lstopo, unfortenately i also crashes, so HT may not be
        the problem.


           It didn't fix the problem in our Nehalem machine here either,
           although it was FC13, and I don't know what OS and kernel
        you're using.


                  Apparently other people got everything (OpenMPI with HT on
               Nehalem)
                  working in more stable distributions (CentOS, RHEL, etc).

                  That problem was likely to be in the FC13 kernel,
                  because even turning off HT I still had the machine
        hanging.
                  Nothing worked with shared memory turned on,
                  so I had to switch OpenMPI to use tcp instead,
                  which is kind of ridiculous in a standalone machine.


               --> very interesting, sm can be the problem

                      im trying to locate the kernel error on logs, but
        after
                      rebooting a crash, the error is not in the
        kern.log (neither
                      kern.log.1).
                      all i remember is that it starts with "Kernel BUG..."
                      and somepart it mentions a certain CPU X, where
        that cpu
               can be
                      any from 0 to 15 (im testing only in main node).
         Someone
               knows
                      where the log of kernel error could be?


                  Have you tried to turn off hyperthreading?


               --> yes, tried, same crashes.
                           In any case, depending on the application, it
        may not help much
                  performance to have HT on.

                  A more radical alternative is to try
                  -mca btl tcp,self
                  in the mpirun command line.
                  That is what worked in the case I mentioned above.


               wow!, this worked really :),  you pointed out the problem, it
               was shared memory.


           Great news!
           That's exactly the problem we had here.
           Glad that the same solution worked for you.

           Over a year ago another fellow reported the same problem on
        Nehalem,
           on the very early days of Nehalem.
           The thread should be in the archives.
           Somebody back then (Ralph, or Jeff, or other?)
           suggested that turning off "sm" might work.
           So, I take no credit for this.


               i have 4 nodes, so anyways there will be node
        comunication, do
               you think i can rely on working with -mca btl tcp,self??
        i dont
               mind small lag.


           Well, this may be it, short from reinstalling the OS.

           Some people reported everything works with OpenMPI+HT+sm in
        CentOS
           and RHEL, see the thread I mentioned in the archives from 1-2
        months
           ago.
           I don't administer that machine, and didn't have the time to
        do OS
           reinstall either.
           So I left it with -mca btl tcp,self, and the user/machine owner
           is happy that he can run his programs right,
           and with a performance that he considers good.


               i just have one more question, is this a problem of the
        ubuntu
               server kernel?? from the Nehalem Cpus?? from openMPI (i dont
               think) ??


           I don't have any idea.
           It may be a problem with some kernels, not sure.
           Which kernel do you have?

           Ours was FC-13, maybe FC-12, I don't remember exactly.
           Currently that machine has kernel 2.6.33.6-147.fc13.x86_64 #1
        SMP.
           However, it may have been a slightly older kernel when I
        installed
           OpenMPI there.
           It may have been 2.6.33.5-124.fc13.x86_64 or
        2.6.32.14-127.fc12.x86_64.
           My colleague here updates the machines with yum,
           so it may have gotten a new kernel since then.

           Our workhorse machines in the clusters that I take care
           of are AMD Opteron, never had this problem there.
           Maybe the kernels have yet to catch up with Nehalem,
           now Westmere, soon another one.


               and on what depends that in the future, sm could be
        possible on
               the same configuration i have?? kernel update?.


           You may want to try CentOS or RHEL, but I can't guarantee the
        results.
           Somebody else in the list may have had the direct experience,
           and may speak out.

           It may be worth the effort anyway.
           After all, intra-node communication should be
           running on shared memory.
           Having to turn it off is outrageous.

           If you try another OS distribution,
           and if it works, please report the results back to the list:
           OS/distro, kernel, OpenMPI version, HT on or off,
           mca btl sm/tcp/self/etc choices, compilers, etc.
           This type of information is a real time saving for everybody.



               Thanks very much Gus, really!
               Cristobal



           My pleasure.
           Glad that there was a solution, even though not the best.
           Enjoy your cluster with vocano-named nodes!
           Have fun with OpenMPI and PETSc!

           Gus Correa
---------------------------------------------------------------------
           Gustavo Correa
           Lamont-Doherty Earth Observatory - Columbia University
           Palisades, NY, 10964-8000 - USA
---------------------------------------------------------------------

My $0.02
                  Gus Correa


                         Cristobal Navarro wrote:


                             On Tue, Jul 27, 2010 at 7:29 PM, Gus Correa
                             <g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>
               <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>> <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>
               <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>>>
                      <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>
               <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>> <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>
               <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>>>>
                             <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>
               <mailto:g...@ldeo.columbia.edu <mailto:g...@ldeo.columbia.edu>>
                      <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>
               <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>>> <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>
               <mailto:g...@ldeo.columbia.edu <mailto:g...@ldeo.columbia.edu>>
                      <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>
               <mailto:g...@ldeo.columbia.edu
        <mailto:g...@ldeo.columbia.edu>>>>>>

                             wrote:

                                Hi Cristobal

                                Does it run only on the head node alone?
                                (Fuego? Agua? Acatenango?)
                                Try to put only the head node on the
        hostfile
               and execute
                             with mpiexec.

                             --> i will try only with the head node, and
        post
               results back
                                This may help sort out what is going on.
                                Hopefully it will run on the head node.

                                Also, do you have Infinband connecting
        the nodes?
                                The error messages refer to the openib
        btl (i.e.
                      Infiniband),
                                and complains of


                             no we are just using normal network 100MBit/s ,
               since i
                      am just
                             testing yet.


                                "perhaps a missing symbol, or compiled for a
               different
                                version of Open MPI?".
                                It sounds as a mixup of versions/builds.


                             --> i agree, somewhere there must be the
        remains
               of the older
                             version

                                Did you configure/build OpenMPI from
        source, or did
                      you install
                                it with apt-get?
                                It may be easier/less confusing to
        install from
               source.
                                If you did, what configure options did
        you use?


                             -->i installed from source, ./configure
                             --prefix=/opt/openmpi-1.4.2 --with-sge
        --without-xgid
                             --disable--static

                                Also, as for the OpenMPI runtime
        environment,
                                it is not enough to set it on
                                the command line, because it will be
        effective
               only on the
                             head node.
                                You need to either add them to the PATH and
                      LD_LIBRARY_PATH
                                on your .bashrc/.cshrc files (assuming these
               files and
                      your home
                                directory are *also* shared with the
        nodes via
               NFS),
                                or use the --prefix option of mpiexec to
        point
               to the
                      OpenMPI
                             main
                                directory.


                             yes, all nodes have their PATH and
        LD_LIBRARY_PATH
               set up
                             properly inside the login scripts ( .bashrc
        in my
               case  )

                                Needless to say, you need to check and
        ensure
               that the
                      OpenMPI
                                directory (and maybe your home
        directory, and
               your work
                             directory)
                                is (are)
                                really mounted on the nodes.


                             --> yes, doublechecked that they are

                                I hope this helps,


                             --> thanks really!

                                Gus Correa

                                Update: i just reinstalled openMPI, with
        the same
                      parameters,
                             and it
                                seems that the problem has gone, i
        couldnt test
                      entirely but
                             when i
                                get back to lab ill confirm.

                             best regards! Cristobal


------------------------------------------------------------------------

                             _______________________________________________
                             users mailing list
                             us...@open-mpi.org
        <mailto:us...@open-mpi.org> <mailto:us...@open-mpi.org
        <mailto:us...@open-mpi.org>>
               <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>
        <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>>
                      <mailto:us...@open-mpi.org
        <mailto:us...@open-mpi.org> <mailto:us...@open-mpi.org
        <mailto:us...@open-mpi.org>>
               <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>
        <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>>>


http://www.open-mpi.org/mailman/listinfo.cgi/users


                         _______________________________________________
                         users mailing list
                         us...@open-mpi.org <mailto:us...@open-mpi.org>
        <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>
               <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>
        <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>>
                      <mailto:us...@open-mpi.org
        <mailto:us...@open-mpi.org> <mailto:us...@open-mpi.org
        <mailto:us...@open-mpi.org>>
               <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>
        <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>>>


                         http://www.open-mpi.org/mailman/listinfo.cgi/users



------------------------------------------------------------------------

                      _______________________________________________
                      users mailing list
                      us...@open-mpi.org <mailto:us...@open-mpi.org>
        <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>
               <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>
        <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>>
                      http://www.open-mpi.org/mailman/listinfo.cgi/users


                  _______________________________________________
                  users mailing list
                  us...@open-mpi.org <mailto:us...@open-mpi.org>
        <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>
               <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>
        <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>>
                  http://www.open-mpi.org/mailman/listinfo.cgi/users



------------------------------------------------------------------------

               _______________________________________________
               users mailing list
               us...@open-mpi.org <mailto:us...@open-mpi.org>
        <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>
               http://www.open-mpi.org/mailman/listinfo.cgi/users


           _______________________________________________
           users mailing list
           us...@open-mpi.org <mailto:us...@open-mpi.org>
        <mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>
           http://www.open-mpi.org/mailman/listinfo.cgi/users



        ------------------------------------------------------------------------

        _______________________________________________
        users mailing list
        us...@open-mpi.org <mailto:us...@open-mpi.org>
        http://www.open-mpi.org/mailman/listinfo.cgi/users


    _______________________________________________
    users mailing list
    us...@open-mpi.org <mailto:us...@open-mpi.org>
    http://www.open-mpi.org/mailman/listinfo.cgi/users



------------------------------------------------------------------------

_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to