Jeff, Thanks for the tip - it started me thinking a bit.
I was using a directory in my /home account with 4.1.5 that I had previously built using GCC 9.4 (Ubuntu 20.04). I rebuilt the system with Ubuntu-22.04 but I did a backup of /home. Then I copied the 4.1.5 directory to /home again. I checked and I did a "make clean" before attempting to build 4.1.5 but with GCC 11.3 that came with Ubuntu 22.04. In fact, I did it several times before I ran configure. Even after running "make clean" I got the error I mentioned in my initial post. This happened several times. This morning, I blew away my 4.1.5 directory and downloaded a fresh 4.1.5. Configure went fine as did compiling it. My theory is that some cruft from building 4.1.5 with GCC 9.4 compilers hung around, even after "make clean". Using a "fresh" download of 4.1.5 did not include this "cruft" so configure and make all proceeds just fine. I don't know if this is correct and I can't point to any smoking gun though. Thanks! Jeff On Mon, Jul 17, 2023 at 2:53 PM Jeff Squyres (jsquyres) <jsquy...@cisco.com> wrote: > That's a little odd. Usually, the specific .h files that are listed as > dependencies came from *somewhere* -- usually either part of the GNU > Autotools dependency analysis. > > I'm guessing that /usr/lib/gcc/x86_64-linux-gnu/9/include/float.h doesn't > actually exist on your system -- but then how did it get into Open MPI's > makefiles? > > Did you run configure on one machine and make on a different machine, > perchance? > ------------------------------ > *From:* users <users-boun...@lists.open-mpi.org> on behalf of Jeffrey > Layton via users <users@lists.open-mpi.org> > *Sent:* Monday, July 17, 2023 2:05 PM > *To:* Open MPI Users <users@lists.open-mpi.org> > *Cc:* Jeffrey Layton <layto...@gmail.com> > *Subject:* [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3 > > Good afternoon, > > I'm trying to build Open MPI 4.1.5 using GCC 11.3. However, I get an error > that I'm not sure how to correct. The error is, > > ... > CC pscatter.lo > CC piscatter.lo > CC pscatterv.lo > CC piscatterv.lo > CC psend.lo > CC psend_init.lo > CC psendrecv.lo > CC psendrecv_replace.lo > CC pssend_init.lo > CC pssend.lo > CC pstart.lo > CC pstartall.lo > CC pstatus_c2f.lo > CC pstatus_f2c.lo > CC pstatus_set_cancelled.lo > CC pstatus_set_elements.lo > CC pstatus_set_elements_x.lo > CC ptestall.lo > CC ptestany.lo > CC ptest.lo > CC ptest_cancelled.lo > CC ptestsome.lo > CC ptopo_test.lo > CC ptype_c2f.lo > CC ptype_commit.lo > CC ptype_contiguous.lo > CC ptype_create_darray.lo > make[3]: *** No rule to make target > '/usr/lib/gcc/x86_64-linux-gnu/9/include/float.h', needed by > 'ptype_create_f90_complex.lo'. Stop. > make[3]: Leaving directory > '/home/laytonjb/src/openmpi-4.1.5/ompi/mpi/c/profile' > make[2]: *** [Makefile:2559: all-recursive] Error 1 > make[2]: Leaving directory '/home/laytonjb/src/openmpi-4.1.5/ompi/mpi/c' > make[1]: *** [Makefile:3566: all-recursive] Error 1 > make[1]: Leaving directory '/home/laytonjb/src/openmpi-4.1.5/ompi' > make: *** [Makefile:1912: all-recursive] Error 1 > > > > Here is the configuration output from configure: > > Open MPI configuration: > ----------------------- > Version: 4.1.5 > Build MPI C bindings: yes > Build MPI C++ bindings (deprecated): no > Build MPI Fortran bindings: mpif.h, use mpi, use mpi_f08 > MPI Build Java bindings (experimental): no > Build Open SHMEM support: false (no spml) > Debug build: no > Platform file: (none) > > Miscellaneous > ----------------------- > CUDA support: no > HWLOC support: external > Libevent support: internal > Open UCC: no > PMIx support: Internal > > Transports > ----------------------- > Cisco usNIC: no > Cray uGNI (Gemini/Aries): no > Intel Omnipath (PSM2): no > Intel TrueScale (PSM): no > Mellanox MXM: no > Open UCX: no > OpenFabrics OFI Libfabric: no > OpenFabrics Verbs: no > Portals4: no > Shared memory/copy in+copy out: yes > Shared memory/Linux CMA: yes > Shared memory/Linux KNEM: no > Shared memory/XPMEM: no > TCP: yes > > Resource Managers > ----------------------- > Cray Alps: no > Grid Engine: no > LSF: no > Moab: no > Slurm: yes > ssh/rsh: yes > Torque: no > > OMPIO File Systems > ----------------------- > DDN Infinite Memory Engine: no > Generic Unix FS: yes > IBM Spectrum Scale/GPFS: no > Lustre: no > PVFS2/OrangeFS: no > > > > Any suggestions! Thanks! > > Jeff > > > >