It's your call, v1.8 RM.  :-)

On the one hand, we've tried to stick with a consistent version of hwloc 
through an entire version series.

But on the other hand, hwloc is wholly internal and shouldn't be visible to 
apps.  So it *might* be harmless to upgrade it.

The only real question is: will upgrading hwloc break anything else inside the 
v1.8 tree?  E.g., did new hwloc abstractions/APIs come in after v1.7 that we've 
adapted to on the trunk, but didn't adapt to on the v1.8 branch?



On Dec 15, 2014, at 10:35 AM, Ralph Castain <r...@open-mpi.org> wrote:

> Sorry, I should have been clearer - that was indeed what I was expecting to 
> see. I guess it begs the question - should we just update to something like 
> 1.9 so Brice doesn't have to worry about back porting future fixes this far 
> back?
> 
> 
> 
> On Mon, Dec 15, 2014 at 7:22 AM, Jeff Squyres (jsquyres) <jsquy...@cisco.com> 
> wrote:
> FWIW, if it would be easier, we can just pull a new hwloc tarball -- that's 
> how we've done it in the past (vs. trying to pull individual patches).  It's 
> also easier to pull a release tarball, because then we can say "hwloc vX.Y.Z 
> is in OMPI vA.B.C", rather than have to try to examine/explain what exact 
> level of hwloc is in OMPI (based on patches, etc.).
> 
> 
> On Dec 15, 2014, at 4:39 AM, Brice Goglin <brice.gog...@inria.fr> wrote:
> 
> > Le 15/12/2014 10:35, Jorge D'Elia a écrit :
> >> Hi Brice,
> >>
> >> ----- Mensaje original -----
> >>> De: "Brice Goglin" <brice.gog...@inria.fr>
> >>> CC: "Open MPI Users" <us...@open-mpi.org>
> >>> Enviado: Jueves, 11 de Diciembre 2014 19:46:44
> >>> Asunto: Re: [OMPI users] OpenMPI 1.8.4 and hwloc in Fedora 14 using a 
> >>> beta gcc 5.0 compiler.
> >>>
> >>> This problem was fixed in hwloc upstream recently.
> >>>
> >>> https://github.com/open-mpi/hwloc/commit/790aa2e1e62be6b4f37622959de9ce3766ebc57e
> >> Great! However, yesterday I downloaded the versions 1.8.3 (stable) and
> >> 1.8.4rc3 of OpenMPI, and tried to use its more traditional configuration.
> >> It was OK on ia64 (as before) but failed again on ia32.  Then again,
> >> I had to use the external installation of hwloc in order to fix it.
> >>
> >
> > It's fixed in "upstream hwloc", not in OMPI yet. I have prepared a long
> > branch of hwloc fixes that OMPI should pull, but it will take some time.
> > thanks
> > Brice
> >
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> > Link to this post: 
> > http://www.open-mpi.org/community/lists/users/2014/12/25995.php
> 
> 
> --
> Jeff Squyres
> jsquy...@cisco.com
> For corporate legal information go to: 
> http://www.cisco.com/web/about/doing_business/legal/cri/
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2014/12/25996.php
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2014/12/25998.php


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/

Reply via email to