[ofa-general] OFED 1.2 Feb-26 meeting summary

Jeff Squyres jsquyres at cisco.com
Fri Mar 2 17:42:22 PST 2007


To be totally clear, there are three issues:

1. *NOT AN MPI ISSUE*: base location of the stack.  Doug has  
repeatedly mentioned that /usr/local/ofed is not good.  This is a  
group issue to decide.

2. *NOT AN MPI ISSUE*: how the RPMs are built is Bad(tm).  Not  
deleting the buildroot is Bad; munging %build into %install is  
Bad; ...etc.  This needs to change.  4 choices jump to mind:

    a. Keep the same scheme.  Ick.
    b. Install while we build (i.e., the normal way to build a pile  
of interdependent RPMs)
    c. Use chroot (Red Hat does this in their internal setup, for  
example)
    d. Only distribute binary RPMs for supported platforms; source is  
available for those who want it.

3. Doug's final point about allowing multiple MPI's to play  
harmoniously on a single system is obviously an MPI issue.  The /etc/ 
alternatives mechanism is not really good enough (IMHO) for this -- / 
etc/alternatives is about choosing one implementation and making  
everyone use it.  The problem is that when multiple MPI's are  
installed on a single system, people need all of them (some users  
prefer one over the other, but much more important, some apps are  
only certified with one MPI or another).  The mpi-selector tool we  
introduced in OFED 1.2 will likely be "good enough" for this purpose,  
but we can also work on integrating the /etc/alternatives stuff if  
desired, particularly for those who only need/want one MPI  
implementation.


On Feb 28, 2007, at 12:56 PM, Doug Ledford wrote:

> On Wed, 2007-02-28 at 16:10 +0200, Tziporet Koren wrote:
>>       * Improved RPM usage by the install will not be part of OFED
>>         1.2
>
> Since I first brought this up, you have added new libraries, iWARP
> support, etc.  These constitute new RPMs.  And, because you guys have
> been doing things contrary to standards like the file hierarchy  
> standard
> in the original RPMs, it's been carried forward to these new RPMs.   
> This
> is a snowball, and the longer you put off fixing it, the harder it  
> gets
> to change.  And not just in your RPMs either.  The longer you put off
> coming up with a reasonable standard for MPI library and executable  
> file
> locations, the longer customers will hand roll their own site specific
> setups, and the harder it will be to get them to switch over to the
> standard once you *do* implement it.  You may end up dooming Jeff to
> maintaining those custom file location hacks in the OpenMPI spec
> forever.
>
> Not to mention that interoperability is about more than one machine
> talking to another machine.  It's also about a customer's application
> building properly on different versions of the stack, without the
> customer needing to change all the include file locations and link
> parameters.  It's also about a customer being able to rest assured  
> that
> if they tried to install two conflicting copies of libibverbs, it  
> would
> in fact cause RPM to throw conflict errors (which it doesn't now  
> because
> your libibverbs is in /usr/local, where I'm not allowed to put  
> ours, so
> since the files are in different locations, rpm will happily let the
> user install both your libibverbs and my libibverbs without a  
> conflict,
> and a customer could waste large amounts of time trying to track  
> down a
> bug in one library only to find out their application is linking  
> against
> the other).
>
>>               * The RPM usage will be enhanced for the next (1.3)
>>                 release and we will decide on the correct way in
>>                 Sonoma.
>
>
>
> There's not really much to decide.  Either the stack is Linux File
> Hierarchy Standard compliant or it isn't.  The only leeway for  
> decisions
> allowed by the standard is on things like where in /etc to put the
> config files (since you guys are striving to be a generic RDMA stack,
> not just an IB stack, I would suggest that all RDMA related config  
> files
> go into /etc/rdma, and for those applications that can reasonably  
> be run
> absent RDMA technology, like OpenMPI, I would separate their config
> files off into either /etc or /etc/openmpi, ditto for the include
> directories, /usr/include/rdma for the generic non-IB specific stuff,
> and possibly /usr/include/rdma/infiniband for IB specific stuff, or  
> you
> could put the IB stuff under /usr/include/infiniband, either way).
>
> The biggest variation from the spec that needs to be dealt with is the
> need for multiple MPI installations, which is problematic if you just
> use generic locations as it stands today, but with a few modifications
> to the MPI stack it could be worked around.
>
>
> -- 
> Doug Ledford <dledford at redhat.com>
>               GPG KeyID: CFBFF194
>               http://people.redhat.com/dledford
>
> Infiniband specific RPMs available at
>               http://people.redhat.com/dledford/Infiniband
> _______________________________________________
> general mailing list
> general at lists.openfabrics.org
> http://lists.openfabrics.org/cgi-bin/mailman/listinfo/general
>
> To unsubscribe, please visit http://openib.org/mailman/listinfo/ 
> openib-general


-- 
Jeff Squyres
Server Virtualization Business Unit
Cisco Systems





More information about the general mailing list