[openib-general] [PATCH] osm: PathRecord prefer 1K MTU for MT23108 devices
Rimmer, Todd
trimmer at silverstorm.com
Mon Sep 18 08:52:18 PDT 2006
> From: Eitan Zahavi [mailto:eitan at mellanox.co.il]
> Sent: Monday, September 18, 2006 11:20 AM
> To: Rimmer, Todd
> Cc: Or Gerlitz; Michael S. Tsirkin; OPENIB
> Subject: Re: [openib-general] [PATCH] osm: PathRecord prefer 1K MTU
for
> MT23108 devices
>
> Hi Todd,
>
> Seems like your knowledge about the specific MTU best for the
> application (MPI) you are running is good
> enough such that you will be able to include the MTU in the PathRecord
> request and thus the patch describe in here will not affect your MPI
at
> all.
> The patch only applies if your request does not provide any MTU & MTU
> SEL comp_mask
Eitan,
The question is not about "our MPI", rather its to ensure the Open
Fabrics and OFED included MPIs and ULPs are capable of being tuned for
optimal performance. When a fabric runs more than 1 application, its
necessary to be able to tune this at the MPI, SDP, etc level, not at the
SM level.
This patch turns on a non-standard behaviour in the SM for the entire
fabric such that some applications will have better performance while
others will suffer. In order to be complete, this patch would need to
include ULP level tunability in all the relevant ULPs (MPI, SDP, uDAPL,
etc) to select the "MAX MTU" to use or to request.
This then begs the question, if proper tuning requires all the ULPs to
have a configurable MAX MTU, why should the SA need to implement the
quirk at all?
Todd Rimmer
> >
> >Putting this in the SM alone and making it a fabric wide setting is
> >inappropriate. The performance difference depends on application
> >message size. Application message size can vary per ULP and/or per
> >application itself. For example one MPI application may send mostly
> >large messages while another may send mostly small messages. The
same
> >could be true of applications for other ULPs such as uDAPL and SDP,
etc.
> >
> >The root issue is the Tavor HCA has 1 too few credits to truly double
> >buffer at 2K MTU. However at message sizes > 1K but < 2K the 2K MTU
> >performs better.
> >
> >Here are some MPI bandwidth results:
> >Tavor w/ 2K MTU:
> >512 140.394173
> >1024 310.553002
> >1500 407.003858
> >1800 435.538752
> >2048 392.831026
> >4096 417.592991
> >
> >Tavor w/ 1K MTU:
> >512 140.261964
> >1024 300.789425
> >1500 379.746835
> >1800 416.726957
> >2048 425.227096
> >4096 501.442289
> >
> >Note that message sizes shown on left do not include MPI headers.
Hence
> >actual IB message size is approx 50 bytes larger.
> >
> >So we see at IB message sizes < 1024 (MPI 512 message), performance
is
> >the same.
> >At IB message sizes > 1024 < 2048 (MPI 1024-1800 messages),
performance
> >is best with 2K MTU.
> >At IB message sizes > 2048 (MPI 2048-4096 messages above),
performance
> >is best with 1K MTU.
> >At larger IB message sizes (MPI 4096 message), performance starts to
> >take off and ultimately at 128K message size (not shown) the 50%
> >difference between 1K and 2K MTU reaches its peak.
> >
> >Todd Rimmer
> >
> >_______________________________________________
> >openib-general mailing list
> >openib-general at openib.org
> >http://openib.org/mailman/listinfo/openib-general
> >
> >To unsubscribe, please visit
http://openib.org/mailman/listinfo/openib-
> general
> >
> >
More information about the general
mailing list