[ofa-general] IB performance stats (revisited)
Eitan Zahavi
eitan at mellanox.co.il
Wed Jul 11 08:30:04 PDT 2007
Hi Marc,
I wish I had a large enough fabric worth testing collectl on...
I did the math for how much data would be collected for 10Knodes
cluster. It is ~7MB for each iteration:
10K ports
* 6 (3 level fabric * 2 ports on each link)
* 32 byte (data/pkts tx/rx) + 22byte (err counters) + 64byte (cong
counters) = 116bytes
Seems reasonable - but adds up to large amount of data over a day period
assuming a collect every second:
24*60*60 *116*10000*6 = 6.01344e+11 Bytes of storage
Eitan Zahavi
Senior Engineering Director, Software Architect
Mellanox Technologies LTD
Tel:+972-4-9097208
Fax:+972-4-9593245
P.O. Box 586 Yokneam 20692 ISRAEL
> -----Original Message-----
> From: Mark Seger [mailto:Mark.Seger at hp.com]
> Sent: Wednesday, July 11, 2007 5:51 PM
> To: Eitan Zahavi
> Cc: Hal Rosenstock; Ira Weiny; general at lists.openfabrics.org;
> Ed.Finn at FMR.COM
> Subject: Re: [ofa-general] IB performance stats (revisited)
>
>
>
> Eitan Zahavi wrote:
>
> >Hi Marc,
> >
> >I published an RFC and later had discussions regarding the
> distribution
> >of query ownership of switch counters.
> >Making this ownership purely dynamic, semi-dynamic or even
> static is an
> >implementation tradeoff.
> >However, it can be shown that the maximal number of switches
> a single
> >compute node would be responsible for is <= number of switch
> levels. So
> >no problem to get counters every second...
> >
> >The issue is: what do you do with the size of data collected?
> >This is only relevant if monitoring is run in "profiling mode"
> >otherwise only link health errors should be reported.
> >
> >
> I use IB data for performance data typically for
> system/application diagnostics. I run a tool I wrote (see
> http://sourceforge.net/projects/collectl/) as a service on
> most systems and it gathers well over hundreds of performance
> metrics/counters on everything from cpu load, memory,
> network, infiniband, disk, etc. The philosophy here is that
> if something goes wrong, it may be too late to then run some
> diagnostic. Rather you need to have already collected the
> data, especially if this is an intemittent problem. When
> there is no need to look at the data, it just gets purged
> away after a week.
>
> There have been situation where someone reports a batch
> program they ran the other day was really slow and they
> didn't change anything. By being able to pull up a
> monitoring log and seeing what the system was doing at the
> time of the run might reveal their network was saturated and
> therefore their MPI job was impacted. You can't very well
> turn on diagnostics and rerun the application because system
> conditions have probably changed.
>
> Does that help? Why don't you try installing collectl and
> see what it does...
>
> -mark
>
>
>
More information about the general
mailing list