[ofa-general] Re: [Bug 506] IPoIB IPv4 multicast throughput is poor
Or Gerlitz
ogerlitz at voltaire.com
Wed Apr 4 02:20:20 PDT 2007
Michael S. Tsirkin wrote:
>> Quoting Or Gerlitz <ogerlitz at voltaire.com>:
>> Subject: Re: [Bug 506] IPoIB IPv4 multicast throughput is poor
>>
>> Michael S. Tsirkin wrote:
>>>> The low throughput is a major issue, though. Shouldn't the IP multicast
>>>> throughput be similar to the UDP unicast throughput?
>>> Is the send side a send only member of multicast group, or full member?
>> The join state (full / sendonly nonmember / nonmember)is something
>> communicated between the ULP through the ib_sa module and the IB SA.
>> I don't see how the host ib driver becomes aware to it.
>>
>> The current ipoib implementation for sendonly joins is to join as full
>> member but not to attach its UD QP for that group.
>
> I think so too. So what does the test do? Is it a sendonly join?
on the client side, when running iperf with -cu ipv4-multicast-address
iperf just send packets to that destination, my understanding is that
ipoib xmit calls ipoib_mcast_send which sense its a sendonly join etc.
on the server side, when running iperf with -suB ipv4-multicast-address
iperf issues an IP_ADD_MEMBERSHIP setsockopt call on its socket and the
kernel uses ip_ib_mc_map to compute the L2 multicast address and then
call the ipoib device set_multicast_list function which initiates a full
join + attach to this MGID.
>>> If it's a full join, HCA creates extra loopback traffic which
>>> has then to be discarded, and which might explain performance degradation.
>> Can you explain what --is-- the trigger for the "looback channel"
>> creation? my thinking it should be conditioned on having any QP attached
>> to this MGID, which does not seem the case in this scenario.
> That's what I'd expect.
Is this documented anywhere (eg Mellanox PRM)?
Or.
More information about the general
mailing list