[Users] IB topology config and polling state

German Anders ganders at despegar.com
Tue Oct 13 06:13:34 PDT 2015


I've try that and in fact I try to put the IP ADDR like this:

SET EBIPA INTERCONNECT XX.XX.XX.XX XXX.XXX.XXX.XXX 3
SET EBIPA INTERCONNECT GATEWAY XX.XX.XX.XX 3
SET EBIPA INTERCONNECT DOMAIN "xxxxxx.net" 3
ADD EBIPA INTERCONNECT DNS 10.xx.xx.xx 3
ADD EBIPA INTERCONNECT DNS 10.xx.xx.xx 3
SET EBIPA INTERCONNECT NTP PRIMARY NONE 3
SET EBIPA INTERCONNECT NTP SECONDARY NONE 3
ENABLE EBIPA INTERCONNECT 3

SAVE EBIPA

But i'm not getting any ip response, also I've try many diff ip addr
with no luck...if i put that ip to one of the blades it works fine,
but not to the interconnect bay :( any other idea?

Cheers,


*German* <ganders at despegar.com>

2015-10-13 10:01 GMT-03:00 Hal Rosenstock <hal.rosenstock at gmail.com>:

> Looks like there are 3 critical errors in system status. Did you look at
> these ?
>
> I don't know if you've seen this but there is some info on configuring the
> management IPs in http://h10032.www1.hp.com/ctg/Manual/c00814176.pdf
>
> Have you looked at/tried the command line interface ?
>
> On Tue, Oct 13, 2015 at 8:28 AM, German Anders <ganders at despegar.com>
> wrote:
>
>> Hi Hal,
>>
>> It does not allow me to setup an IP ADDR to the Internal SW so I can't
>> access from outside, except from the tools that I mentioned before, also it
>> doesn't allow me to access through serial connection from inside the
>> enclosure. I've attach some screen-shots about the connectivity.
>>
>>
>>
>> *German*
>>
>> 2015-10-13 9:13 GMT-03:00 Hal Rosenstock <hal.rosenstock at gmail.com>:
>>
>>> Hi German,
>>>
>>> Are the cards in the correct bays and slots ?
>>>
>>> Do you have the HP Onboard Administrator tool ? What does it say about
>>> internal connectivity ?
>>>
>>> -- Hal
>>>
>>>
>>>
>>> On Tue, Oct 13, 2015 at 7:44 AM, German Anders <ganders at despegar.com>
>>> wrote:
>>>
>>>> Hi Ira,
>>>>
>>>> I've some HP documentation but it quite short, also it doesn't describe
>>>> any 'config' or 'impl' steps in order to get the internal switch up and
>>>> running. The version of SW that came with enclosure does not had any
>>>> management module at all, so it depends on external management. During
>>>> weekend I've found a way to upgrade the firmware of the HP switch with the
>>>> following command (mlxburn -d lid-0x001D -fw fw-IS4.mlx), and the I've run
>>>> (flint -d /dev/mst/SW_MT48438_0x2c902004b0918_lid-0x001D dc
>>>> /home/ceph/HPIBSW.INI) and found the following inside that file:
>>>>
>>>> [PS_INFO]
>>>> Name = 489184-B21
>>>> Description = HP BLc 4X QDR IB Switch
>>>>
>>>> [ADAPTER]
>>>> PSID = HP_0100000009
>>>>
>>>> (...)
>>>>
>>>> [IB_TO_HW_MAP]
>>>> PORT1=14
>>>> PORT2=15
>>>> PORT3=16
>>>> PORT4=17
>>>> PORT5=18
>>>> PORT6=12
>>>> PORT7=11
>>>> PORT8=10
>>>> PORT9=9
>>>> PORT10=8
>>>> PORT11=7
>>>> PORT12=6
>>>> PORT13=5
>>>> PORT14=4
>>>> PORT15=3
>>>> PORT16=2
>>>> PORT17=20
>>>> PORT18=22
>>>> PORT19=24
>>>>
>>>> PORT20=26
>>>> PORT21=28
>>>> PORT22=30
>>>> PORT23=35
>>>> PORT24=33
>>>> PORT25=21
>>>> PORT26=23
>>>> PORT27=25
>>>> PORT28=27
>>>> PORT29=29
>>>> PORT30=36
>>>> PORT31=34
>>>> PORT32=32
>>>> PORT33=1
>>>> PORT34=13
>>>> PORT35=19
>>>> PORT36=31
>>>>
>>>> [unused_ports]
>>>> hw_port1_not_in_use=1
>>>> hw_port13_not_in_use=1
>>>> hw_port19_not_in_use=1
>>>> hw_port31_not_in_use=1
>>>>
>>>> (...)
>>>>
>>>> I don't know if maybe there's some issue with the port mapping, anyone
>>>> had used this kind of switch?
>>>>
>>>> The summary of the problem is correct, the connectivity between the IB
>>>> network (MLNX switches/gw) and the HP IB switch is working since I was able
>>>> to upgrade the firmare of the switch and get information about it. But, the
>>>> connection between the mezzanine cards of the blades and the internal IB sw
>>>> enclosure is not working at all. Note, that if I go to the OA
>>>> administration of the enclosure I can see the 'green' ports mapping of each
>>>> of the blades and the interconnection switch, so I'm guessing that it
>>>> should be working.
>>>>
>>>> Regarding the questions:
>>>>
>>>> 1)      What type of switch is in the HP chassis?
>>>>
>>>>
>>>> *QLogic HP BLc 4X QDR IB Switch*
>>>>
>>>> *PSID = HP_0100000009*
>>>>
>>>> *Image type:   FS2*
>>>>
>>>> *FW ver:         7.4.3000*
>>>>
>>>> *Device ID:     48438*
>>>> *GUI:              0002c902004b0918*
>>>>
>>>> 2)      Do you have console access or http access to that switch?
>>>>
>>>> *No, since it didn't had any manage module mezzanine card inside the
>>>> switch, it only come with a i2c port. But, i can have access through the
>>>> mlxburn and flint tools from one host that's connected to the ib network
>>>> (outside the enclosure).*
>>>>
>>>> 3)      Does that switch have an SM in it?
>>>>
>>>> *No*
>>>>
>>>> 4)      What version of the kernel are you running with the qib cards?
>>>>
>>>> a.       I assume you are using the qib driver in that kernel.
>>>>
>>>> *Ubuntu 14.04.3 LTS - kernel 3.18.20-031820-generic*
>>>>
>>>>
>>>>
>>>> At some point Hal spoke of “LLR being a Mellanox thing”  Was that to
>>>> solve the problem of connecting the “HP switch” to the Mellanox switch?
>>>>
>>>>
>>>>
>>>> *No, since LLR is only supported between mlnx devices, the ISL are up
>>>> and working, since it's possible for me to query the switch*
>>>>
>>>>
>>>>
>>>> I would like it if you could verify that the
>>>>
>>>>
>>>>
>>>> /usr/sbin/truescale-serdes.cmds
>>>>
>>>>
>>>>
>>>> Is being run?
>>>>
>>>>
>>>> *When trying to run the command:*
>>>>
>>>>
>>>>
>>>> *# /usr/sbin/truescale-serdes.cmds/usr/sbin/truescale-serdes.cmds: 100:
>>>> /usr/sbin/truescale-serdes.cmds: Syntax error: "(" unexpected (expecting
>>>> "}")*
>>>>
>>>>
>>>>
>>>> Also what version of libipathverbs do you have?
>>>>
>>>>
>>>>
>>>>
>>>> *# rpm -qa | grep libipathverbslibipathverbs-1.3-1.x86_64*
>>>> Thanks in advance,
>>>>
>>>> Cheers,
>>>>
>>>>
>>>>
>>>> *German*
>>>> 2015-10-13 2:14 GMT-03:00 Weiny, Ira <ira.weiny at intel.com>:
>>>>
>>>>> German,
>>>>>
>>>>>
>>>>>
>>>>> Do you have any documentation on the HP blade system?  And the switch
>>>>> which is in that system?
>>>>>
>>>>>
>>>>>
>>>>> I have to admit I have not followed everything in this thread
>>>>> regarding your configuration but it seems like you have some mellanox
>>>>> switches connected into an HP chassis which has both a switch and blades
>>>>> with qib (Truescale) cards.
>>>>>
>>>>>
>>>>>
>>>>> The connection from the mellanox switch to the “HP chassis switch” is
>>>>> linkup (active) but the connections to the individual qib HCAs are not even
>>>>> linkup.
>>>>>
>>>>>
>>>>>
>>>>> Is that a correct summary of the problem?
>>>>>
>>>>>
>>>>>
>>>>> If so here are some questions:
>>>>>
>>>>>
>>>>>
>>>>> 1)      What type of switch is in the HP chassis?
>>>>>
>>>>> 2)      Do you have console access or http access to that switch?
>>>>>
>>>>> 3)      Does that switch have an SM in it?
>>>>>
>>>>> 4)      What version of the kernel are you running with the qib cards?
>>>>>
>>>>> a.       I assume you are using the qib driver in that kernel.
>>>>>
>>>>>
>>>>>
>>>>> At some point Hal spoke of “LLR being a Mellanox thing”  Was that to
>>>>> solve the problem of connecting the “HP switch” to the Mellanox switch?
>>>>>
>>>>>
>>>>>
>>>>> I would like it if you could verify that the
>>>>>
>>>>>
>>>>>
>>>>> /usr/sbin/truescale-serdes.cmds
>>>>>
>>>>>
>>>>>
>>>>> Is being run?
>>>>>
>>>>>
>>>>>
>>>>> Also what version of libipathverbs do you have?
>>>>>
>>>>>
>>>>>
>>>>> Ira
>>>>>
>>>>>
>>>>>
>>>>> *From:* users-bounces at lists.openfabrics.org [mailto:
>>>>> users-bounces at lists.openfabrics.org] *On Behalf Of *Weiny, Ira
>>>>> *Sent:* Wednesday, October 07, 2015 1:31 PM
>>>>> *To:* Hal Rosenstock; German Anders
>>>>>
>>>>> *Cc:* users at lists.openfabrics.org
>>>>> *Subject:* Re: [Users] IB topology config and polling state
>>>>>
>>>>>
>>>>>
>>>>> Agree with Hal here.
>>>>>
>>>>>
>>>>>
>>>>> I’m not familiar with those blades/switches.  I’ll ask around.
>>>>>
>>>>>
>>>>>
>>>>> Ira
>>>>>
>>>>>
>>>>>
>>>>> *From:* Hal Rosenstock [mailto:hal.rosenstock at gmail.com
>>>>> <hal.rosenstock at gmail.com>]
>>>>> *Sent:* Wednesday, October 07, 2015 1:26 PM
>>>>> *To:* German Anders
>>>>> *Cc:* Weiny, Ira; users at lists.openfabrics.org
>>>>> *Subject:* Re: [Users] IB topology config and polling state
>>>>>
>>>>>
>>>>>
>>>>> That's the gateway to the switch in the enclosure. It's the internal
>>>>> connectivity in the blade enclosure that's (physically) broken.
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Oct 7, 2015 at 4:24 PM, German Anders <ganders at despegar.com>
>>>>> wrote:
>>>>>
>>>>> cabled
>>>>>
>>>>> the blade it's:
>>>>>
>>>>> vendid=0x2c9
>>>>> devid=0xbd36
>>>>> sysimgguid=0x2c902004b0918
>>>>> switchguid=0x2c902004b0918(2c902004b0918)
>>>>> Switch    32 "S-0002c902004b0918"        # "Infiniscale-IV Mellanox
>>>>> Technologies" base port 0 *lid 29* lmc 0
>>>>> [1]    "S-e41d2d030031e9c1"[9]        # "MF0;GWIB01:SX6036G/U1" lid 24
>>>>> 4xQDR
>>>>>
>>>>>
>>>>> *German*
>>>>>
>>>>>
>>>>>
>>>>> 2015-10-07 17:21 GMT-03:00 Hal Rosenstock <hal.rosenstock at gmail.com>:
>>>>>
>>>>> What are those HCAs cabled to or is it internal to the blade enclosure
>>>>> ?
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Oct 7, 2015 at 3:24 PM, German Anders <ganders at despegar.com>
>>>>> wrote:
>>>>>
>>>>> Yeah, is there any command that I can run in order to change the port
>>>>> state on the remote switch? I mean everything looks good but in the hp
>>>>> blades still getting:
>>>>>
>>>>>
>>>>>
>>>>> # ibstat
>>>>> CA 'qib0'
>>>>>     CA type: InfiniPath_QMH7342
>>>>>     Number of ports: 2
>>>>>     Firmware version:
>>>>>     Hardware version: 2
>>>>>     Node GUID: 0x0011750000791fec
>>>>>     System image GUID: 0x0011750000791fec
>>>>>     Port 1:
>>>>>         State: *Down*
>>>>>         Physical state: *Polling*
>>>>>         Rate: 40
>>>>>         Base lid: 4660
>>>>>         LMC: 0
>>>>>         SM lid: 4660
>>>>>         Capability mask: 0x0761086a
>>>>>         Port GUID: 0x0011750000791fec
>>>>>         Link layer: InfiniBand
>>>>>     Port 2:
>>>>>         State: *Down*
>>>>>         Physical state: *Polling*
>>>>>         Rate: 40
>>>>>         Base lid: 4660
>>>>>         LMC: 0
>>>>>         SM lid: 4660
>>>>>         Capability mask: 0x0761086a
>>>>>         Port GUID: 0x0011750000791fed
>>>>>         Link layer: InfiniBand
>>>>>
>>>>> Also on working hosts I only see devices from the local network, but
>>>>> didn't see any of the blades hca connections.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> *German*
>>>>>
>>>>>
>>>>>
>>>>> 2015-10-07 16:21 GMT-03:00 Hal Rosenstock <hal.rosenstock at gmail.com>:
>>>>>
>>>>> The screen shot looks good :-) SM brought the link up to active.
>>>>>
>>>>>
>>>>>
>>>>> Note that the ibportstate command you gave was for switch port 0 of
>>>>> the Mellanox IS-4 switch in the QLogic HP BLc 4X QDR IB Switch.
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Oct 7, 2015 at 3:06 PM, German Anders <ganders at despegar.com>
>>>>> wrote:
>>>>>
>>>>> Yes, find attached an screenshot of the port information (# 9) the one
>>>>> that makes the ISL to the QLogic HP BLc 4X QDR IB Switch, also from one of
>>>>> the hosts that are connected to one of the SX6018F I can see the 'remote'
>>>>> HP IB SW:
>>>>>
>>>>> # *ibnodes*
>>>>>
>>>>> (...)
>>>>> Switch    : 0x0002c902004b0918 ports 32 "Infiniscale-IV Mellanox
>>>>> Technologies" base port 0 *lid 29* lmc 0
>>>>> Switch    : 0xe41d2d030031e9c1 ports 37 "MF0;GWIB01:SX6036G/U1"
>>>>> enhanced port 0 lid 24 lmc 0
>>>>> (...)
>>>>>
>>>>> # *ibportstate -L 29 query*
>>>>> Switch PortInfo:
>>>>> # Port info: Lid 29 port 0
>>>>> LinkState:.......................Active
>>>>> PhysLinkState:...................LinkUp
>>>>> Lid:.............................29
>>>>> SMLid:...........................2
>>>>> LMC:.............................0
>>>>> LinkWidthSupported:..............1X or 4X
>>>>> LinkWidthEnabled:................1X or 4X
>>>>> LinkWidthActive:.................4X
>>>>> LinkSpeedSupported:..............2.5 Gbps or 5.0 Gbps or 10.0 Gbps
>>>>> LinkSpeedEnabled:................2.5 Gbps or 5.0 Gbps or 10.0 Gbps
>>>>> LinkSpeedActive:.................10.0 Gbps
>>>>> Mkey:............................<not displayed>
>>>>> MkeyLeasePeriod:.................0
>>>>> ProtectBits:.....................0
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> *German*
>>>>>
>>>>>
>>>>>
>>>>> 2015-10-07 16:00 GMT-03:00 Hal Rosenstock <hal.rosenstock at gmail.com>:
>>>>>
>>>>> One more thing hopefully before playing with the low level phy
>>>>> settings:
>>>>>
>>>>>
>>>>>
>>>>> Are you using known good cables ? Do you have FDR cables on the FDR
>>>>> <-> FDR links ? Cable lengths can matter as well.
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Oct 7, 2015 at 12:57 PM, Hal Rosenstock <
>>>>> hal.rosenstock at gmail.com> wrote:
>>>>>
>>>>> Were the ports mapped to the phy profile shutdown when you changed
>>>>> this ?
>>>>>
>>>>>
>>>>>
>>>>> LLR is a proprietary Mellanox mechanism.
>>>>>
>>>>>
>>>>>
>>>>> You might want 2 different profiles: one for the interfaces connected
>>>>> to other gateway interfaces (which are FDR (and FDR-10) capable and the
>>>>> other for the interfaces connecting to QDR (the older equipment in your
>>>>> network). By configuring the Switch-X interfaces to the appropriate
>>>>> possible speeds and disabling the proprietary mechanisms there, the link
>>>>> should not only come up but also this will occur faster than if FDR/FDR10
>>>>> are enabled.
>>>>>
>>>>>
>>>>>
>>>>> I suspect that due to the Switch-X configuration that the links to
>>>>> the switch(es) in the HP enclosures do not negotiate properly (as shown by
>>>>> down rather than LinkUp).
>>>>>
>>>>>
>>>>>
>>>>> Once you get all your links to INIT, negotiation has occurred and then
>>>>> it's time for SM to bring links to active.
>>>>>
>>>>>
>>>>>
>>>>> Since you have down links, the SM can't do anything about those.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Oct 7, 2015 at 12:44 PM, German Anders <ganders at despegar.com>
>>>>> wrote:
>>>>>
>>>>> Anyone had any experience with HP BLc 4X QDR IB Switch?? I know that
>>>>> this kind of SW does not come with an embedded sm, but I don't know how to
>>>>> access any mgmt at all on this particularly switch, I mean for example to
>>>>> setup speed or anything like that, is possible to access through the
>>>>> chassis?
>>>>>
>>>>>
>>>>> *German*
>>>>>
>>>>>
>>>>>
>>>>> 2015-10-07 13:19 GMT-03:00 German Anders <ganders at despegar.com>:
>>>>>
>>>>> I think so, but when trying to configured the phy-profile on the
>>>>> interface in order to negotiate on QDR it failed to map the profile:
>>>>>
>>>>> GWIB01 [proxy-ha-group: master] (config) # show phy-profile
>>>>> high-speed-ber
>>>>>
>>>>>   Profile: high-speed-ber
>>>>>   --------
>>>>>   llr support ib-speed
>>>>>   SDR: disable
>>>>>   DDR: disable
>>>>>   QDR: disable
>>>>>   FDR10: enable-request
>>>>>   FDR: enable-request
>>>>>
>>>>> GWIB01 [proxy-ha-group: master] (config) # show phy-profile
>>>>> hp-encl-isl
>>>>>
>>>>>   Profile: hp-encl-isl
>>>>>   --------
>>>>>   llr support ib-speed
>>>>>   SDR: disable
>>>>>   DDR: disable
>>>>>   QDR: enable
>>>>>   FDR10: enable-request
>>>>>   FDR: enable-request
>>>>>
>>>>> GWIB01 [proxy-ha-group: master] (config) #
>>>>> GWIB01 [proxy-ha-group: master] (config) # interface ib 1/9
>>>>> phy-profile map hp-encl-isl
>>>>> *% Cannot map profile hp-encl-isl to port:  1/9*
>>>>>
>>>>>
>>>>> *German*
>>>>>
>>>>>
>>>>>
>>>>> 2015-10-07 13:17 GMT-03:00 Weiny, Ira <ira.weiny at intel.com>:
>>>>>
>>>>> The driver ‘qib’ is loading fine.  As can be seen by the ibstat
>>>>> output.  The ib_ipath is an older card.
>>>>>
>>>>>
>>>>>
>>>>> The problem is the link is not coming up to init.  Like Hal said the
>>>>> link should transition to “link up” without the SMs involvement.
>>>>>
>>>>>
>>>>>
>>>>> I think you are on to something with the fact that it seems like your
>>>>> switch ports are not configured to do QDR.
>>>>>
>>>>>
>>>>>
>>>>> Ira
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> *From:* German Anders [mailto:ganders at despegar.com]
>>>>> *Sent:* Wednesday, October 07, 2015 9:05 AM
>>>>> *To:* Weiny, Ira
>>>>> *Cc:* Hal Rosenstock; users at lists.openfabrics.org
>>>>>
>>>>>
>>>>> *Subject:* Re: [Users] IB topology config and polling state
>>>>>
>>>>>
>>>>>
>>>>> Yes I've that file:
>>>>>
>>>>> /usr/sbin/truescale-serdes.cmds
>>>>>
>>>>> Also I've done the install of libipathverbs:
>>>>>
>>>>> # apt-get install libipathverbs-dev
>>>>>
>>>>> But I try to load the ib_ipath module but I'm getting the following
>>>>> error msg:
>>>>>
>>>>> # modprobe ib_ipath
>>>>> modprobe: ERROR: could not insert 'ib_ipath': Device or resource busy
>>>>>
>>>>>
>>>>> *German*
>>>>>
>>>>>
>>>>>
>>>>> 2015-10-07 12:54 GMT-03:00 Weiny, Ira <ira.weiny at intel.com>:
>>>>>
>>>>> There are a few issues for routing in that diagram but the links
>>>>> should come up.
>>>>>
>>>>>
>>>>>
>>>>> I assume there is some backplane between the blade servers and the
>>>>> switch in that chassis?
>>>>>
>>>>>
>>>>>
>>>>> Have you gotten libipathverbs installed?
>>>>>
>>>>>
>>>>>
>>>>> In ipathverbs there is a serdes tuning script.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> https://github.com/01org/libipathverbs/blob/master/truescale-serdes.cmds
>>>>>
>>>>>
>>>>>
>>>>> Does your libipathverbs include that file?  If not try the latest from
>>>>> github.
>>>>>
>>>>>
>>>>>
>>>>> Ira
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> *From:* users-bounces at lists.openfabrics.org [mailto:
>>>>> users-bounces at lists.openfabrics.org] *On Behalf Of *German Anders
>>>>> *Sent:* Wednesday, October 07, 2015 8:41 AM
>>>>> *To:* Hal Rosenstock
>>>>> *Cc:* users at lists.openfabrics.org
>>>>> *Subject:* Re: [Users] IB topology config and polling state
>>>>>
>>>>>
>>>>>
>>>>> Hi Hal,
>>>>>
>>>>> Thanks for the reply, I've attach a pdf with the diagram topology, I
>>>>> don't know if this is the best way to go or if there's another way to
>>>>> connect and setup the IB network, tips and suggestions will be very
>>>>> appreciated, also the mezzanine cards are already installed on the blade
>>>>> hosts:
>>>>>
>>>>> # lspci
>>>>> (...)
>>>>> 41:00.0 InfiniBand: QLogic Corp. IBA7322 QDR InfiniBand HCA (rev 02)
>>>>>
>>>>>
>>>>>
>>>>> Thanks in advance,
>>>>>
>>>>> Cheers,
>>>>>
>>>>>
>>>>> *German*
>>>>>
>>>>>
>>>>>
>>>>> 2015-10-07 11:47 GMT-03:00 Hal Rosenstock <hal.rosenstock at gmail.com>:
>>>>>
>>>>> Hi again German,
>>>>>
>>>>>
>>>>>
>>>>> Looks like you made some progress from yesterday as the qib ports are
>>>>> now Polling rather than Disabled.
>>>>>
>>>>>
>>>>>
>>>>> But since they are Down, do you have them cabled to a switch ? That
>>>>> should bring the links up and the port state will be Init. That is the
>>>>> "starting" point.
>>>>>
>>>>>
>>>>>
>>>>> You will also then need to be running SM to bring the ports up to
>>>>> Active.
>>>>>
>>>>>
>>>>>
>>>>> -- Hal
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Oct 7, 2015 at 10:37 AM, German Anders <ganders at despegar.com>
>>>>> wrote:
>>>>>
>>>>> Hi all,
>>>>>
>>>>> I don't know if this is the mailist list for this kind of topic but
>>>>> I'm really new to IB and I've just install two SX6036G gateways connected
>>>>> to each other through two ISL ports, then I've configured a proxy-arp
>>>>> between both nodes (sm is disable on both gw's):
>>>>>
>>>>> GWIB01 [proxy-ha-group: master] (config) # show proxy-arp ha
>>>>>
>>>>> Load balancing algorithm: ib-base-ip
>>>>> Number of Proxy-Arp interfaces: 1
>>>>>
>>>>> Proxy-ARP VIP
>>>>> =============
>>>>> Pra-group name: proxy-ha-group
>>>>> HA VIP address: 10.xx.xx.xx/xx
>>>>>
>>>>> Active nodes:
>>>>> ID                   State                IP
>>>>> --------------------------------------------------------------
>>>>> GWIB01               master               10.xx.xx.xx1
>>>>> GWIB02               standby              10.xx.xx.xx2
>>>>>
>>>>> Then I setup two SX6018F switches (*SWIB01* and *SWIB02*), one
>>>>> connected to GWIB01 and the other connected to GWIB02. The SM is configured
>>>>> locally on both SWIB01 & SWIB02 switches. So far so good, after this config
>>>>> I setup a commodity server with a MLNX IB ADPT FDR to the SWIB01 & SWIB02
>>>>> switches, config the drivers, etc and then get it up & running fine.
>>>>>
>>>>> Finally I've setup a HP Enclosure with an internal IB SW (then connect
>>>>> port 1 of the internal SW to GWIB01 - link is up but LLR status is
>>>>> inactive), install one of the blades and I see the following:
>>>>>
>>>>> # ibstat
>>>>> CA 'qib0'
>>>>>     CA type: InfiniPath_QMH7342
>>>>>     Number of ports: 2
>>>>>     Firmware version:
>>>>>     Hardware version: 2
>>>>>     Node GUID: 0x0011750000791fec
>>>>>     System image GUID: 0x0011750000791fec
>>>>>     Port 1:
>>>>>         State: Down
>>>>>         Physical state: Polling
>>>>>         Rate: 40
>>>>>         Base lid: 4660
>>>>>         LMC: 0
>>>>>         SM lid: 4660
>>>>>         Capability mask: 0x0761086a
>>>>>         Port GUID: 0x0011750000791fec
>>>>>         Link layer: InfiniBand
>>>>>     Port 2:
>>>>>         State: Down
>>>>>         Physical state: Polling
>>>>>         Rate: 40
>>>>>         Base lid: 4660
>>>>>         LMC: 0
>>>>>         SM lid: 4660
>>>>>         Capability mask: 0x0761086a
>>>>>         Port GUID: 0x0011750000791fed
>>>>>         Link layer: InfiniBand
>>>>>
>>>>> So I was wondering if maybe the SM is not being recognized on the
>>>>> Blade system and that's why is not passing the Polling state, is that
>>>>> possible? Or maybe is not possible to connect an ISL between the GW and the
>>>>> HP internal SW so that the sm is available or maybe the inactive LLR is
>>>>> causing this thing, any ideas? I thought about connecting the ISL of
>>>>> the HP IB SW to the SWIB01 or SWIB02 instead of the GW's but I don't have
>>>>> any available ports.
>>>>>
>>>>> Thanks in advance,
>>>>>
>>>>> Cheers,
>>>>>
>>>>>
>>>>> *German*
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list
>>>>> Users at lists.openfabrics.org
>>>>> http://lists.openfabrics.org/mailman/listinfo/users
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openfabrics.org/pipermail/users/attachments/20151013/42a25f2d/attachment.html>


More information about the Users mailing list