[ofa-general] SRP target/LVM HA configuration
Vu Pham
vuhuong at mellanox.com
Wed Mar 12 15:20:40 PDT 2008
I would second to Stanley - client hosts
fail-over/load-balancing would be straightforward
For linux host you have several tools: sw raid, lvm or
dm-multipath
-vu
> I had looked at this configuration as well and decided to use the volume
> management at the clients to mirror the data. Windows LDM mirrored
> across 2 SRPT servers and Linux md RAID 1 mirrored.
>
> This provides transparent failover and the SRP client/host will rebuild
> the slices that went offline.
>
>> -----Original Message-----
>> From: general-bounces at lists.openfabrics.org
>> [mailto:general-bounces at lists.openfabrics.org] On Behalf Of
>> Daniel Pocock
>> Sent: Tuesday, March 11, 2008 4:26 PM
>> To: general at lists.openfabrics.org
>> Subject: [ofa-general] SRP target/LVM HA configuration
>>
>>
>>
>>
>>
>>
>> I'm contemplating a HA configuration based on SRP and LVM (or
>> maybe EVMS).
>>
>> There are many good resources based on NFS and drbd (see
>> http://www.linux-ha.org/HaNFS) but it would be more flexible to work
>> with block level (e.g SRP) rather than file level (NFS). Obviously,
>> SRP/RDMA offers a major performance benefit compared with drbd (which
>> uses IP).
>>
>> Basically, I envisage the primary server having access to the
>> secondary
>> (passive) server's disk using SRP, and putting both the local
>> (primary)
>> disk and SRP (secondary) disk into RAID1. The RAID1 set
>> would contain a
>> volume group and multiple volumes - which would, in turn, be
>> SRP targets
>> (for VMware to use) or possibly NFS shares.
>>
>> This leads me to a few issues:
>>
>> - Read operations - would it be better for the primary to
>> read from both
>> disks, or just it's own disk? Using drbd, the secondary disk is not
>> read unless the primary is down. However, given the
>> performance of SRP,
>> I suspect that reading from both the local and SRP disk would give a
>> boost to performance.
>>
>> - Does it make sense to use md or LVM to combine a local disk
>> and an SRP
>> disk into RAID1 (or potentially RAID5)? Are there technical
>> challenges
>> there, given that one target is slightly faster than the other?
>>
>> - Fail-over - when the secondary detects that the primary is
>> down, can
>> it dynamically take the place of the failed SRP target? Will the
>> end-user initiators (e.g. VMWare, see diagram below) be confused when
>> the changeover occurs? Is there the possibility of data
>> inconsistency
>> if some write operations had been acknowledged by the primary but not
>> propagated to the secondary's disk at the moment when the
>> failure occurred?
>>
>> - Recovery - when the old primary comes back online as a
>> secondary, it
>> will need to resync it's disk - is a partial resync possible,
>> or is full
>> rebuild mandatory?
>>
>>
>> Diagram:
>>
>>
>> Disk--Primary Server-------------------SRP Initiator (e.g. VMware ESX)
>> | +------NFS client
>> | .
>> SRP .
>> (RAID1 of primary's .
>> disk and secondary's .
>> disk) . (fail-over path to storage
>> | . when primary is down)
>> Disk--Secondary Server. . . . . .
>>
>>
>>
>> _______________________________________________
>> general mailing list
>> general at lists.openfabrics.org
>> http://lists.openfabrics.org/cgi-bin/mailman/listinfo/general
>>
>> To unsubscribe, please visit
>> http://openib.org/mailman/listinfo/openib-general
>>
> _______________________________________________
> general mailing list
> general at lists.openfabrics.org
> http://lists.openfabrics.org/cgi-bin/mailman/listinfo/general
>
> To unsubscribe, please visit http://openib.org/mailman/listinfo/openib-general
More information about the general
mailing list