Requirements for stretched systems with Fibre Channel connections
If you are configuring a stretched system that uses Fibre Channel connections, ensure
that all SAN and Fibre Channel-specific requirements are met.
Notes:
Stretched system Fibre Channel configurations with active/passive controllers such as IBM®
DS5000, IBM
DS4000®, and IBM DS3000 systems must be configured with
sufficient connections such that all sites have direct access to both external storage systems. For
iSCSI configurations with two or more active/passive controllers such as Storwize® family systems, the systems must be
configured with sufficient connections such that all sites have direct access to both external storage
systems. Quorum access for stretched system is possible only through the current owner of the MDisk that
is being used as the active quorum disk.
Use the following requirements to configure a stretched system with Fibre Channel
connections:
Directly connect each node to two or more SAN fabrics at the primary and secondary sites (2 - 8
fabrics are supported). In iSCSI connections, connect each node to two or more Ethernet fabrics at the
primary and secondary sites. Sites are defined as independent failure domains. A failure domain is a
part of the system within a boundary such that any failure (such as a power failure, fire, or flood)
within that boundary is contained within the boundary and the failure does not propagate or affect parts
outside of that boundary. Failure domains can be in the same room or across rooms in the data center,
buildings on the same campus, or buildings in different towns. Different kinds of failure domains
protect against different types of faults.
If a storage system
is used at the third site, it must support extended quorum disks. More information is available in the
interoperability matrixes that are available at the following website:
Place independent storage systems at the primary and secondary sites, and use volume mirroring to
mirror the host data between storage systems at the two sites. Where possible, set the preferred node of
each volume to the node in the same site as the host that the volume is mapped to.
Connections can vary based on fibre type and small form-factor
pluggable (SFP) transceiver (longwave and shortwave).
Nodes that are in the same I/O group and separated by more than
100 meters (109 yards) must use longwave Fibre Channel or Ethernet connections. A longwave small form-factor
pluggable (SFP) transceiver can be purchased as an optional
component, and must be one of the longwave SFP transceivers listed at the following website:
Avoid using inter-switch links (ISLs) in paths between nodes and external storage systems. If this is
unavoidable, do not oversubscribe the ISLs because of substantial Fibre Channel traffic across the ISLs.
For most configurations, trunking is required. Because ISL problems are difficult to diagnose,
switch-port error statistics must be collected and regularly monitored to detect failures.
Using a single switch at the third site can lead to the creation of a single fabric rather than two
independent and redundant fabrics. A single fabric is an unsupported configuration.
Ethernet port 1 on every node must be connected to the same subnet or subnets. Ethernet port 2 (if
used) of every node must be connected to the same subnet (this might be a different subnet from port 1).
The same principle applies to other Ethernet ports.
Some service actions require physical access to all nodes in a system. If nodes in a stretched system
are separated by more than 100 meters, service actions might require multiple service personnel. Contact
your service representative to inquire about multiple site support.
Use a third, dedicated site to house a quorum disk or an IP quorum
application.
For Fibre Channel-connected stretched system, quorum disks or IP quorum applications provide redundancy
if communication is lost between the primary and secondary site. In addition both contain configuration
metadata that is used to recover the system, if necessary. IP quorum applications are used for when the
stretched system connects to iSCSI-attached storage systems. iSCSI storage systems cannot be configured
on a third site.
A stretched system locates the active quorum disk or an IP quorum application at a
third site. If communication is lost between the primary and secondary sites, the site with access to the
active quorum disk continues to process transactions. If communication is lost to the active quorum disk,
an alternative quorum disk at another site can become the active quorum disk.
Although a system of nodes can be configured to use up to three quorum disks, only one
quorum disk can be elected to resolve a situation where the system is partitioned into two sets of nodes
of equal size. The purpose of the other quorum disks is to provide redundancy if a quorum disk fails
before the system is partitioned.
Figure 1 illustrates an example stretched
system configuration. When used with volume mirroring, this configuration provides a high
availability solution that is tolerant of a failure at a single site. If either the primary or
secondary site fails, the remaining sites can continue doing I/O operations. In this configuration,
the connections between
the nodes in the
system are greater than 100 meters apart, and therefore must be longwave Fibre Channel
connections.Figure 1. A stretched system with a quorum disk at a third site
In Figure 1, the storage system that hosts
the third-site quorum disk is attached directly to a switch at both the primary and secondary sites
by using longwave Fibre Channel connections. If either the primary site or the secondary site fails,
you must ensure that the remaining site retains direct access to the storage system that hosts the
quorum disks.
Restriction: Do not connect a storage system in one site directly to a
switch fabric in the other site.
An alternative configuration can use an additional Fibre Channel switch at the third site with
connections from that switch to the primary site and to the secondary site.
A stretched system configuration is supported only when the storage system that hosts the quorum
disks supports extended quorum. Although
other types of
storage systems can be used to provide quorum disks, access to these quorum disks is always through
a single path.
For quorum disk configuration requirements, see the technote Guidance for Identifying and
Changing Managed Disks Assigned as Quorum Disk Candidates.