Reproduction of these materials in any manner whatsoever without the written permission of Dell Inc.
is strictly forbidden.
Trademarks used in this text: Dell, the DELL logo, PowerEdge, and PowerVault are trademarks of
Dell Inc.; Active Directory, Microsoft, Windows, Windows Server, and Windows NT are either
trademarks or registered trademarks of Microsoft Corporation in the United States and/or other
countries.; EMC, Navisphere, and PowerPath are registered trademarks and MirrorView, SAN Copy, and SnapView are trademarks of EMC Corporation.
Other trademarks and trade names may be used in this document to refer to either the entities
claiming the marks and names or their products. Dell Inc. disclaims any proprietary interest in
trademarks and trade names other than its own.
A failover cluster combines specific hardware and software components to
provide enhanced availability for applications and services that are run on the
cluster. A failover cluster is designed to reduce the possibility of any single
point of failure within the system that can cause the clustered applications or
services to become unavailable. It is recommended that you use redundant
components like server and storage power supplies, connections between the
nodes and the storage array(s), connections to client systems or other servers
in a multi-tier enterprise application architecture in your cluster.
This document provides information and specific configuration tasks that
enable you to configure your Microsoft
with Dell/EMC AX4-5F (2 Fibre Channel ports per Storage Processor) or
Dell/EMC AX4-5FX (4 Fibre Channel ports per Storage Processor) storage
array(s).
NOTE: Throughout this document, Dell/EMC AX4-5 refers to Dell/EMC AX4-5F and
Dell/EMC AX4-5FX storage arrays.
NOTE: Throughout this document, Windows Server 2008 refers to Windows Server
2008 and Windows Server 2008 R2 operating systems.
For more information on deploying your cluster with Windows Server 2003
operating systems, see the Dell™ Failover Clusters with Microsoft Windows Server 2003 Installation and Troubleshooting Guide on the Dell Support
website at support.dell.com/manuals. For more information on deploying
your cluster with Windows Server 2008 operating systems, see the Dell
Failover Clusters with Microsoft Windows Server 2008 Installation and
Troubleshooting Guide on the Dell Support website at
support.dell.com/manuals.
For a list of recommended operating systems, hardware components, and
driver or firmware versions for your Dell Windows Server Failover Cluster, see
the Dell Cluster Configuration Support Matrices located on the Dell High
Availability Clustering website at dell.com/ha.
®
Windows Server® failover cluster
Introduction7
Cluster Solution
Your cluster implements a minimum of two node to a maximum of either eight
nodes (for Windows Server 2003) or sixteen nodes (for Windows Server 2008)
clustering and provides the following features:
•8-Gbps and 4-Gbps Fibre Channel technologies
•High availability of resources to network clients
•Redundant paths to the shared storage
•Failure recovery for applications and services
•Flexible maintenance capabilities, allowing you to repair, maintain, or
upgrade a node or storage system without taking the entire cluster offline
Implementing Fibre Channel technology in a cluster provides the following
advantages:
•
Flexibility
switches without degrading the signal.
•
Availability
providing multiple data paths and greater availability for clients.
•
Connectivity
SCSI. Because Fibre Channel devices are hot-swappable, you can add or
remove devices from the nodes without bringing down the cluster.
— Fibre Channel allows a distance of up to 10 km between
— Fibre Channel components use redundant connections,
— Fibre Channel allows more device connections than
Cluster Hardware Requirements
Your cluster requires the following hardware components:
•Cluster nodes
•Cluster storage
8Introduction
Cluster Nodes
Table 1-1 lists the hardware requirements for the cluster nodes.
Table 1-1. Cluster Node Requirements
ComponentMinimum Requirement
Cluster nodesA minimum of two identical Dell™ PowerEdge™ servers are
required. The maximum number of nodes that is supported
depends on the variant of the Windows Server operating
system used in your cluster, and on the physical topology in
which the storage system and nodes are interconnected.
RAMThe variant of the Windows Server operating system that is
installed on your cluster nodes determines the minimum
required amount of system RAM.
HBA portsTwo Fibre Channel HBAs per node, unless the server employs
an integrated or supported dual-port Fibre Channel HBA.
Where possible, place the HBAs on separate PCI buses to
improve availability and performance.
NICs (public and
private networks)
Internal disk
controller
At least two NICs: one NIC for the public network and
another NIC for the private network.
NOTE: It is recommended that the NICs on each public network
are identical, and that the NICs on each private network are
identical.
One controller connected to at least two internal hard drives
for each node. Use any supported RAID controller or disk
controller.
Two hard drives are required for mirroring (RAID 1) and at
least three are required for disk striping with parity (RAID 5).
NOTE: It is strongly recommended that you use hardware-
based RAID or software-based disk-fault tolerance for the
internal drives.
NOTE: For more information about supported systems, HBAs and operating system
variants, see the Dell Cluster Configuration Support Matrices located on the Dell
High Availability Clustering website at dell.com/ha.
Introduction9
Cluster Storage
Cluster nodes can share access to external storage systems. However, only one
of the nodes can own any redundant array of independent disks (RAID)
volume in the external storage system at any time. Microsoft Cluster Services
(MSCS) controls which node has access to each RAID volume in the shared
storage system.
Table 1-2 lists supported storage systems and the configuration requirements
for the cluster nodes and stand-alone systems connected to the storage
systems.
Table 1-2. Cluster Storage Requirements
Hardware Components Requirement
Supported storage
systems
Cluster nodesAll nodes must be directly attached to a single storage
Multiple clusters and
stand-alone systems
One to four supported Dell/EMC storage systems. For
specific storage system requirements see Table 1-3.
system or attached to one or more storage systems
through a SAN.
Can share one or more supported storage systems using
optional software that is available for your storage system.
See "Installing and Configuring the Shared Storage
System" on page 41.
The storage systems work together with the following hardware components:
•Disk processor enclosure (DPE)—Configured with storage processors that
control the RAID arrays in the storage system and provide storage
functionalities such as snapshots, LUN masking, and remote mirroring.
•Disk array enclosure (DAE)—Provides additional storage and is attached
to the disk processor enclosure.
•Standby power supply (SPS)—Provides backup power to protect the
integrity of the disk processor write cache. The SPS is connected to the
disk processor enclosure
10Introduction
Table 1-3 lists hardware requirements for the disk processor enclosures DPE
DAE, and SPS.
Table 1-3. Dell/EMC Storage System Requirements
Storage
System
AX4-5 1 DPE with at least 4
NOTE: Ensure that the core software version running on the storage system is
supported by Dell. For specific version requirements, see the Dell Cluster Configuration Support Matrices located on the Dell High Availability Cluster website
at dell.com/ha.
Minimum Required
Storage
and up to 12 hard
drives
Possible Storage
Expansion
Up to 3 DAE with a
maximum of 12 harddrives each
SPS
First is required,
the second SPS is
optional
Supported Cluster Configurations
Direct-Attached Cluster
In a direct-attached cluster, all nodes of the cluster are directly attached to a
single storage system. In this configuration, the RAID controllers (or storage
processors) on the storage systems are connected by cables directly to the
Fibre Channel HBA ports in the nodes.
Figure 1-1 shows a basic direct-attached, single-cluster configuration.
EMC® PowerPath® Limitations in a Direct-Attached Cluster
EMC PowerPath provides failover capabilities and multiple path detection as
well as dynamic load balancing between multiple ports on the same storage
processor. However, direct-attached clusters supported by Dell connect to a
single port on each storage processor in the storage system. Because of the
single port limitation, PowerPath can provide only failover protection, not
load balancing, in a direct-attached configuration.
SAN-Attached Cluster
In a SAN-attached cluster, all of the nodes are attached to a single storage
system or to multiple storage systems through a SAN using redundant switch
fabrics. SAN-attached clusters are superior to direct-attached clusters in
configuration flexibility, expandability, and performance.
Figure 1-2 shows a SAN-attached cluster.
12Introduction
Figure 1-2. SAN-Attached Cluster
cluster node
cluster node
private network
Fibre Channel
connections
storage system
Fibre Channel
switch
Fibre Channel
switch
public network
Fibre Channel
connections
Other Documents You May Need
CAUTION: The safety information that is shipped with your system provides
important safety and regulatory information. Warranty information may be
included within this document or as a separate document.
•The
•The
•The
•The
NOTE: To configure Dell blade server modules in a Dell PowerEdge cluster, see the
Using Dell Blade Servers in a Dell PowerEdge High Availability Cluster document
located on the Dell Support website at support.dell.com/manuals.
Rack Installation Guide
how to install your system into a rack.
Getting Started Guide
system.
Dell Failover Clusters with Microsoft Windows Server 2003 Installation
and Troubleshooting Guide
cluster with the Windows Server 2003 operating system.
Dell Failover Clusters with Microsoft Windows Server 2008 Installation
and Troubleshooting Guide
cluster with the Windows Server 2008 operating system.
included with your rack solution describes
provides an overview of initially setting up your
provides more information on deploying your
provides more information on deploying your
Introduction13
•The
Dell Cluster Configuration Support Matrices
provides a list of
recommended operating systems, hardware components, and driver or
firmware versions for your Dell Windows Server Failover Cluster.
•The HBA documentation provides installation instructions for the HBAs.
•Systems management software documentation describes the features,
requirements, installation, and basic operation of the software.
•Operating system documentation describes how to install (if necessary),
configure, and use the operating system software.
•The Dell PowerVault™ tape library documentation provides information
for installing, troubleshooting, and upgrading the tape library.
•The EMC PowerPath documentation that came with your HBA kit(s) and
Dell/EMC Storage Enclosure User’s Guides.
•Updates are sometimes included with the system to describe changes to
the system, software, and/or documentation.
NOTE: Always read the updates first because they often supersede
information in other documents.
•Release notes or readme files may be included to provide last-minute
updates to the system or documentation, or advanced technical reference
material intended for experienced users or technicians.
14Introduction
Cabling Your Cluster Hardware
NOTE: To configure Dell blade server modules in a Dell™ PowerEdge™ cluster, see
the Using Dell Blade Servers in a Dell PowerEdge High Availability Cluster
document located on the Dell Support website at support.dell.com/manuals.
Cabling the Mouse, Keyboard, and Monitor
When installing a cluster configuration in a rack, you must include a switch
box to connect the mouse, keyboard, and monitor to the nodes. See the
documentation included with your rack for instructions on cabling each
node’s connections to the switch box.
Cabling the Power Supplies
Refer to the documentation for each component in your cluster solution to
ensure that the specific power requirements are satisfied.
The following guidelines are recommended to protect your cluster solution
from power-related failures:
•For nodes with multiple power supplies, plug each power supply into a
separate AC circuit.
•Use uninterruptible power supplies (UPS).
•For some environments, consider having backup generators and power
from separate electrical substations.
Figure 2-1, and Figure 2-2 illustrate recommended methods for power cabling
for a cluster solution consisting of two PowerEdge systems and two storage
systems. To ensure redundancy, the primary power supplies of all the
components are grouped onto one or two circuits and the redundant power
supplies are grouped onto a different circuit.
Cabling Your Cluster Hardware15
Figure 2-1. Power Cabling Example With One Power Supply in the PowerEdge Systems
redundant power
supplies on one AC
power strip (or on one
AC PDU [not shown])
primary power
supplies on one AC
power strip (or on one
AC PDU [not shown])
NOTE: This illustration is intended only to demonstrate the power
distribution of the components.
SPS
and One Standby Power Supply (SPS) in an AX4-5 Storage System
16Cabling Your Cluster Hardware
Figure 2-2. Power Cabling Example With Two Power Supplies in the PowerEdge
A
B
A
B
0Fibre 1Fibre0Fibre 1Fibre
A
B
A
B
primary power supplies
on one AC power strip
(or on one AC PDU [not
shown])
redundant power supplies
on one AC power strip (or
on one AC PDU [not
shown])
Systems and Two SPS(s) in an AX4-5 Storage System
Cabling Your Cluster for Public and Private
Networks
The network adapters in the cluster nodes provide at least two network
connections for each node, as described in Table 2-1.
NOTE: To configure Dell blade server modules in a Dell PowerEdge cluster, see the
Using Dell Blade Servers in a Dell PowerEdge High Availability Cluster document
located on the Dell Support website at support.dell.com/manuals.
Cabling Your Cluster Hardware17
Table 2-1. Network Connections
cluster node 1
cluster node 2
public network
private network
private
network
adapter
public
network
adapter
Network ConnectionDescription
Public networkAll connections to the client LAN.
At least one public network must be configured for
Mixed mode for private network failover.
Private networkA dedicated connection for sharing cluster health and
status information only.
Figure 2-3 shows an example of cabling in which dedicated network adapters
in each node are connected to each other (for the private network) and the
remaining network adapters are connected to the public network.
Figure 2-3. Example of Network Cabling Connection
18Cabling Your Cluster Hardware
Loading...
+ 42 hidden pages
You need points to download manuals.
1 point = 1 manual.
You can buy points or you can get point for every manual you upload.