Download Dell PowerEdge Cluster FL100 Troubleshooting guide

Transcript
SAN GUIDE
SAN www.dell.com
support.dell.com
SAN GUIDE
www.dell.com
support.dell.com
Notes, Notices, Cautions, and Warnings
Throughout this guide, blocks of text may be accompanied by an icon and printed in
bold type or in italic type. These blocks are notes, notices, cautions, and warnings,
and they are used as follows:
NOTE: A NOTE indicates important information that helps you make better use of your
computer system.
NOTICE: A NOTICE indicates either potential damage to hardware or loss
of data and tells you how to avoid the problem.
CAUTION: A CAUTION indicates a potentially hazardous situation which, if
not avoided, may result in minor or moderate injury.
WARNING: A WARNING indicates a potentially hazardous situation which,
if not avoided, could result in death or serious bodily injury.
____________________
Information in this document is subject to change without notice.
© 2000 Dell Computer Corporation. All rights reserved.
Reproduction in any manner whatsoever without the written permission of Dell Computer Corporation is strictly forbidden.
Trademarks used in this text: Dell, the DELL E COM logo, PowerEdge, PowerVault, and Dell
OpenManage are trademarks of Dell Computer Corporation; ClusterX is a registered trademark
of VERITAS Software; Intel is a registered trademark of Intel Corporation.
Other trademarks and trade names may be used in this document to refer to either the entities
claiming the marks and names or their products. Dell Computer Corporation disclaims any proprietary interest in trademarks and trade names other than its own.
June 2000
P/N 078UD Rev. A01
Contents
MSCS Cluster Support for Dell PowerVault Storage Area Networks . . . . . . . . . . . 1-1
Direct-Attached Cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-2
SAN-Attached Cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-2
Cluster Consolidation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-3
PowerVault SAN Components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-4
SAN-Attached Clusters. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-5
SAN-Attached Cluster Rules and Guidelines. . . . . . . . . . . . . . . . . . . . . . . 1-6
Cluster Installation Overview. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-8
Windows NT 4.0 Cluster Installation Overview . . . . . . . . . . . . . . . . . . . . . . . . 1-8
Windows 2000 Cluster Installation Overview . . . . . . . . . . . . . . . . . . . . . . . . . 1-9
Using Dell PowerVault Fibre Channel Switches. . . . . . . . . . . . . . . . . . . . . . . 1-11
SAN-Attached Cluster Configurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-11
Connecting a Single PowerEdge Cluster to a Single PowerVault 65xF
Storage System . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-11
Connecting a PowerVault 130T DLT Library and PowerVault 35F
Fibre Channel Bridge to a Cluster-Attached PowerVault SAN . . . . . . . . . . . . 1-12
Connecting a Single PowerEdge Cluster to Multiple
PowerVault 65xF DPEs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-13
Configurations With Two to Four PowerVault DPEs . . . . . . . . . . . . . . . . . . . 1-13
Configurations With SAN Backup. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-14
Configuring Hard-Disk Drive Letters When Using Multiple Shared
Storage Units . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-15
Connecting Multiple PowerEdge Clusters to a PowerVault SAN. . . . . . . . . . 1-16
Two PowerEdge Clusters Sharing a PowerVault SAN . . . . . . . . . . . . . . . . . . 1-17
Three PowerEdge Clusters Sharing a PowerVault SAN . . . . . . . . . . . . . . . . . 1-19
Configurations With Multiple SAN-Attached Clusters, Each Using
Multiple PowerVault 65xF DPEs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-20
Cluster Consolidation Configurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-21
Cluster Consolidation Rules and Guidelines. . . . . . . . . . . . . . . . . . . . . . . . . . 1-22
iii
Basic Installation Overview for Cluster Consolidation . . . . . . . . . . . . . . . . . .
Installing the QLogic Fibre Channel Configuration Utility . . . . . . . . . . . .
Cluster Consolidation Switch Zoning Requirements . . . . . . . . . . . . . . . . . . .
Example Configurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Additional PowerEdge Cluster Maintenance Procedures . . . . . . . . . . . . . . . . . . .
Using the QLogic Fibre Channel Configuration Utility for Storage
Processor Failure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Upgrading to a Cluster Consolidation Configuration . . . . . . . . . . . . . . . . . . .
1-23
1-26
1-27
1-29
1-33
1-33
1-33
Index
Figures
Figure 1-1.
Figure 1-2.
Figure 1-3.
Figure 1-4.
Figure 1-12.
Figure 1-13.
Figure 1-14.
Figure 1-15.
Figure 1-16.
Direct-Attached Cluster Configuration. . . . . . . . . . . . . . . . . . . . . . . . 1-2
SAN-Attached Cluster Configuration . . . . . . . . . . . . . . . . . . . . . . . . . 1-3
Cluster Consolidation Configuration . . . . . . . . . . . . . . . . . . . . . . . . . 1-4
How a SAN-Attached PowerEdge Cluster Uses
Networking Segments. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-5
Example of Cabling a Single PowerEdge Cluster With Two
PowerVault Fibre Channel Switches . . . . . . . . . . . . . . . . . . . . . . . . 1-12
Example of Cabling a PowerVault Storage System and a
PowerVault 130T DLT Library. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-13
Example of Cabling PowerEdge Cluster F-Series Nodes to
Four PowerVault 65xF DPEs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-14
Example of Cabling PowerEdge Cluster F-Series Nodes to Four
PowerVault 65xF DPEs and SAN Backup Using the PowerVault
Fibre Channel Bridge and Tape Library . . . . . . . . . . . . . . . . . . . . . . 1-15
Example of a Two-Cluster Configuration Sharing a
PowerVault 130T DLT Library . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-18
Example of a Three-Cluster Configuration Sharing a
PowerVault SAN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-19
Example of Cabling Two PowerEdge Cluster F-Series
Systems in a SANS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-21
Example Showing Two Zones . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-28
Example Showing Two Zones . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-29
Five MSCS Clusters Sharing a PowerVault 65xF Storage System . 1-30
Ten MSCS Clusters Sharing a PowerVault 65xF Storage System. . 1-31
Multiple Cluster, Multiple Storage SAN Environment Example . . . 1-32
Table 1-1.
Table 1-2.
Rules and Guidelines for SAN-Attached Clusters . . . . . . . . . . . . . . . 1-6
Cluster Consolidation Rules. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-22
Figure 1-5.
Figure 1-6.
Figure 1-7.
Figure 1-8.
Figure 1-9.
Figure 1-10.
Figure 1-11.
Tables
iv
Dell™ PowerEdge™ Cluster
F-Series SAN Guide
This document supplements the information in the Dell PowerEdge Cluster FE100/
FL100 Installation and Troubleshooting Guide and the Dell PowerVault™ storage area
network (SAN) documentation for using Dell PowerEdge Cluster F-Series configurations running Microsoft® Cluster Server (MSCS) and Microsoft Windows® 2000
Cluster Service software in conjunction with Dell PowerVault SAN 3.0. This document
provides supplemental information and procedures specific to SAN configurations.
See the Dell PowerEdge Cluster FE100/FL100 Installation and Troubleshooting Guide
for a detailed list of additional related documentation, in-depth information, and stepby-step procedures for configuring a cluster.
Use this document in conjunction with the Dell PowerEdge Cluster FE100/FL100
Installation and Troubleshooting Guide for installing Dell PowerEdge F-Series configurations into a SAN environment.
This document and other PowerEdge cluster documents use the term F-Series to
refer to PowerEdge FE100 and FL100 systems.
MSCS Cluster Support for Dell PowerVault
Storage Area Networks
A Dell PowerVault SAN is a configuration of computer and storage systems that incorporates dedicated Fibre Channel connectivity between the computer systems and
storage. A SAN bypasses traditional network bottlenecks and supports direct highspeed data transfer between the servers and storage devices. SAN storage may
consist of both hard-disk drive and tape storage devices that are connected through
switches and bridges to one or more servers or clusters. See the Dell PowerVault
SAN documentation for more information.
PowerEdge Cluster F-Series configurations can be either direct-attached or SANattached clusters.
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-1
The following options are available for PowerEdge Cluster F-Series configurations:
•
•
•
Direct-attached cluster
SAN-attached cluster
Cluster consolidation
The following sections provide detailed information and examples for these options.
Direct-Attached Cluster
In a direct-attached cluster configuration, both nodes of the cluster are directly
attached to a single PowerVault 65xF storage system.
NOTE: A direct-attached cluster configuration does not require a SAN.
Figure 1-1 shows a basic direct-attached cluster configuration.
Public LAN
PowerEdge
server
PowerEdge
server
cluster interconnect
Fibre Channel
connections
Fibre Channel
connections
PowerVault
storage system
Figure 1-1. Direct-Attached Cluster Configuration
See the Dell PowerEdge Cluster FE100/FL100 Installation and Troubleshooting Guide
for more information about installing, managing, and troubleshooting direct-attached
cluster configurations.
SAN-Attached Cluster
In a SAN-attached cluster configuration, both cluster nodes are attached to a single
PowerVault 65xF storage system or to multiple PowerVault 65xF storage systems
through a Dell PowerVault SAN using a redundant Fibre Channel switch fabric. One or
more PowerVault Fibre Channel switches make up a Fibre Channel fabric—an infrastructure that is the backbone for deploying and managing information technology (IT)
resources as a network.
1-2
Dell PowerEdge Cluster F-Series SAN Guide
Zoning helps to segment a SAN into logical fabric subsets by setting up barriers
between different operating environments, or between a cluster and other servers or
clusters. By using switch zoning, you can attach multiple clusters to a SAN.
See the Dell PowerVault 5xF Switches Zoning Guide for information on zoning concepts and implementations.
Figure 1-2 shows an advanced SAN-attached cluster configuration.
Public LAN
PowerEdge
server
PowerEdge
server
cluster interconnect
Fibre Channel
connections
Fibre Channel
connections
Fibre Channel
switch
Fibre Channel
switch
PowerVault
storage system
Figure 1-2. SAN-Attached Cluster Configuration
See “SAN-Attached Cluster Configurations,” found later in this document for more
information.
Cluster Consolidation
In a cluster consolidation configuration, both nodes of multiple clusters are attached
to a single PowerVault 65xF storage system through a PowerVault SAN using a redundant Fibre Channel switch fabric and switch zoning.
Figure 1-3 shows an example of a cluster consolidation configuration.
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-3
Public LAN
PowerEdge
Cluster pair
cluster
interconnect
PowerEdge
Cluster pair
cluster
interconnect
Fibre Channel
connections
Fibre Channel
switch
Fibre Channel
switch
PowerVault
storage system
Figure 1-3. Cluster Consolidation Configuration
See “Cluster Consolidation Configurations,” found later in this document, for more
information about supported configurations and installation procedures.
PowerVault SAN Components
Dell PowerVault SAN products supported in cluster configurations include the following components:
•
•
•
•
•
PowerVault 650F, 651F, and 630F — Fibre Channel storage systems
•
•
PowerVault 130T — Digital Linear Tape (DLT) library
PowerVault 50F, 51F, and 56F — Fibre Channel switches
QLogic QLA-2100/33 — Fibre Channel host bus adapters (HBAs) (copper only)
QLogic QLA-2200/33 or 2200/66— Fibre Channel HBAs (copper only)
PowerVault 35F — Fibre Channel-to-small computer system interface (SCSI)
bridge
PowerVault 120T — DLT autoloader
NOTE: Not all of the products in the preceding list are supported in all SAN-attached
cluster configurations. See the section in this guide for your specific configuration and
its corresponding supported components.
See the Dell PowerVault SAN documentation and the appropriate SAN component
documentation for configuration information.
1-4
Dell PowerEdge Cluster F-Series SAN Guide
SAN-Attached Clusters
SAN-attached clusters are cluster configurations in which you cable redundant Fibre
Channel HBAs to a redundant Fibre Channel switch fabric. You make the connection
to the storage system through the switch fabric. SAN-attached configurations can
share certain resources with other servers, storage systems, and backup systems on
the SAN.
Figure 1-4 shows how a SAN-attached PowerEdge Cluster F-Series configuration uses
three networking segments: public network, private network, and SAN. Each segment may vary in components and complexity.
Public LAN
public
network
PowerEdge
server
PowerEdge
server
private network
(cluster interconnect)
PowerVault Fibre Channel switch
PowerVault Fibre
Channel switch
Fibre Channel bridge
PowerVault 130T
PowerVault
storage system
storage area network
Figure 1-4. How a SAN-Attached PowerEdge Cluster Uses Networking Segments
support.dell.com
•
Public network — A public network connects cluster nodes to the client network.
This network consists of devices such as Ethernet NICs installed in the cluster
nodes.
•
Private network (node-to-node cluster interconnect) — A dedicated network that
connects the servers to each other. The primary function of the private network
in a cluster configuration is to pass heartbeat information that allows each cluster
node to monitor the health of the other cluster node.
Dell PowerEdge Cluster F-Series SAN Guide
1-5
•
Storage area network (SAN) — A configuration of computer and storage systems
that incorporates dedicated Fibre Channel connectivity between the computer
systems and storage. A SAN bypasses traditional network bottlenecks and supports direct high-speed data transfer between the servers and storage devices.
SAN-Attached Cluster Rules and Guidelines
The following rules and requirements described in Table 1-1 apply to SAN-attached
clusters. For rules and requirements for cluster consolidation configurations, see
“Cluster Consolidation Rules and Guidelines” found later in this document. See the
Dell PowerVault SAN documentation for more information on general SAN rules and
guidelines.
Table 1-1. Rules and Guidelines for SAN-Attached Clusters
1-6
Rule/Guideline
Description
Number of servers
supported
Up to 10 two-node clusters attached to a SAN.
Combinations of standalone servers and cluster pairs
not to exceed 20 server maximum.
Cluster pair support
All homogeneous and heterogeneous PowerEdge
Cluster F-Series configurations supported in directattach configurations are supported in SAN-attached
configurations. See the Dell PowerEdge Cluster FE100/
FL100 Platform Guide for more information on supported cluster pairs.
Primary storage
Up to 8 PowerVault Fibre Channel disk arrays are supported on a SAN. Primary storage can be owned by a
single cluster and multiple clusters (through SANattached cluster consolidation configurations discussed
later in this document). A mixture of these types of
ownership on a single storage system is not supported.
For example, the following are supported configurations: one cluster to one storage system, one cluster to
multiple storage systems, and multiple clusters to one
storage system. However, a configuration with multiple
clusters to multiple storage systems is not supported.
A mixture of configurations is supported on a SAN.
Secondary storage
Up to 4 Fibre Channel Bridges are supported, each with
up to 2 PowerVault 130T tape libraries or up to 4 PowerVault 120T tape autoloaders. Any server attached to the
SAN can share these devices.
Dell OpenManage™
Storage
Consolidation
Not required for standard SAN-attached clusters.
Fibre Channel switch
configuration
Redundant switch fabrics required for clusters.
Fibre Channel switch
zoning
Required whenever a cluster shares a SAN with other
cluster(s) or standalone server(s).
Dell PowerEdge Cluster F-Series SAN Guide
Table 1-1. Rules and Guidelines for SAN-Attached Clusters (continued)
Rule/Guideline
Description
Fibre Channel
switches supported
PowerVault 50F, 51F, and 56F.
Fibre Channel switch
firmware
For 50F, version 1.6d5-dell or later; for 51F and 56F, version 2.1.7 or later.
Fibre Channel HBA
supported
QLogic 2100/33, 2200/33, or 2200/66 megahertz (MHz)
HBA (copper cabling). HBAs within a single cluster must
be the same.
QLogic driver
version
Version 7.04.08 or later for Windows NT.
Version 7.04.08 or later for Windows 2000.
QLConfig version
Version 1.24 or later. See “Installing the QLogic Fibre
Channel Configuration Utility,” found later in this document, for information on installing QLConfig.
PowerVault
firmware version
Version 5.11.08 or later. All PowerVault storage systems
must be running the same version of firmware.
Operating system
Each cluster attached to the SAN can run either
Windows NT 4.0 Server, Enterprise Edition or
Windows 2000 Advanced Server. A SAN configuration
with one or more clusters running Windows 2000 and
one or more clusters running Windows NT is supported.
Windows NT or
Windows 2000 service pack
Windows NT configurations require Service Pack 6a or
later. Windows 2000 configurations require the latest
supported service pack. See the Dell PowerEdge Cluster FE100/FL100 Platform Guide for more information
on the latest supported service pack.
PowerVault 530F
SAN appliance
Not supported.
NOTES: See “Installing the QLogic Fibre Channel Configuration Utility,” found later in
this document, for information on installing QLConfig in SAN-attached cluster
configurations.
Dell OpenManage Storage Consolidation software must not be running on cluster
nodes.
With SAN 3.0, you can configure a SAN of up to 20 servers and 8 storage systems.
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-7
Cluster Installation Overview
This section provides an overview sequence for installing PowerEdge Cluster F-Series
systems with Windows NT 4.0 Server, Enterprise Edition or Windows 2000 Advanced
Server operating systems. More detailed instructions are provided later in this
document.
NOTICE: Before installing the cluster, ensure that your site can handle the
power requirements of the cluster equipment. Contact your Dell sales representative for information about your region’s power requirements.
WARNING: Hardware installation should be performed only by trained
service technicians. Before working inside the computer system, see the
safety instructions in your PowerEdge system documentation to avoid a
situation that could cause serious injury or death.
Windows NT 4.0 Cluster Installation Overview
The following is an overview sequence for installing Windows NT 4.0 on a PowerEdge
Cluster FE100/FL100.
1.
Add NICs, HBAs, redundant array of independent disks (RAID) controllers
(optional), SCSI hard-disk drives, Fibre Channel hard-disk drives, and other components to the existing system hardware to meet the requirements for a
PowerEdge Cluster FE100/FL100 configuration.
For more information on upgrading existing non-clustered systems to a cluster
configuration, see the Dell PowerEdge Cluster FE100/FL100 Installation and Troubleshooting Guide.
2.
Cable the system hardware for clustering.
3.
If you are using hardware-based RAID for the internal SCSI hard-disk drives, configure them using the controller’s basic input/output system (BIOS) utility.
4.
Perform the low-level configuration of the HBAs.
5.
Install and configure the Windows NT 4.0 Server, Enterprise Edition operating
system on each node. Configure the public and private NIC interconnects in each
node, and place the interconnects on separate Internet protocol (IP) subnetworks
using static IP addresses.
NOTES: Public refers to the NIC used for client connections. Private refers to the
dedicated cluster interconnect.
If you are using Giganet cluster local area network (cLAN) Host Adapters or a
Giganet cLAN Cluster Switch, see the Dell PowerEdge Cluster FE100/FL100
Installation and Troubleshooting Guide for more information.
6.
1-8
Install the device driver for the integrated video controller.
Dell PowerEdge Cluster F-Series SAN Guide
7.
Install Windows NT Service Pack 6a or later.
See the Dell PowerEdge Cluster FE100/FL100 Platform Guide for more information on the latest supported service pack.
8.
Install the miniport driver for the Fibre Channel HBAs in each node.
9.
Install the QLogic Fibre Channel configuration utility.
10. Install Dell OpenManage Application Transparent Failover (ATF) on each node and
reboot.
11. Install Dell OpenManage Managed Node (Data Agent) on each node.
12. Install Dell OpenManage Data Supervisor or Dell OpenManage Data Administrator on node A. From node A, configure the RAID level on the storage system.
13. Using Windows NT Disk Administrator, partition, format, and assign drive letters
to the Fibre Channel hard-disk drives in the shared storage system. Repeat drive
letter assignments on node B.
14. Install and configure the MSCS software on each node. Before rebooting, you
must reinstall the service pack.
NOTICE: To avoid damage to the system, do not reboot the nodes until
you reinstall the service pack.
NOTE: If you reinstall MSCS, you must reinstall the Windows NT service pack.
15. Verify the functionality of the cluster.
16. Configure the Dell OpenManage Managed Node (Agent) to run in a cluster
environment.
17. Install and set up your application programs.
18. Install the Dell OpenManage Cluster Assistant With ClusterX software from the
management console (optional).
19. Record the configuration of the cluster on the cluster data sheets provided in the
Dell PowerEdge Cluster FE100/FL100 Installation and Troubleshooting Guide.
Windows 2000 Cluster Installation Overview
The following is an overview sequence for installing Windows 2000 on a PowerEdge
Cluster FE100/FL100. Specific steps are provided throughout this document.
support.dell.com
1.
Add NICs, HBAs, RAID controllers (optional), SCSI hard-disk drives, Fibre Channel
hard-disk drives, and other components to the existing system hardware to meet
the requirements for a PowerEdge Cluster F-Series. For more information on
upgrading existing non-clustered systems to a cluster configuration, see the Dell
PowerEdge Cluster FE100/FL100 Installation and Troubleshooting Guide.
2.
Cable the system hardware for clustering.
Dell PowerEdge Cluster F-Series SAN Guide
1-9
3.
If you are using hardware-based RAID for the internal SCSI hard-disk drives, configure them using the controller's BIOS utility.
4.
Perform the low-level configuration of the HBAs.
5.
Install and configure the Microsoft Windows 2000 Advanced Server operating
system on each node and the latest Windows 2000 Service Pack.
See the Dell PowerEdge Cluster FE100/FL100 Platform Guide for more information on the latest supported service pack.
6.
During the installation, check to install the Cluster Service files when prompted.
You will configure the Cluster Service later.
7.
Configure the public and private NIC interconnects in each node, and place the
interconnects on separate IP subnetworks using static IP addresses.
NOTES: Public refers to the NIC used for client connections. Private refers to the
dedicated cluster interconnect.
If you are using Giganet cLAN Host Adapters or a Giganet cLAN Cluster Switch,
see the Dell PowerEdge Cluster FE100/FL100 Installation and Troubleshooting
Guide for more information.
8.
Update the miniport driver for the Fibre Channel HBAs in each node.
9.
Install the QLogic Fibre Channel configuration software.
10. Install Dell OpenManage Application Transparent Failover (ATF) on each node and
reboot.
11. Install Dell OpenManage Managed Node (Data Agent) on each node.
12. Install Dell OpenManage Data Supervisor or Dell OpenManage Data Administrator on node A. From node A, configure the RAID level on the storage system.
13. Using the Windows 2000 Disk Management applet, partition, format, and
assign drive letters to the Fibre Channel hard-disk drives in the shared storage
system. Repeat drive letter assignments on node B for each volume.
NOTICE: Do not upgrade cluster disks or volumes to dynamic disks.
Microsoft requires basic disk configuration for cluster disks.
14. Install and configure the Cluster Service software on each node. Before rebooting, you must reinstall the service pack.
NOTE: If you reinstall the Cluster Service software, you must reinstall the
Windows 2000 service pack.
15. Verify the functionality of the cluster.
16. Configure the Dell OpenManage Managed Node (Agent) to run in a cluster
environment.
17. Install and set up your application programs.
1-10
Dell PowerEdge Cluster F-Series SAN Guide
18. Install Dell OpenManage Cluster Assistant With ClusterX from the management
console (optional).
19. Record the configuration of the cluster on the cluster data sheets provided in the
Dell PowerEdge Cluster FE100/FL100 Installation and Troubleshooting Guide.
Using Dell PowerVault Fibre Channel Switches
You can connect cluster nodes to the PowerVault 65xF shared storage system by
using redundant PowerVault Fibre Channel switches. When cluster nodes are connected to the storage system through Fibre Channel switches, the cluster
configuration is technically attached to a SAN.
NOTE: For Windows NT cluster configurations, you must first install Microsoft
Windows NT Service Pack 6a or later before you connect Fibre Channel switches to
the cluster. See the cluster’s basic installation procedure, found in the Dell
PowerEdge Cluster FE100/FL100 Installation and Troubleshooting Guide, for installation information.
To connect your cluster nodes to the PowerVault 65xF shared storage system, your
system must have the following features:
•
Two PowerVault 50F, 51F, or 56F Fibre Channel switches (more switches may be
necessary depending on the configuration)
NOTICE: PowerVault 50F Fibre Channel switches are not supported in
SAN-attached cluster consolidation configurations; however, these
switches may be used for standard SAN-attached clusters.
•
Copper high-speed, serial data connector (HSSDC) cables to connect the cluster
nodes to the Fibre Channel switches
•
Copper cables with an HSSDC connector on one end and a DB-9 connector on
the other to connect the Fibre Channel switch to the PowerVault 65xF disk processor enclosure (DPE) storage systems
NOTE: For cluster configurations, you can use only copper cables (not optical
cables) to connect the cluster nodes to Fibre Channel switches.
SAN-Attached Cluster Configurations
The following sections provide configuration and cabling examples for SAN-attached
clusters.
Connecting a Single PowerEdge Cluster to a Single
PowerVault 65xF Storage System
Connect each cluster node to both Fibre Channel switches and connect each Fibre
Channel switch to both PowerVault 65xF storage processors on the cluster’s shared
storage, as shown in Figure 1-5.
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-11
PowerEdge
server
PowerEdge
server
cluster interconnect
PowerVault 56F
Fibre Channel
switch
PowerVault 56F
Fibre Channel
switch
PowerVault
storage system
Figure 1-5. Example of Cabling a Single PowerEdge Cluster With
Two PowerVault Fibre Channel Switches
NOTICE: Switch zoning is required if a cluster or standalone server is
already attached to the SAN. Information on using switch zoning with cluster configurations is provided throughout this document. See the Dell
PowerVault system documentation for more information about switch
cabling.
Connecting a PowerVault 130T DLT Library and
PowerVault 35F Fibre Channel Bridge to a ClusterAttached PowerVault SAN
You can use the PowerVault 35F Fibre Channel SCSI bridge to support the
PowerVault 130T DLT library or the PowerVault 120T tape autoloader on PowerEdge
Cluster F-Series configurations. Figure 1-6 shows a supported PowerEdge Cluster
F-Series configuration using redundant Fibre Channel switches, Fibre Channel bridge,
and PowerVault 130T DLT library. In this configuration, each of the cluster nodes is
attached to the backup device and the backup local disk resources, as well as to the
owned cluster disk resources. See the storage and tape backup documentation for
more information on configuring these components.
NOTE: Contact the tape backup software vendor for more information on using your
software in a cluster configuration.
1-12
Dell PowerEdge Cluster F-Series SAN Guide
PowerEdge
server
PowerEdge
server
PowerVault 130T
cluster interconnect
Fibre Channel
bridge
PowerVault 56F
Fibre Channel
switch
PowerVault 56F
Fibre Channel
switch
PowerVault
storage system
Figure 1-6. Example of Cabling a PowerVault Storage System and a
PowerVault 130T DLT Library
NOTE: The PowerVault 35F Fibre Channel SCSI bridge, the PowerVault 130T DLT
library, and their cables are made through a single switch fabric; their operation is not
redundant. Therefore, tape backups may be incomplete if one of these components
fails.
Connecting a Single PowerEdge Cluster to Multiple
PowerVault 65xF DPEs
With the use of redundant PowerVault Fibre Channel switch fabrics, the PowerEdge
Cluster F-Series systems support configurations with multiple storage units attached
to a single pair of clustered servers. In this scenario, the MSCS software can fail over
disk drives in each cluster-attached PowerVault 65xF shared storage array between
the two cluster nodes.
Configurations With Two to Four PowerVault DPEs
When using multiple PowerVault 65xF storage arrays with your cluster, the following
rules apply:
•
•
•
•
support.dell.com
Two to four PowerVault DPEs can be used with the cluster nodes.
All PowerVault 65xF systems must use the same firmware revision. Contact your
Dell representative for more information on storage system firmware.
Redundant Fibre Channel switch fabrics are required.
Dell OpenManage Storage Consolidation software must not be running on cluster nodes.
Dell PowerEdge Cluster F-Series SAN Guide
1-13
Figure 1-7 provides an example of cabling PowerEdge Cluster F-Series nodes to four
PowerVault 65xF DPEs.
Cluster
node 1
Cluster
node 2
cluster interconnect
PowerVault Fibre
Channel switch 1
PowerVault Fibre
Channel switch 2
PowerVault PowerVault PowerVault PowerVault
storage
storage
storage
storage
system 1
system 2
system 3 system 4
Figure 1-7. Example of Cabling PowerEdge Cluster F-Series Nodes
to Four PowerVault 65xF DPEs
Configurations With SAN Backup
When using multiple PowerVault 65xF DPEs and SAN backup with your cluster, the
following requirements apply:
•
All PowerVault 65xF systems must use the same firmware revision. Contact your
Dell representative for more information on storage system firmware.
•
Redundant Fibre Channel switch fabrics are required.
Figure 1-8 provides an example of cabling PowerEdge Cluster F-Series nodes to
PowerVault 65xF DPEs and SAN backup using the PowerVault Fibre Channel bridge
and tape library.
1-14
Dell PowerEdge Cluster F-Series SAN Guide
l
cluster 1
cluster 2
PowerVault Fibre
Channel switch
PowerVault Fibre
Channel switch
Fibre Channel
bridge
PowerVault
130T
PowerVault
storage
system
for cluster 1
PowerVault
storage
system
for cluster 2
PowerVault
storage
system
for cluster 2
Figure 1-8. Example of Cabling PowerEdge Cluster F-Series Nodes
to Four PowerVault 65xF DPEs and SAN Backup Using the PowerVault
Fibre Channel Bridge and Tape Library
Configuring Hard-Disk Drive Letters When Using Multiple
Shared Storage Units
Before installing MSCS, it is very important to make sure that both nodes have the
same view of the shared storage. Because each node has access to hard-disks that
are in a common storage array, each node must have the hard-disk drive letters
assigned identically. Perform the following steps to ensure that hard-disk drive letter
assignments are identical:
1.
Ensure the proper order of your cables.
The Windows NT Disk Administrator or Windows 2000 Disk Management
utility displays all disks that are accessible from the first QLogic HBA followed by
those that are accessible from the second QLogic HBA. Among the disks accessible from one HBA, the disks attached to a port with a lower port number on the
Fibre Channel switch will be displayed first, followed by those attached to a port
with a higher port number.
To ensure proper drive letter assignments, ensure that the first QLogic HBA
detected by each node is connected to the same switch (the switch connected to
PowerVault storage processor A) and that the second QLogic HBA detected is
connected to the other switch (the switch connected to PowerVault storage
processor B).
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-15
2.
Assign drive letters and volume labels to the disks.
a.
Turn off node A.
b.
Assign the drive letters on node B by using the Windows NT Disk Administrator or Windows 2000 Disk Management utility to create the drive
letters and volume labels. For example, create volumes labeled "Volume E"
for disk E and "Volume F" for disk F.
c.
Turn off node B. Turn on the node A, and start the Windows NT Disk
Administrator or Windows 2000 Disk Management utility.
d.
Assign the drive letters on node A.
If the cables are connected properly, the drive order will be the same as on
node B, and the drive letter assignments should follow the same order as on
the node B. The volume labels can also be used to double-check the drive
order by observing that the disk with volume label "Volume E" is assigned to
drive letter E and so on for each disk on each node. The drive letter assignment should be done on each of the shared disks, even if the disk displays
the drive letter correctly.
Connecting Multiple PowerEdge Clusters to a PowerVault
SAN
You can connect multiple PowerEdge clusters to a PowerVault SAN, each with its own
PowerVault 65xF storage system. Examples of supported configurations are:
•
•
•
one direct-attached cluster to one storage system
the SAN-attached configurations of one cluster to multiple PowerVault 65xF storage systems
one PowerVault 65xF storage system to multiple clusters
However, a configuration with multiple clusters to multiple storage systems is
not supported.
More than one PowerEdge Cluster F-Series configuration can share a PowerVault SAN
through the use of Fibre Channel switch zoning. By using PowerVault Fibre Channel
switches to implement zoning, you can arrange Fibre Channel switch fabric-connected
devices into logical groups over the physical switch fabric configuration (a Fibre Channel switch fabric consists of one or more Fibre Channel switches). Zoning
automatically and transparently enforces access of information to the zone devices.
You can segment the SANs by setting up zones between different clusters.
NOTICE: If any cluster shares a PowerVault Fibre Channel switch fabric with
any other clustered or nonclustered servers and storage systems, you must
configure zoning on the Fibre Channel switches.
See the Dell PowerVault 5xF Switches Zoning Guide for information on zoning concepts and implementations.
1-16
Dell PowerEdge Cluster F-Series SAN Guide
NOTE: Using zoning in SAN configurations containing multiple MSCS clusters prevents the nodes from discovering a storage device that belongs to another cluster. For
example, cluster nodes 1 and 2 that connect to shared storage system A should not
“see” shared storage B to which nodes 3 and 4 are connected. Zoning prevents this
scenario, while allowing all of the nodes to see the tape backup devices.
The PowerEdge Cluster F-Series configurations support the following zoning methods:
•
Port worldwide name
To use this zoning method, add one cluster at a time to the switch fabric. Then
zone the newly added cluster immediately after you add it to prevent it from
being accessed by other clusters or servers in the switch fabric. Redefine the
zoning for the cluster after you replace a component; for example, after you
replace a Fibre Channel HBA, a storage processor, or a Fibre Channel SCSI
bridge. The advantage of this method of zoning is that you can move cables from
port to port within the switch fabric without requiring the zones to be updated.
NOTICE: You must reconfigure the zones if you have replaced any Fibre
Channel storage components, such as Fibre Channel HBAs or
PowerVault 65xF storage processors.
•
Physical switch fabric port number
With this zoning method, you cannot move the cables that connect the servers,
the PowerVault 650F storage system, or the PowerVault 35F Fibre Channel SCSI
bridge to the Fibre Channel switches to a port outside the currently defined zone.
If you move the cables to a port outside the zone, you must redefine the zoning
for the cluster. The advantage of this method of zoning is that you can replace
components without requiring the zones to be updated.
NOTICE: You must reconfigure the zones if you have moved a cable
from a port inside the currently defined zone to a port outside the currently defined zone.
Two PowerEdge Clusters Sharing a PowerVault SAN
Figure 1-9 provides an example of a two-cluster configuration that shares a
PowerVault 130T DLT library.
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-17
Zone 1
server 3
server 1 server 2
server 4
PowerVault
Fibre Channel
bridge
PowerVault
56F Fibre
Channel
switch 2
PowerVault
56F Fibre
Channel
switch 1
PowerVault
tape library
PowerVault
65xF 1
PowerVault
65xF 2
Zone 2
Figure 1-9. Example of a Two-Cluster Configuration Sharing a
PowerVault 130T DLT Library
The example shown in Figure 1-9 contains the following configuration information.
Switch fabric 1 comprises switch 1:
•
Zone 1 — server 1, server 2, PowerVault 65xF storage system (number 1), and
Fibre Channel-to-SCSI bridge
•
Zone 2 — server 3, server 4, PowerVault 65xF storage system (number 2), and
Fibre Channel-to-SCSI bridge
Switch fabric 2 comprises switch 2:
•
•
Zone 1 — server 1, server 2, and PowerVault 65xF storage system (number 1)
Zone 2 — server 3, server 4, and PowerVault 65xF storage system (number 2)
To implement the zoning in the preceding configuration, perform the following steps:
1.
Ensure that the PowerVault 65xF storage system (number 2), server 3, and
server 4 are turned off.
2.
Configure zone 1 on switch fabric 1 and zone 1 on switch fabric 2.
3.
Turn on the PowerVault 65xF storage system (number 2), server 3, and server 4.
4.
Configure zone 2 on switch fabric 1 and zone 2 on switch fabric 2.
See the Dell PowerVault 5xF Switches Zoning Guide for more information on configuring zones on the Fibre Channel switches.
1-18
Dell PowerEdge Cluster F-Series SAN Guide
Three PowerEdge Clusters Sharing a PowerVault SAN
Figure 1-10 provides an example of a three-cluster configuration that shares a
PowerVault 130T DLT library.
Zone 2
Zone 1
server 1
server 2
server 3
server 4
PowerVault
51F Fibre
Channel
switch 1
PowerVault
Fibre Channel
bridge
PowerVault
tape library
PowerVault
51F Fibre
Channel
switch 3
PowerVault
65xF 1
PowerVault
65xF 2
Fibre Channel media (copper)
PowerVault
65xF 3
PowerVault
51F Fibre
Channel
switch 4
PowerVault
51F Fibre
Channel
switch 2
server 5
server 6
Zone 3
Figure 1-10. Example of a Three-Cluster Configuration Sharing a
PowerVault SAN
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-19
The example shown in Figure 1-10 contains the following configuration information.
Switch fabric 1 comprises switches 1 and 2:
•
Zone 1 — server 1, server 2, PowerVault 65xF storage system (number 1), and
Fibre Channel-to-SCSI bridge
•
Zone 2 — server 3, server 4, PowerVault 65xF storage system (number 2), and
Fibre Channel-to-SCSI bridge
•
Zone 3 — server 5, server 6, PowerVault 65xF storage system (number 3), and
Fibre Channel-to-SCSI bridge
Switch fabric 2 comprises switch 3:
•
•
Zone 1 — server 1, server 2, and PowerVault 65xF (number 1)
Zone 2 — server 3, server 4, and PowerVault 65xF (number 2)
Switch fabric 3 comprises switch 4. No zoning is required.
See the Dell PowerVault 5xF Switches Zoning Guide for more information on configuring zones on the Fibre Channel switches.
NOTE: The three-cluster configuration requires cascading the PowerVault 51F Fibre
Channel switches. You also can implement this configuration by using two
PowerVault 56F Fibre Channel switches. See the Dell PowerVault 5xF Fibre Channel
switch documentation for more information on cascading the switches.
To implement the zoning in the preceding configuration, perform the following steps:
1.
Make sure that the PowerVault 65xF storage system (number 2 and 3), server 3,
server 4, server 5, and server 6 are turned off.
2.
Configure zone 1 on switch fabric 1 and zone 1 on switch fabric 2.
3.
Turn on the PowerVault 65xF storage system (number 2), server 3, and server 4.
4.
Configure zone 2 on switch fabric 1 and zone 2 on switch fabric 2.
5.
Turn on the PowerVault 65xF storage system (number 3), server 5, and server 6.
6.
Configure zone 3 on switch fabric 1.
Configurations With Multiple SAN-Attached Clusters,
Each Using Multiple PowerVault 65xF DPEs
Using zoning (as described earlier in this document), a PowerEdge Cluster F-Series
system with multiple storage units can share the same SAN with another cluster. The
applicable zoning rules and guidelines apply, as described in Table 1-1, found earlier in
this document.
Figure 1-11 shows the cabling of a SAN configuration of two PowerEdge Cluster
F-Series clusters. One cluster uses multiple PowerVault 65xF DPEs and one cluster
uses a single PowerVault 65xF DPE. SAN backup is shown for both clusters.
1-20
Dell PowerEdge Cluster F-Series SAN Guide
Zone 1
server 3 server 4
server 1 server 2
PowerVault
56F Fibre
Channel
switch 1
PowerVault
56F Fibre
Channel
switch 2
PowerVault
Fibre Channel bridge
PowerVault
tape library
PowerVault 65xF 1
PowerVault
65xF 2
Zone 2
Figure 1-11. Example of Cabling Two PowerEdge Cluster F-Series
Systems in a SAN
Cluster Consolidation Configurations
With Dell OpenManage Storage Consolidation 3.0 software, you can partition a
PowerVault 65xF storage system into multiple disjoint storage systems, each with
one or multiple logical unit numbers (LUNs), and assign each storage system to a twonode MSCS cluster. Configurations with multiple two-node MSCS clusters are
referred to as cluster consolidation configurations. These configurations require a Dell
PowerVault SAN, Dell OpenManage Storage Consolidation 3.0 or later software, the
QLogic Fibre Channel configuration utility, and Fibre Channel switch zoning. Example
configurations and procedures are covered in the following sections. See “Installing
QLogic Fibre Channel Configuration Utility,” found later in this document, for information on installing QLConfig.
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-21
Cluster Consolidation Rules and Guidelines
Table 1-2 describes the important requirements for cluster consolidation
configurations.
Table 1-2. Cluster Consolidation Rules
1-22
Rule/Guideline
Description
Number of supported
clusters
10 cluster pairs (20 cluster nodes).
Cluster pair support
Any supported homogeneous server pair with the
QLogic 2200/33- or 2200/66-MHz HBA. See the Dell
PowerEdge Cluster FE100/FL100 Platform Guide for
more information.
Primary storage
1 PowerVault Fibre Channel disk array.
Secondary storage
Up to 4 Fibre Channel Bridges are supported, each with
up to 2 PowerVault 130T tape libraries or up to 4
PowerVault 120T tape autoloaders. Any server attached
to the SAN can share these devices.
Dell OpenManage
Storage
Consolidation
Version 3.0 is required for cluster consolidation
configurations.
Fibre Channel switch
configuration
Redundant switch fabrics are required.
Fibre Channel switch
zoning
Required for each cluster pair and the storage array.
Fibre Channel switches
supported
PowerVault 51F and 56F.
Fibre Channel switch
firmware
Version 2.1.7 or later.
Fibre Channel HBA
supported
QLogic 2200/33 MHz HBA (copper cabling).
QLogic 2200/66 MHz HBA (copper cabling).
QLogic driver version
Version 7.04.08 or later for Windows NT.
Version 7.04.08 or later for Windows 2000.
QLConfig version
Version 1.24 or later. See “Installing the QLogic Fibre
Channel Configuration Utility,” found later in this document, for information on installing QLConfig.
PowerVault firmware
version
Version 5.11.08 or later.
Dell PowerEdge Cluster F-Series SAN Guide
Table 1-2. Cluster Consolidation Rules (continued)
Rule/Guideline
Description
Operating system
All clusters attached to a single PowerVault 65xF storage system must be running either Windows NT Server
4.0, Enterprise Edition or Windows 2000 Advanced
Server. A mixture of operating systems within a cluster
is not supported.
Windows NT or
Windows 2000
service pack
Windows NT configurations require Service Pack 6a or
later. Windows 2000 configurations require the latest
supported service pack. See the Dell PowerEdge Cluster FE100/FL100 Platform Guide for more information
on the latest supported service pack.
PowerVault 530F
SAN Appliance
Not supported.
Disks
Each cluster has its own set of disks within the
PowerVault Fibre Channel disk array and cannot have
access to any other cluster’s disks.
SAN support
A cluster consolidation configuration consists of no
more than 10 clusters or 20 individual servers in several
combinations. For example, you can have a configuration consisting of 5 clusters (10 servers) and 10
stand-alone servers for a total of 20 servers.
Standalone servers
A mixture of standalone servers and cluster server
pairs on the cluster consolidation storage system is not
supported.
Additional software
application programs
Dell OpenManage ATF version 2.2.0.0 or later for Windows NT; version 2.3.2.5 or later for Windows 2000.
Dell OpenManage Managed Node (Data Agent)
version 4.1.4.0 or later.
Dell OpenManage Data Supervisor version 3.02.19 or
later, or Dell OpenManage Data Administrator
version 3.02.19 or later.
Basic Installation Overview for Cluster Consolidation
NOTICE: Before installing the cluster, ensure that your site can handle the
power requirements of the cluster equipment. Contact your Dell sales representative for information about your region’s power requirements.
WARNING: Hardware installation should be performed only by trained
service technicians. Before working inside the computer system, see the
safety instructions in your PowerEdge system documentation to avoid a
situation that could cause serious injury or death.
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-23
The following is an overview of the necessary steps for configuring multiple
PowerEdge F-Series clusters configured for cluster consolidation. See the PowerEdge
Cluster F-Series documentation and the PowerVault storage system documentation
for detailed installation information.
To configure multiple PowerEdge F-Series clusters configured for cluster consolidation, perform the following steps:
1.
Verify that all of the components adhere to the requirements and restrictions
described in the Dell PowerEdge Cluster FE100/FL100 Installation and Troubleshooting Guide, as well as to the additional requirements described throughout
this document.
2.
Make low-level configurations and hardware settings as described in the
PowerEdge Cluster F-Series documentation.
3.
Configure the storage system as described in the PowerVault storage system
documentation.
4.
Calculate the Execution Throttle value and change the Execution Throttle value in
the nonvolatile random-access memory (NVRAM) for each HBA. Ensure that the
sum of all of these values does not exceed 250 and that each value does not
exceed 90. See the PowerVault system documentation for instructions for calculating and changing the Execution Throttle value.
5.
Disconnect all Fibre Channel interconnects attached to all PowerVault 65xF storage systems in the SAN.
6.
Perform the following steps for each server:
a.
Install Windows NT Server 4.0, Enterprise Edition with Service Pack 6a or
later, or Windows 2000 Advanced Server with the latest supported service
pack. See the Dell PowerEdge Cluster FE100/FL100 Platform Guide for information on the supported service packs.
b.
Install the QLogic driver.
c.
Install the QLConfig configuration utility.
See “Installing the QLConfig Fibre Channel Configuration Utility,” found later
in this document, for information on installing QLConfig.
7.
1-24
d.
Install Dell OpenManage ATF and reboot the server.
e.
Install Dell OpenManage Storage Consolidation software (StorageC) for a
cluster environment. During the installation, check Yes in response to the
question Will this server be used in a Microsoft Cluster Server environment?
f.
Shut down Windows and turn off the server.
After you have configured all of the servers in the SAN, reconnect all Fibre Channel interconnects to the PowerVault 65xF storage systems.
Dell PowerEdge Cluster F-Series SAN Guide
8.
Configure zoning on the Fibre Channel switches by turning on the servers one at
a time. Each zone consists of a cluster pair (2 servers), the PowerVault 65xF, and
the Fibre Channel-to-SCSI bridge (if that zone is participating in the SAN backup).
See the Dell PowerVault SAN documentation for information about configuring
zoning.
9.
Start the Storage Consolidation Administrator on the Storage Consolidation
Master node. For each cluster, use the Server Group Assign option to assign
LUNs to both cluster nodes. It is important to use the Server Group Assign
option to assign one disk to multiple servers.
See the Dell OpenManage Storage Consolidation documentation for more
information.
10. Shutdown Windows on all of the servers.
11. Install MSCS software on each cluster. Because each cluster has its own set of
disks now, follow the standard PowerEdge Cluster F-Series installation (format
the disks, assign the drive letters, and install MSCS software, and so on).
12. Install Dell OpenManage Data Agent on both nodes of one cluster and create the
Generic Service cluster resource for the Agent.
Installing the QLogic Fibre Channel Configuration Utility
To install the QLogic Fibre Channel configuration utility, perform the following steps:
1.
Turn on or reboot the server.
2.
Insert the Dell PowerVault Fibre Channel Utilities CD into the CD-ROM drive. You
can load the application program directly from the Dell PowerVault Fibre Channel
Utilities CD.
3.
Start the QLogic Fibre Channel configuration utility.
4.
Read the component information, and click Next.
5.
Select the desired destination folder, and click Next.
6.
Select QLConfig Application and QLConfig Agent, and click Next.
The QLogic Fibre Channel configuration utility installation starts and displays an
information screen.
7.
Click Next.
8.
Read the displayed information, and click Next.
9.
Select the destination folder, and click Next.
10. Select the program folder, and click Next.
The SAN Configuration Agent installation starts and displays an information
screen.
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-25
11. Click Next.
12. Read the displayed information, and click Next.
13. Select the destination folder, and click Next.
14. Click Finish.
15. After the Intel® DMI 2.0 Service Provider SDK Version 1.10 installation starts, click
Next.
16. Read the license agreement, and click Yes.
17. Make sure that the win32sl service process is stopped.
18. Make sure that both of the following components are selected:
—
DMI 2.0 Service Provider
—
DMI 2.0 Client
19. Select the appropriate program folder, and click Next.
20. Click OK.
21. Reboot the system.
The installation is complete and the application program is ready to use. See “Using
the QLogic Fibre Channel Configuration Utility for Storage Processor Failure,” found
later in this document, for information on using QLConfig.
Cluster Consolidation Switch Zoning Requirements
Fibre Channel switch zoning is required for cluster consolidation configurations. The
following examples describe how to create a zone configuration for any supported
configuration.
Example 1: Create a zone around each pair of cluster nodes, the Fibre Channel-toSCSI bridge, and the PowerVault storage system. For example, a configuration with
two clusters attached to a single PowerVault storage system would have two zones as
shown in the Figure 1-12.
1-26
Dell PowerEdge Cluster F-Series SAN Guide
Zone 1
Zone 2
PowerEdge cluster
PowerEdge cluster
cluster
interconnect
PowerVault
51F Fibre
Channel
switch 1
PowerVault
51F Fibre
Channel
switch 2
PowerVault
Fibre Channel
storage
up to 11 disk
array enclosures
Figure 1-12. Example Showing Two Zones
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-27
Example 2: A configuration with two clusters attached to a single PowerVault storage
system would have two zones as shown in Figure 1-13.
Zone 1
cluster
interconnect
PowerEdge cluster
PowerEdge cluster
PowerVault
51F Fibre
Channel
switch 1
PowerVault
51F Fibre
Channel
switch 2
PowerVault3
5F Fibre
Channel
bridge
PowerVault
tape library
PowerVault
Fibre Channel
storage
up to 11
disk array
enclosures
Zone 2
Figure 1-13. Example Showing Two Zones
Example Configurations
This section provides examples of supported configurations for SAN-attached cluster
consolidation configurations.
Figure 1-14 shows an example of a configuration with five MSCS clusters sharing one
PowerVault 65xF storage system and a pair of PowerVault 56F 16-port Fibre Channel
switches. Figure 1-15 provides an example of a maximum configuration with one
PowerVault 65xF storage system. Figure 1-16 shows a combination of different groups
of MSCS clusters using different PowerVault 65xF storage systems.
1-28
Dell PowerEdge Cluster F-Series SAN Guide
Figure 1-14 shows a configuration with five MSCS clusters sharing a single
PowerVault 65xF storage system. These five MSCS clusters also share the two
PowerVault 130T tape libraries through the SAN with the use of the PowerVault 35F
Fibre Channel bridge. This configuration requires two PowerVault 56F, 16-port Fibre
Channel switches and the creation of five zones. Each zone contains an individual
two-node MSCS cluster (both servers), the Fibre Channel-to-SCSI bridge, and the
PowerVault 65xF storage system.
cluster
interconnect
PowerVault
56F Fibre Channel
switches (2)
PowerVault
35F Fibre
Channel bridge
PowerVault
Fibre Channel
storage
up to 11
disk array
enclosures
5 PowerEdge
cluster pairs
PowerVault tape libraries
Figure 1-14. Five MSCS Clusters Sharing a PowerVault 65xF Storage
System
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-29
Figure 1-15 shows a configuration with ten MSCS clusters sharing a single
PowerVault 65xF storage system. These ten MSCS clusters also share the two
PowerVault 130T tape libraries through the SAN with the use of the PowerVault 35F
Fibre Channel bridge. This configuration requires four Fibre Channel switches and the
creation of ten zones. Each zone contains an individual two-node MSCS cluster (both
servers), the Fibre Channel-to-SCSI bridge, and the PowerVault 65xF storage system.
cluster
interconnect
PowerVault
56F Fibre
Channel
switches (4)
PowerVault
35F Fibre
Channel
bridge
5 PowerEdge clusters
PowerVault
Fibre Channel
storage
PowerVault
tape libraries
5 PowerEdge clusters
Figure 1-15. Ten MSCS Clusters Sharing a PowerVault 65xF Storage
System
1-30
Dell PowerEdge Cluster F-Series SAN Guide
Figure 1-16 shows a configuration with six MSCS clusters. Clusters one through four
share a single PowerVault 65xF storage system, cluster five has its own storage
system, and cluster six uses four storage systems. All six MSCS clusters also share
the two PowerVault 130T tape libraries through the SAN with the use of the
PowerVault 35F Fibre Channel Bridge. This configuration requires four PowerVault 56F,
16-port Fibre Channel switches and the creation of six zones. Each of the first five
zones includes two nodes in the same MSCS cluster, the Fibre Channel-to-SCSI
bridge, and the PowerVault 65xF storage system. These five zones are for the five
clusters and the two storage systems on the left side of the figure. The sixth zone
includes the two nodes of the cluster on the right side of the figure, the Fibre
Channel-to-SCSI bridge, and its four storage systems.
cluster
interconnect
storage for
4 clusters
PowerVault
56F Fibre
Channel
switches (2)
1 PowerEdge
cluster
PowerVault
51F Fibre
Channel
switches (2)
4 PowerEdge
Clusters
1 PowerEdge
cluster
storage for
1 cluster
backup for
all clusters
storage for 1 cluster
PowerVault
35F Fibre
Channel bridge
Figure 1-16. Multiple Cluster, Multiple Storage SAN Environment
Example
support.dell.com
Dell PowerEdge Cluster F-Series SAN Guide
1-31
Additional PowerEdge Cluster
Maintenance Procedures
The following sections provide procedures for maintaining your PowerEdge Cluster
F-Series configurations.
Using the QLogic Fibre Channel Configuration Utility for
Storage Processor Failure
The QLogic Fibre Channel configuration utility is used for a storage processor failure.
When the failed storage processor is replaced with the new storage processor, the
QLogic Fibre Channel configuration utility updates the list of worldwide name associations stored in the firmware of the HBA and eliminates the need to reboot the server
in the event of an storage processor failure. This update must be performed on both
nodes of a cluster.
You must run the QLogic Fibre Channel configuration utility each time the hardware
on the SAN changes to create the appropriate settings for the replaced device.
If the failed storage processor is part of a zone, you need to update the alias, zone,
and configuration information on the PowerVault 5xF switch after the storage processor replacement. See the Dell PowerVault Systems Zoning Guide for details.
Upgrading to a Cluster Consolidation Configuration
The following is an overview of the necessary steps for adding clusters to an existing
PowerEdge Cluster F-Series configuration using the existing PowerVault 65xF storage
system.
1.
Verify that all the components adhere to the requirements and restrictions for
multiple cluster configurations, as described throughout this document.
2.
Perform the following steps on each node of the existing cluster:
a.
Verify that the QLogic driver, the ATF drivers, and the storage firmware are
running SAN 3.0 level or later. If not, update the drivers and firmware with
the proper version.
b.
Install the QLogic Fibre Channel configuration utility, if it has not been
installed.
See “Installing the QLogic Fibre Channel Configuration Utility,” found earlier
in this document, for information on installing QLConfig.
c.
For Windows NT Server 4.0, Enterprise Edition, select the Devices applet in
the Control Panel. Change the start-up setting for Cluster Disk from
System to Disabled.
For Windows 2000, right-click the My Computer desktop icon, point to
Properties, select the Hardware tab, and click Device Manager. Click View,
and select Show Hidden Devices. Expand Non-Plug and Play drivers.
1-32
Dell PowerEdge Cluster F-Series SAN Guide
Click Cluster Disk, and then select the Driver tab. Change the startup setting for Cluster Disk from System to Disabled.
d.
For Windows NT Server 4.0, Enterprise Edition, select the Services applet in
the Control Panel. Change the start-up setting for Cluster Service from
Automatic to Disabled.
For Windows 2000, select Start, point to Programs, point to Administrative Tools, and then point to Services. Change the start-up setting for
Cluster Service from Automatic to Disabled.
e.
Install Dell OpenManage Storage Consolidation software (StorageC) for a
cluster environment. During the installation, check Yes in response to the
question Will this server be used in a Microsoft Cluster Server environment?
f.
Shutdown Windows and turn off the node.
3.
Add the switches, if needed, and the additional servers to the SAN switch fabric.
4.
Configure the hardware according to the Dell PowerEdge Cluster FE100/FL100
Installation and Troubleshooting Guide and the storage system according to the
PowerVault storage system documentation.
5.
Calculate the Execution Throttle value and change the Execution Throttle value in
the NVRAM for each HBA. Ensure that the sum of all of these values does not
exceed 250 and that each value does not exceed 90.
See the PowerVault system documentation for instructions for calculating and
changing the Execution Throttle value.
6.
Disconnect only the cables attached to the PowerVault 65xF storage system.
7.
Perform the following steps on each additional server:
a.
Install Windows NT Server 4.0, Enterprise Edition with Service Pack 6a or
later, or Windows 2000 Advanced Server with the latest supported service
pack. See the Dell PowerEdge Cluster FE100/FL100 Platform Guide for information on the supported service packs.
b.
Install the QLogic driver.
c.
Install the QLogic Fibre Channel configuration utility.
See “Installing the QLogic Fibre Channel Configuration Utility,” found earlier
in this document, for information on installing QLConfig.
support.dell.com
d.
Install the Dell OpenManage ATF and reboot the server.
e.
Install Dell OpenManage Storage Consolidation software (StorageC) for a
cluster environment. During the installation, check Yes in response to the
question Will this server be used in a Microsoft Cluster Server environment?
f.
Shut down Windows and turn off the server.
Dell PowerEdge Cluster F-Series SAN Guide
1-33
8.
After you have configured all of the servers, reconnect all of the cables to the
PowerVault 65xF storage systems.
9.
Turn on all of the servers, one at a time. Configure zoning on the Fibre Channel
switches. Each zone consists of a cluster pair (2 servers), the PowerVault 65xF,
and the Fibre Channel-to-SCSI bridge (if that zone is participating in the SAN
backup).
See the Dell PowerVault SAN documentation for information about configuring
zoning.
10. Start the Storage Consolidation Administrator on the Storage Consolidation
Master node. Use the Server Group Assign option to assign the current disks
used by the existing PowerEdge Cluster F-Series system back to both of its
nodes. Assign other disks to both nodes of each additional cluster.
11. Repeat the following steps on each node of the existing PowerEdge Cluster
F-Series system:
a.
For Windows NT Server 4.0, Enterprise Edition, select the Devices applet in
the Control Panel. Change the start-up setting for Cluster Disk from
Disabled back to System.
For Windows 2000, right-click the My Computer desktop icon, point to
Properties, select the Hardware tab, and click Device Manager. Click View,
and then select Show Hidden Devices. Expand Non-Plug and Play drivers.
Click Cluster Disk, and then select the Driver tab. Change the startup setting for Cluster Disk from Disabled back to System.
b.
For Windows NT Server 4.0, Enterprise Edition, select Services applet in the
Control Panel. Change the start-up setting for Cluster Service from
Disabled back to Automatic.
For Windows 2000, select Start, point to Programs, point to Administrative Tools, and then point to Services. Change the startup setting for
Cluster Service from Disabled back to Automatic.
c.
Reboot the node.
12. Shutdown Windows on all additional servers.
13. Install MSCS software on each additional cluster.
Since each cluster now has its own set of disks, follow the installation procedure
for a typical PowerEdge Cluster F-Series system installation (format the disks,
assign the drive letters, and install MSCS software, and so on).
1-34
Dell PowerEdge Cluster F-Series SAN Guide
Index
C
E
cluster cabling
connecting multiple PowerEdge
clusters to a PowerVault SAN ,
1-16
examples of supported configurations for
cluster consolidation configurations ,
1-29
cluster configuration
three PowerEdge clusters sharing a
PowerVault SAN , 1-19
two PowerEdge clusters sharing a
PowerVault SAN , 1-17
cluster consolidation configurations ,
1-21
cluster consolidation rules and guidelines
, 1-22
configurations with multiple SANattached clusters , 1-20
configurations with two to four
PowerVault DPEs , 1-13
configurations with twoor threer
PowerVault DPEs , 1-14
connecting a single PowerEdge cluster to
multiple PowerVault 65xF DPEs ,
1-13
connecting PowerVault 130T DLT library
and PowerVault 35F Fibre Channel
Bridge to a cluster-attached
PowerVault SAN , 1-12
H
hard-disk drive
configuring drive letters with multiple
shared storage systems , 1-15
P
PowerVault 130T DLT library and
PowerVault 35F Fibre Channel SCSI
bridge , 1-12
PowerVault 65xF
with multiple SAN-attached clusters ,
1-20
S
SAN-attached cluster configurations ,
1-11
SAN-attached cluster rules and
guidelines , 1-6
support for Dell PowerVault SAN , 1-1
switch zoning requirements , 1-23
support.dell.com
Index
37
U
upgrading to a cluster consolidation
configuration , 1-33
using Dell PowerVault Fibre Channel
Switches , 1-11
W
Windows 2000
Disk Management applet , 1-10
installation , 1-9
Windows 2000 Disk Management , 1-15
Windows NT
installation , 1-8
Windows NT Disk Administrator , 1-15
Windows NT Service Pack , 1-9
Z
zoning , 1-16
methods , 1-17
38
Dell PowerEdge Cluster F-Series SAN Guide
www.dell.com
support.dell.com
!
"#$%&'%(
!"#$%&'()*$&+$,-./01234
'5$672890':$;<=3
>?672894@ABCD$E2FGHI./01234
HI672894JK'(LM2FGHI./01234
!"#$
%&'()*+, -.
/ *+0#$ !""
# 1
2*+0#$ 12*+0#$
%034456*+789:;&'()<=;>?"#$)*+78
9:@A BC#0>?"DE$
F $ G%&'()*+!,
NOPQRS#TUOVWX$ WYN*Z[O )**+,-./ *+,-./
,-./01
23456 *+,-./
*+,-./789:;-: WYN*\]NO^$_`
! ,-./:.64<= !!! ,-./:.64<= >?:@4AB5.:CAD
ab$WYN*cd
EF "# ,-./ $%
.6G4H.IJK*+
! & -:L-M % >?:@4AB5
LMCH,-./#*+N K*+ EF "# ,-./OP $%" K*+ " Q RST RU
VW @C,XCYVW '
OPZ[.6G4H\]C6Q%^_`4;ab.,;-:L
cde8 (
OP "# ,-./ K*+ QZ[f
R "# ,-./
QZ[f
R "# ,-./
g,-./#OP $%" h
OP *+
,-./e8 WYN*efcd ,-./01789:;-: ,-./01VWi:.64<= (
)&#* >?:@4AB52>bjG4k\4IbMIb
:.64 !
,-./01.:CAl4]j=m VWn "# WYN*]gh]Nij '
.6G4HYopCqrst )&#* >?:@4AB52>bjG4
k\4IbMIbD '
,-./01VWKXCYjG4; '
u u u u u u u '
u (
u u !
u u u u u u )**+,-./VW *+,-./VW ,-./01VW *+ "# ,-./5C6v4,pj6
Dw EF "# ,-./ >?:@4AB5.:CA
Rx4L*+n .6G4H.IJ ! & -:LMx4L*+n y $%" K "#% M4z{4
;x4L*+n '
"#% M4z{4; $%" y
>?:@4AB5LMCHI4Y-:L-MQf
N @C,XCYKx4L*+n (
! & -:L-MQZ[f
,-./
VWn QZ[f
,-./VWn R "#% M4z.IJ
x4L*+n l4 Rn l4 Rn $% .6G4H.IJ yQZ[f
R
,-./ $% .6G4H.IJ yQZ[f
! ,-./ OP,-./OP.6G4H |}n ~ ~ *+,-./789:;-: '
,-./0178 *+,€ *+!!!
,* >6‚ƒXQ ! „…_†‡
ˆ‰f
"#% M4zVWDw#R
Š"#
%"!!-%&!! :.64‹6-LŒ4Ibj .
.6G4HŽMX5C6v4,€]ŒX#‘
’“Q”•f
–%( VW#—8”•’“˜™Qš›f
–%(
,-./e8#R
œ0]ŒX'žf’“˜™#R
žŸ Š"#%"!!-%&!! :.64‹6-LŒ4
Ibj9:;Q ¡f
¢£(
"#% M4zVW |}K:.64Dw#R
Š"#%"!!-%&!! :.64‹6-LŒ4Ibj
Q_†‡
¢£(
œ "# ]ŒX "#%"!! %&!!
.IJQ % M4z¤¥–%(
2¦Œ4/.IJ.6G4H.IJ§¨
>?:@4AB5*+h„…©–N2¦Œ4/.6G4H.IJVW
%(
&ª5C6v4,65C,Q«¬[email protected]
a@:.§)*­®a4/¯°Qq346f–%( .6G4H R±
OPq4@± ,-./#.:CALMCHQ²³f
*+
`4;ab.,;-:LI4Y.6G4Ha@:.´DVW%
µh¶–%(žŸ ]ŒXQ ¡f
¢£(
"#% M4z )*#*+±
#*+
,
-./VW%µh¶–%(
·¸Ykh "#% M4zVW¹º»%(
/
)**+,-./
/
*+,-./
/
,-./01
·¼ µT¸YkžŸn#R
½¾f
–%(
)**+,-./VW ,-./´D{4; )* y % .6G4H.IJ#*+
–%(
kaab$WYN*cd( 2l`4mn
u #i¿À)**+,-./VWQÁf–%(
o kaab$WYN*cd
)**+,-./VW:.64ÂÃ6-LŒ4Ibj#
R
žŸ Š"#%"!!-%&!! :.64 0 6-L
Œ4IbjQ ¡f
¢£(
*+,-./VW ÄÅ>?:@4AB5.:CA>?LMC,Q
U
Q²³f
y % .6G4H.I
J± OP % .6G4H.IJ#´D,-./{4;
h*+–%( R± OP >?:@4AB5.:CAh
5C6v4,f
1 M4.ÆÇÂÃQRSÈ@C,4>?:
@4AB5>?LMC,VÉiÊQVWf–%(
ËÀUNÌÍ|}§± ,-./œq4@Î,-./§}ÏQeÐf
ÑÃ>?LMC,qLpC6# QÒÓ%#l4ÔhÕ#Öז%(.
:CAl4]jQf
OP,-./Q #*+¶–%(
l4Ô<؈Ù#R
žŸ Š
$% .:CAl4]j
9:;Q ¡f
¢£(
u #­ÚÀ *+,-./VWQÁf–%(
o ab$WYN*cd
žŸ#R
ÛÜ%Ý *+,-./VWÞQ ¡f
¢£
(
,-./01VW ÄÅ>?:@4AB5.:CA>?LMC,.:CA
l4]jQ Q²³f
OP,-./´D{4;h y % .6G4H.IJ#*+
–%(
u #,-./01VWnQÁf–%(
o WYN*efcd
q346
VW:.64˜™#R
žŸ ÛÜ%
Ý,-./01VWÞQ ¡f
¢£(
!
,-./VW#ßàf
áâ# ·23456h
㖖%(
/
!%% !% >?:@4AB5.6G4H.IJ
/
!%% % >?:@4AB5.:CA
/
)&#*)&!!- >?:@4AB5ä[email protected]åY/23€
æá
…€
/
)&#*)&!!- –N
!!- >?:@4AB5 23æᅀ
/
% >?:@4AB5 1 §LMCH
/
! aH/M]XI4Y& €-:L-M
/
! & ¸46o4å
$p$qrs=t
ab$WYN*cdXZ[O
0
uEX(/vmnw$cdXZ[O
0x][T]Osy
(z{sE0|}$~2€4

VW’“#R
]ŒX 23456#ç
è]ŒXQ ¡f
¢£(
*+,-./ ÄÅ>?:@4AB5 23 ÄÅ>?:@4AB5
.:CA>?LMC,Qx4L*+f
,-./VWµ%(.:C
A>?LMC,Q²³f
.6G4H.IJ*+QéÖf–%( *+V
W œ[email protected]@C,XCY.IJ—
8M4.QZ[%µh¶–%(
u # *+ "#% M4zVW5C6v4,pj6
Q R%DwQÁf–%(êZ5C6v4,ëe5C6v4,
R%(gpj6 23456ÎOìhËÀµh±'–
%(
!
"#
$#
o ab "# WYN*$TUOVW‚ƒ]Ow„;<
/
êZ5C6v4,íêZ5C6v4, ,-./{4;Q,-:X65C6
v4,#*+f–%(µ5C6v4, ,-./{4;#î'çïTN
:4q5C6 1 ÀÈa@:.VW
–%(
/
ëe5C6v4,{4;{4;,-./§*+€
íq4@Qðñ*+%
¨5C6v4,(,-./VWëe5C6v4,òÀó» g,./{4;hœ,-./{4;ôÍõöQ÷ø¶’“Qù%µ%(
/
.6G4HŽMX5C6v4,€
í2¦Œ4/.6G4H.IJ§
¨>?:@4AB5*+h„…©–N2¦Œ4/.IJ.6
G4H.IJVW(
&ª5C6v4,65C,Q«¬
[email protected]@:.§)*­®a4/¯°Qq346f–%(
~ #½¾
78=mh *+,-./#ú–%(,-.
/01VW78=m#R
ÛÜ%Ý,-./01789:
;-:ÞQ ¡f
¢£(Fû¿À 789:;-:žŸ#R
]ŒXQ ¡f
¢£(
… ab$WYN*†‡ˆ\ Y\]
!"#$%#&
q346
q4@P
,-./þXq
346
'(
#*+
{4;,-./Qüý !
–(
./;Xoq4@,-./þX„…_†‡
hq4@üýP ! QÿÀµ(
)**+VWq346
%
ó
Ëó "#% M4zVW *+VWq346
–%(q346
,-./þX#R
žŸ Š"#%"!!-%&!! Y-C6>4J9
:;Q ¡f
¢£(
Y-:M.6G4
H
>?:@4AB5ab.,X
G:h ( R–q346
–%(Y-:M.
6G4H R,-./OP,-./
[%µh¶–% *+,-./01V
W²³^_#R
ÛÜf–%€
( y.
6G4H.IJËÀUN[ö q346
–‡¥(N·VW q346
–%( y.6G4H.IJ# R,-./
OP.6G4H.IJ# R,-./ y
.6G4H.IJ#OP,-./(OP.6
G4H.IJ#OP,-./VW q346
–‡¥( _VWhq346
–%(
p
;.6G4H
R–>?:@4AB5LMCHhq346
–%(gLMCH# ! I4Y
-:L-Mh R–± !
I4Y¸46o4åh R–q346–%(
#*+
q4@ µTa@:.
QZ[¶–%(
45
#
.6G4H
01
*+,-./^_ =±'–‡¥(
>?:@4AB5
.:CAVW
,-./#ÄÅ.:CA>?:@4h=%(
>?:@4AB5
.:CAl4]j
,-./hœ,-./± ./;Xoq4
@ QZ[%^_ =%(
q346
>?:@4AB5
.:CA
!%% %(
!
… ab$WYN*†‡ˆ\ Y\]¡b ¢
!"#$%#&
'(
>?:@4AB5
.:CA>?4J
‚ƒX
!% @4Hk !% % @4
Hk ' !(
q346
>?:@4AB5
23
)&#*!!-!!-± !!-26
23æx4L€
( R,-./ 23
±=h±'–%(
)&#* ;-:@
@4Hk
@4Hk '!!( !(
!!! @4Hk '!!( !(
)&+# @4Hk
@4Hk ! )&+# :.64žŸ
#R
ÛÜ%Ý)&#* >?:@4AB
52>bjG4k\4IbMIb:.
64ÞQ ¡f
¢£(
>?4J‚ƒX@4
Hk
@4Hk !( !%
.6G4
H.IJ @4Hk>?4J‚ƒXQˆ‰
%=h±'–%(
¸þG4Ibj
.IJ
#*+
g,-./ !
,7"5"± !!!
,
*, SQˆ‰%µh¶–
%(!!! Qˆ‰f
R± OP
,-./ Qˆ‰f
R
± OP,-./ VWhq346
–%(
±
VW# ,*
*8
!h=
%(!!! VW# üßàq4.
C,h=%(üßàq4.C,#R
žŸ Š"#%"!!-%&!! Y
-C6>4J9:;Q ¡f
¢£(
!!!
q4.C,
!%
XY-:X.
q346
–‡¥(
ab$WYN*cd‰$ )&+# $\]NO^$Š‹sy
(
z{XŒ0Ž)&#* \‘’“T^x]”ƒP•–]—g”Sg”
$\]NO^˜2€4

45
#
#
™O!"#2WYN*š X›œ4
0l`/v
! X(Z‘ ! NOPQ•Ngž ( X$ 2Ÿ‡X "
¼ !,"5" ±
!!!
,
*, ¸þG4Ibj.IJ"#% M4z.
IJQ:.64%˜™<=Q½¾f–%('žŸÀ˜™ ÛÜf
–%(
WYN*2\]NO^0£s¤¥$Z\OWYN*¦§sl`
¨©$`ª2«'4
012¬­4
®$¯°$¨©$`ª
sy
( $±²³|´s6µum
>? !"#$¶·¸E(¹º2»E'Z¼N½¾´$¿œG1
s
x]ÀÁ*•NgžÂÃ$Ä2œG£sFÅ/0(J
KÆÇ2ÈÉ':Ê"# •NgžËÌÁ#^Ís/0ŽÎÏsw'
':$˜2п
"!#$%
·#"#%"!!-%&!! K ! :.64˜™
<=#R
½¾f–%(
"#%"!!-%&!! cd`ª2«'':ÑÒ$•Ngž
!"#s12391 x]OÓY¡ÔՕ–]¢
1 )”
NW Y\֏\‘’“T^ )”NW Y\Öp$x][T
]O2×Ø4
,-./.IJQ,-./VW#XCYjG4;%Dw#R
žŸ Š"#%"!!-%&!! :.64‹6-L
Œ4IbjQ ¡f
¢£(
WYN*S]ƒ„$•Ngž !"#2ÙÖ^ab4
ÂÚ 1 )”NW Y\Ös !"#ÛN$ 91 2w$Ü
µ Y\Ö2x]OÓY$ 314 —g”Sg”2w
Ÿ‡4
23 $ÝPÛ^Ÿ‡2œ
!,7"5" ÔÞPg”]ƒ•Ngž2ߚ
s\]NO^àះ4ߚ $âãàáäŸ 1 åæab
2Ÿ‡4çè 1 # PN2w
éê$ 1¡\]*TUOÕÓOx^¢
ZÖTUOsåæab2Ÿ‡4
#
âã(1 WY\#]Oabsw„
012ë4
äŸ(ì„WYN*åæab2ë4
:#
*& íNO#îÕ*'( :#
*& WYN*N\U’
2ïw„$ܵŠ‹sy
(Ê"#%"!!-%&!! \]
NO^ðOYÖ^•Ág”]ƒÍ2€4

ÂÚ¼)Ôx]OÓY$)‘\N Y\‘2\]NO^4
'
,*
*8
ñò2\]NO^4
üßàq4.C,#R
žŸ Š"#%"!!%&!! Y-C6>4J9:;Q ¡f
¢£(
(
ߚ s\‘’“T^ 23 „óÌ[O Y\‘2\]NO^4
)&#* \‘’“T^x]”ƒP•–]—g”Sg”2\]NO
^4
! 45
#55*
5
%
¡ %¢2ߚ s\]N
O^4ôõö4
45
#
#¡
#¢2ߚ s\]NO^4
š
s 45
#
5,/0( 45
#
;
2\]NO^4š
÷øNOPQ•Ngž$
91 PÛ^2Ÿ‡4
8;
2w„4
ãùNOPQ•Ngž$\
‘’“T^ )”NW Y\Ö$úg”•–]ûËUOà
á Y\Öüý$þv|
2œš
3 X Y\Öüý‚ƒ]O
2ÿv4
™O!"#2ߚ s\]NO^àះ4ôõö0
£sZ¼NúUW2ô\]NO^0l`/v
•Ngž$&Å2ÈÉ':Z¼NúUW2ô\]NO^0X
š 2ôõö4X
2ô\]NO^0ܵ Z¼NúUW2ô\
]NO^0l`/v
$
WYN*$-2¬­4
45
#
#¡#¢2WYN*Xö0àGŸ‡
4
' w$#ÕSٕ–]ÕӃYž2\]NO^àá‚UO#UÕ4
( x]™^÷ø 45
#
<= ™O
!"#2\]NO^4¡ÔՕ–]¢
Ê"#%"!!-%&!! \]NO^ðOYÖ^•Ág”]
ƒÍs/0WYN*)*•OsWYN*cd2{ :
"!#$%
·#"#%"!!-%&!! K !!! :.64˜™
<=#R
½¾f–%(˜™žŸ#R
g½¾f
–%(
"#% •S`ª2«'':ÑÒ$•Ngž !"#
s12391 x]OÓY¡ÔՕ–]¢
1 )”NW Y
\֏\‘’“T^ )”NW Y\Öp$x][T]O2×
Ø4ÑÒ$WYN*•Ngž2WYN*cds#UՃP 0;
<sy
$Š‹(
Ê"#%"!!-%&!! \]NO^ðO
YÖ^•Ág”]ƒÍ2€4

WYN*S]ƒ„$•Ngž !"#2ÙÖ^ab4
ÂÚ 1 )”NW Y\Ös !"#ÛN$ 91 2w$Ü
µ Y\Ö2x]OÓY$ 314 —g”Sg”2w
Ÿ‡4
23 $ÝPÛ^Ÿ‡2œ
ߚ s *+!!!,
*, ÔÞPg”]ƒ•Ng
ž
$ !!!,*
*8 $\]NO^Ÿ‡2œ
üßàq4.C,#R
žŸ Š"#%"!!%&!! Y-C6>4J9:;Q ¡f
¢£(
\]NO^BÕÓ]ÕO…3'ø ,* \^2’"U
W4
\]NO^4,* (ŒXŸ‡4
'
ߚ $âãàáäŸ 1 åæab2Ÿ‡4çè 1 # PN2w
éê$ 1 ZÖTUOsåæab2Ÿ‡4
âã(1 WY\#]Oabsw„
012ë4
äŸ(ì„WYN*åæab2ë4
:#
*& íNO#îÕ*'( :#
*& WYN*N\U’
2ïw„$ܵŠ‹sy
(Ê"#%"!!-%&!! \]
NO^ðOYÖ^•Ág”]ƒÍ2€4

(
ߚ X\‘’“T^ 23 „óÌ[O Y\‘2
4
)&#* \‘’“T^x]”ƒP•–]™O!"#2\]NO^
4
! 45
#55*
5
%
¡ %¢2ߚ s\]N
O^4
45
#
#¡
#¢2ߚ s\]NO^4
š
s 45
#
5,/0( 45
#
;
2\]NO^4š
÷øNOPQ•Ngž$
91 PÛ^2Ÿ‡4
!!!8
#; 2w„4
ãùNOPQ•Ngž$
\‘’“T^ )”NW Y\Ö$#ÕPUOúg”•–]
ûËUOàá Y\Öüý$þv|
2œßSÁž$
š
3 X Y\Öüý‚ƒ]O2ÿv4
î\hóUW)”NWs(WYN*)”NWSÁž$#UՃ
P (œuX*+ $ªX(WYN*)”NWs
z)”NWcdl`X
,* ™O!"#2ߚ s\]NO^àះ4ô
õö0£sZ¼NúUW2ô\]NO^0l`/v
,* ™O!"#2ô\]NO^0ܵ
!!! Z¼NúUW2\]NO^0l`/v
WYN*$-2¬­4
45
#
#¡#¢2WYN*Xö0àGŸ‡
4
' w$#ÕSٕ–]ÕӃYž2\]NO^àá‚UO#UÕ4
( 45
#
<= 2x]™^÷ø\]N
O^4¡ÔՕ–]¢
Ê"#%"!!-%&!! \]NO^ðOYÖ^•Ág”]
ƒÍs/0WYN*)*•OsWYN*cd2{ :
&'"()*
#"+),-.
ÄÅ >?:@4AB5.:CAQU
,-./{4;Q
% Z[.6G4H.IJ#*+¶–%(>?:@4AB5.
:CAQf
,-./{4;Q.6G4H.IJ#*+%8,./VW
#*+–%(
WYN*cdX(s *+ ,*
*8
ñò2\]NO^4
÷ø\‘’“T^N\U’2WYN*sab0l
`/v\]NO^$Š‹sy
(
Ê"#%"!!%&!! \]NO^ðOYÖ^•Ág”]ƒÍs/0WYN*$z\]NO
^ij2€4

,-./{4;Q % Z[.6G4H.IJ#*+%# .IJ#·ó»h=%(
/
!%%±
% >?:@4AB5.:CA yVW
#U
T#.:CAh=#Àµh±'–%€
!% \‘’“T^N\U’( ab$WYN*e
fcdX(Z[O
mn1ø$N\U’($ abW
YN*Xws
/
,-./{4;Q>?:@4AB5.:CA#*+%Næá­®M
Xa4/25,/2€x4L
/
>?:@4AB5.:CAQ % ab.,YopCqŽ,
o4HB"€.6G4H.IJ#*+%ND# 2
25,/ FD# 3 25,/hç
æáx4L
WYN*cdX(WYN*š 2\‘’“T^N\U’sa
b0s(qÙÖ^$¿¡ÙÖ^X(¢w„X ·¼ *+,-./VWx4L*+nQÁf–%(
/0 !12
345
u Á% #g,-./{4;Q´D>?:@4AB5.:CA#*
+fg>?:@4AB5.:CAQ,-./Z[.6G4H#±´D
% .6G4HYopCq#*+f–%(
o "# WYN* \‘’“T^N\U’ y$
ÙÖ^ab$
WYN*/0(N*] #Ó]Z‘Xs sab
0Ü
µN\U’$Ì]ƒl`XWYN*cdX$N\U’Ì]ƒ$w
„;<sy
$Š‹(z{ßs
N\U’$ÙÖ^ab
sy
$Š‹(
•NgžËÌÁ#^2€4

!""6 ! &'
"()* #6+2789: 5
% >?:@4AB5 1 LMCHQf
,-./ % M4
zVW "# ! & -:L-M±
! I4Y¸46o4å4Qq346¶–%(u #ÄÅ>?:@4AB5
.:CA>?:@4AB5LMCH ! & -:LMQUNq346
"#% M4zVWQÁf–%(
µVW g,-./{4; [%,-./ab.,M4.£ïÀ¢
@C,XCYa@:.@C,XCYo4
ab.,M4.#*+
–%(µT23456VW#%žŸ .6G4HI4Y
@C,XCY]ŒXQ ¡f
¢£(
WYN*cdXw$™O!"#$w„;<sy
(gՑUW
#UՙO!"#Û]îs6µum
"#
o NOPQ•Ngžàá ! & Y\ÖYS$
ÙÖ^ab$
% \‘’“T^ 1 ÖSUQ
! &
Y\ÖYSàá5ø$ÙÖ^(N\U’ÖSUW24
Ÿ‡
vÔՕ–](
mn5$':1ø$x][T
]O$ yÇ4'ܵgՑUW#UÕ !Ïs01/v
/0 ;< 5
ÄÅ >?:@4AB5.:CA>?LMC,Qf
"#
% M4z.IJh,-./q4@ RþX#*+
OP
.6G4H\]C6VWQq346f–%(µ^_ >6‚ƒX g
,-./#*+
R,-./{4;§ % Z[XG:
ab.,;-:L>ƒ4¸4@h¶–%(
= >?@ >,ABCD
,-./OP % .6G4HXG:Q%^_·=mQ
!Nf
µh=%(
/
/
RST R " h,-./{4;¶µ(
%
% .IJh>?4J‚ƒX"#$Qf
µ(.6G4H.IJ>?4J‚ƒXžŸ#R
%
&'Ã(K)_†‡¢£(
/
>?:@4AB5.:CA>?LMC,hÄÅÔ
µ(
/
45
#
#
>6‚ƒXh,-./{4;ˆ‰
˵(
u ' # y %" K "#% M4z{4;
x4L*+nQÁf–%(
%&(
%&'
(
(
'
'
)
*
o ' $ $%" ‰$ "#% •Sš $
ÙÖ^ab$
(+E+FGCD
,-./OP %" @C,XCYQ%^_·
=mQ!Nf
µh=%(
/
%
% .IJh>?4J‚ƒXMHkQ
f
µ(.6G4H.IJ>?4J‚ƒXžŸ#R
%&'Ã(K)_†‡¢£(
/
>?:@4AB5.:CA>?LMC,hÄÅÔ
µ(
u ( #"#% M4z{4; %" >?:@4AB5LMCHI4Y-:L-MQfN @C
,XCYKx4L*+nQÁf–%(
!
(
'
"#
(+
'+
'+
o ("#% •Sš $ $%" 
\‘’“T^ÖSUQgÕY\ÖYS2w„4' ‘U
W#UՉ$ÙÖ^ab$
;<HI!12JK+!=,LMNOPQRST
R"6UVWX
Q:.64%*#´D{4;STZ[.6G4Hh #~
ÁSé+%=h±'–%(g{4; Z&.6G4HXG:#±
`4;ab.,#X,p.%g{4; `4;ab.,;-:Lcd
Ó',
hF±=h±'–%(`4;ab.,;-:LcdÓ',
h
FSQéS# ·˜™Qˆ‰f
¢£(
ÙÖ^$j"#4÷2¬­4
8;
±
!!!8
#; \4
IbMIbhü- )&#*23 STX,p.º»À%
ab.,Q~
Áf·# ./ )&#*23 STX,p.º»Àab.,Q~Áf–
%(23 STX,p.º»Àab.,>?:@4AB5.:CA
"0346.1346#*+
ab.,h–~Á·#0
346.1346#*+
ab.,h~Á–%(
2f;-:LcdÓ',
±µQé+%# g{4;ü-#3
4 )&#*23 h.:CA#*+
.6G4HYo
"
pCq #*+
.:CA€
·#34 )&#*23 h
R
.:CA
.6G4HYopCq 3 #*+
.:CA€
#*+
µQé+f–%(
)”NWs Y\ÖüýSÁžYÛ^2þv|
{4; 56Q7'–%(
>
{4; 3 #;-:LcdQÓ',
–%(;-:LcdMŒ4J-8
QÍW%# 8;
±
!!!8
#; \4IbMIbQ–%(NMŒ4J
-8ÝMŒ4J "ÞQab., " #
ÝMŒ4J %ÞQab., %
#ÍWf–%(
*
{4; 3 56Q7'–%({4; 56Q9
8
;
±
!!!8
#; \4IbMIbQ
:ôf–%(
{4; #;-:LcdQÓ',
–%(
x4Lh2f¢*+
^_;-:L™. {4; 3 #À';-:LcdÓ',
{4; 3 ™.#À %(
ÝMŒ4J "Þh;-:Lcd " #–Ng{4;gab.,hD
wÓ',
T
µQé+f
MŒ4J-8;-:L™
;h<=#é+¶–%(ab.,;-:Lcdh2f¢~Á
;-:LcdÓ',
gZ[ab.,‰ÀU
¢£(
;< 5
% .6G4H.IJhˆ>
#OP
"# ,-./Q*+¶–%(q346
VWn# ·h±
'–%(
/
)**+,-./ RQ.6G4H.IJ y#*+%
/
,-./ R#OP % .6G4H.IJQ*+fN *
+VW
/
% .6G4H.IJ yQOP,-./#*+%
OP.6G4H.IJOP,-./VW q346
–‡¥(
#
OP "#% M4zVW >?:@4AB5.:CAl4]
jQf
QZ[¶–%(
>?:@4AB5
.:CAQf
l4]jQˆ>%µ#'>?:@4AB5.:C
A>?LMC,*+a@:.Q?Ã.:CA>?LMC,VW@A#†NÑÃj
4Yf
Bж–%>?:@4AB5.:CA>?LMC, y–
N OP>?:@4AB5.:CAVW–%€
(l4Qe8%C
ô¿SR\4DhE%µÀ¢l4a@:.#ßf
’“KX,p.h‰
À†–%(
ËÀ,-./§#l4Qe8f Qpj6#Òïµh¶–%(
\‘’“T^N\U’ÖSUW2$$WYN*/0
(WYN*Z‘àáNOPQ•Ngžãù4
0WYN*/0Ü
µÌ]ƒ(\‘’“T^N\U’sŸ‡0l`/v
l4Ô<؈Ù#R
žŸ Š
$% .:CAl4]j
9:;Q ¡f
¢£(
%&$ WYN*2'( cdXÌ]ƒ2w„0š
$$WYN*s)0NOPQ)‘\N2*+4v'·,ã
ùNOPQ•Ngž sab
0WYN*š -.÷ø(š /
0ab
0ãùNOPQ 3 Ž1·˜
(Emn]2w„
01È2X 4t
$š (gՑUW#UÕ)‘\N2­
3X "#% M4zVW ·l4]jDwQq346f
–%(
/
346v4;v:;F
µl4]jDwQ%# ,-./QFÚ# RR.:CA>?
LMC,#GHf–%(N#,-./QGHfNT%I#,-./Q
l4Ôf>?LMC,œ,-./Îq4@STX,p.À
#f–%(23456Qî'JNÛN>?:@4AB5
23.6G4HYopCq–N >?:@4AB5 1 LMCHÀÈ
Qî'JNÛ# ,-./l4QK8f–%(µl4]jDw
Q%l4QL‡#>?LMC,346ST346Kx4
LQRÀMJµh¶–%(
\‘’“T^ 23 '( % NOPQÕӂUZ
p$\‘’“T^NOPQx][T]O24v5·'ܵ
$
]2ôŸ‡0l`/v
/
?Ã.:CA>?LMC,346.1
µl4]jDw q4@
!% .6G4H.IJ–
N
% >?:@4AB5 1 LMCHQ>?:@4AB5
.:CA#*+f
x4LQNO8
l4PQ#±
346#RÀMJµ ¶–‡¥(x4LQl4PQ346#R
ÀMJ^_,-./l4QK8%=h±'–%(µl4]
jDwQ l4QL‡#23456Qî'Jµh
¶–%(
67‡8
0]$ÂÃs/0[O÷ø67‡8
0
]$9Ãs/0[OsÙÖ^2y:5·0ܵ]2ôŸ‡
0l`/v
=HIYB-N > u #
! & -:L-MQZ[f
,-./VWnQ
Áf–%(
)*& ,
"#
,"#
)*& o ! & Y\ÖYS2ãù4
0 WYN*cd$
u n# ·VW’“hã–
–%(
.:CA>?LMC, /
!".:CA VW
–%(
l4 Rq4@ q4@ % .6G4H.IJ?€
>?:@4AB5 1 §LMCH
/
l4 Rq4@ q4@ % .6G4H.IJ?€
>?:@4AB5 1 §LMCH
.:CA>?LMC, /
!".:CA VW
–%(
l4 Rq4@ q4@ % .6G4H.IJ
?€
/
l4 Rq4@ q4@ % .6G4H.IJ
?€
*ÜVWl4]jQe8%# ·˜™Qˆ‰f
¢£(
% NOPQ•Ngž¡?¢
Z‘ àáZ‘ $¨©
;
012¬­4
N\U’ÖSUW s] 2N\U’ÖSUW s] 2
Ÿ‡4
% NOPQ•Ngž¡?¢
Z‘ àáZ‘ $¨©
2<
N\U’ÖSUW s] 2N\U’ÖSUW s] 2
Ÿ‡4
>?:@4AB5.:CAl4VWžŸ#R
Š
$% l4]j9:;Q ¡f
¢£(
=HIYB-N ! > u ! #
! & -:L-MQZ[f
,-./VWnQ
Áf–%(
)*& )*& ,
"#
-./
01!
)*& -
o !
2ãù4
0 WYN*cd$
u ! n# ·VW’“hã–
–%(
.:CA>?LMC, /
!".:CA VW
–%(
l4 Rq4@ q4@ % .6G4H.IJ?€
>?:@4AB5 1 §LMCH
/
l4 Rq4@ q4@ % .6G4H.IJ?€
>?:@4AB5 1 §LMCH
/
l4 Rq4@ q4@ % .6G4H.IJ?€
>?:@4AB5 1 §LMCH
.:CA>?LMC, !".:CA VW
–%(
/
l4 Rq4@ q4@ %?€
/
l4 Rq4@ q4@ %?€
.:CA>?LMC, .:CA VW
–%(l4Ô =±'
–‡¥(
>?:@4AB5.:CAl4VWžŸ#R
Š
$% l4]j9:;Q ¡f
¢£(
1$ WYN*cds(
% \‘’“T^N\U’$=
NÙ abl`X
% \‘’“T^N\U’2 yw
=1$cd2Ÿ‡X N\U’$=NÙ ab$Š‹sy
( $% \‘’“T^N\U’ËÌÁ#^2€4

*ÜVWl4]jQe8%# ·˜™Qˆ‰f
¢£(
% NOPQ•Ngž¡? ?¢
Z‘ Z‘ Z‘
àáZ‘ $¨©;
012¬­4
N\U’ÖSUW s] 2N\U’ÖSUW s] 2
Ÿ‡4
% NOPQ•Ngž¡?¢
Z‘ àáZ‘ $¨©
2<
N\U’ÖSUW s] 2N\U’ÖSUW s] 2
Ÿ‡4
% NOPQ•Ngž¡?¢
Z‘ àáZ‘ $¨©
2<
N\U’ÖSUW s] 2Ÿ‡4
Z7;< [,L89B-N;
< WX
l4]j*ÜfN€QU
OP.6G4H\]C6±
"#% M4z.IJhœ,-./ QZ[¶–%(
*ÜfN~ #$‘
l4]j789:;-:hú
–%(
u #"#% M4z,-./SRQUN VWx4L
*+QÁf–%(,-./ R OP %" Qf
R,-./
%" Q Rf
–%( @C,XCY
´D,-./Á
–%(
)*& ,
"#
$
,
"#
$
)*& o X$ y$ "#% •S•Ngž$
ÙÖ^ab$
45
#
#
! >6‚ƒXQU
% .6G4H.IJQOPÒT.6G4H.IJ#4Ib
kÒÓ¶–%(g.6G4H.IJ#
R± OPÑÃ\]C6
.1&@€h±'g.6G4H.IJQ {4; ,-./#Ó',
µh¶–%(OP {4; ,-./ ,-./01VW¤
–%(µTVW# 45
#
#
! !>6‚ƒX)&#* >?:@4AB52>bjG4
k\4IbMIb>?:@4AB5.:CAl4]jh=%(
VWn˜™#R
·¼½¾f–%()&+# :.64žŸ#R
ÛÜ%Ý)&#* >?:@4AB52>bjG4k\4
IbMIb:.64ÞQ ¡f
¢£(
\XG]"R"
~ #,-./01VW===mQÁf–%(
… WYN*ef†‡
!"#$%#&
'(
q346
,-./P
,-./þX ! „,-./{4; !€
(
,-./þX
q346
)&#*!!- ± !!-2623 U‘
q346
óq4@þX(žŸ#R
Š"#%"!!-%&!! Y
-C6>4J9:;Q ¡f
¢£(
Y-:M.6G4H
>?:@4AB5ab.,XG:(
p
;.6G4H
R–>?:@4AB5LMCHhq346
–%(gLMCH# ! I4
Y-:L-Mh R–± ! I4Y¸46o4åh R–q346–
%( #*+
q4@ µTa
@:.QZ[¶–%(
45
#
.6G4H
01
,-./01VW @4Hk ! h=%(
>?:@4AB5
.:CAVW
ÄÅ.:CA>?LMC,h=%(
>?:@4AB5
.:CAl4]j
g,-./þX.6G4HXG:#=%(
q346
>?
:@4AB5
.:CA
% %(
… WYN*ef†‡¡b ¢
!"#$%#&
'(
>?:@4AB5
.:CA>?4J‚ƒX
@4Hk ' !(
q346
>?:@4AB5 23
)&#*!!-2623æáx4L€
(
)&#*!!-2623æáx4L€
(
)&#* ;-:@@4
Hk
@4Hk '!!( !(
!!! @4Hk '!!( !(
)&+# @4Hk
@4Hk ! )&+# :.64ž
Ÿ#R
ÛÜ%Ý)&#* >?:@4
AB52>bjG4k\4IbMIb:
.64ÞQ ¡f
¢£(
>?4J
‚ƒX@4Hk
@4Hk !( !(
¸þG4Ibj
.IJ
y % .6G4H.IJ#*+
%
,-./ ,
!7"5"± !!!
,
*, Qˆ‰f
=h±'–%(
,-./¸þG4Ibj.IJO q
346
–‡¥(
±
!!!
q4.C,
VW# ,*
*8
!h=
%(!!! VW# üßàq4.
C,h=%(üßàq4.C,#R
žŸ Š"#%"!!%&!! Y-C6>4J9:;Q ¡f
¢£
(
!%
XY-:X.
q346
–‡¥(
ab.,
g,-./ >?:@4AB5ab.
,XG:#VCab.,pC6QWU
œ,-./ab.,# X,p.¶–‡¥(
q346
,-./01VW üý ! ,-./±
! XYq4@„…_†‡VW
–
%(N R,-./! yq4@€
! y./;Xoq4@_Z ! yq4@
VW%µh¶–%(
./;Xoq4@
,-./01.6G4H.IJ./;Xo
q4@,-./q4@þXO q346
–‡¥(
!
… WYN*ef†‡¡b ¢
!"#$%#&
GH>6‚ƒX
XYMx4k
Yoj-J
'(
45
# % @4Hk
!! ! !!! @4Hk !
(
45
#
#
#€
@4Hk ! !(
45
#
5, @4Hk
! !± 45
#
;
@4Hk ! !(
CD^_"!#$%
WYN*2\]NO^0£s¤¥$Z\OWYN*¦§sl`
¨©$`ª2«'4
012¬­4
®$¯°$¨©$`ª
sy
( $±²³|´s6µum
>? !"#$¶·¸E(¹º2»E'Z¼N½¾´$¿œG1
s
x]ÀÁ*•NgžÂÃ$Ä2œG£sFÅ/0(J
KÆÇ2ÈÉ':Ê"# •NgžËÌÁ#^Ís/0ŽÎÏsw'
':$˜2п
· ,-./01#e8NOP "#% M4z,-./Qe8%
#=À˜™<=%(:.64žŸ#R
"#
% M4z]ŒX
.6G4H.IJ]ŒXQ ¡f
¢£(
,-./01#e8NOP "#% M4z,-./Qe8%# ·˜™Qˆ‰f
¢£(
t
$x][T]OÊ"#%"!!-%&!! \]NO
^ðOYÖ^•Ág”]ƒÍX>?
0`ª†@àáz{$ß
X>?
05ñ9$`ªs=A4
012¬­4
"#% •SËÌÁ#^X>?
0BCPÛ^$cd
!"#Ÿ‡2œ
NOPQ•NgžËÌÁ#^X>?
0àGsNOP
Q•Ngž2Ÿ‡4
"
ß 23 $ "$* < D2EF4
9 $ "$* < D2
G4t
$D$µE ! 2H·IßD ! 2J·àGs4
"$* < D$EF;<G;<$ijsy
(
•NgžËÌÁ#^2€4

$t
$ % NOPQ•Ngžsab
0t
$\‘’“T^$åæab2KL4
ßZ‘sM$Ä2œ
,!7"5"<,*
*8
!±
ßà%üq4.C,± !!!,
*,
Q:.64f–%(ßàq4.C,#R
žŸ Š
"#%"!!-%&!! Y-C6>4J9:;Q ¡f
¢
£(
>
)&#* ;-:@Q:.64f–%(
*
)&+# 2>bjG4k\4IbMIbQ:.64f–%(
)&+# :.64žŸ#R
ÛÜ%Ý)&+#
>?:@4AB52>bjG4k\4IbMIb:.64Þ
Q ¡f
¢£(
45
# % Q:.64f
q4@QK:ôf–%(
,-./|} 45
#
#
>6‚ƒX

#€Q:.64f–%(:.64[#<,>
*+,,;Aµq4@Q *+
, |}f–%S€ \)#]B^QAƒC,f–
%(
+
'
QBC6å‚f
q4@56Q7'–%(
Xt
$Z‘2Ÿ‡4Nu'øt
$\‘’“T^åæ
ab2 % NOPQ•Ngžsôab4
(
Ds IyZ‘$¨©2<
\‘’“T^N\U’$Ì]
ƒ2Ÿ‡4ß](WYN*Þ# y¡Z‘ ¢
%àᏐ\‘’“T^ 1 O$ÖSUQXcd
¡] ‘UW#UÕsØ4
0ܵ¢
#
l4]je8#R
žŸ ]ŒXQ ¡f
¢£(
#
;
2 #
š Xõ
ö4ßWYN*X,:5# ÔՕ–]2w
P;$WY
N*š s &@ 2þv|
y$)”NW2%&$Z‘sþv|
0s( ,:5# ÔՕ–]2lIw„4

žŸ 45
#
#
]ŒXQ ¡f
¢£
(
! t
$Z‘X 2•“UOî!]4
ßWYN*s ™O!"#2\]NO^4ßWYN*s(QR
$)”NW‚UO/v$X$ "#% •S\]
NO^ijsS
¡)”NW$ûËUO Y\Öüý$þ
v|
™O!"#$\]NO^pX¢
45
#
# 2 y$WYN*$P;$š s\]NO^
4# „ :*,* WYN*S™N2d4
)&#* >?:@4AB52>bjG4k\4IbMIbQ:.64%
# ·˜™Qˆ‰f
¢£(
Z‘$¨©2<0÷ôõö4
94
Y\Ös %><
@ 2T<4
%><
@ ÷øka#ÕSٕ–]ÕӃY
ž2Ó X )&#* \‘’“T^x]”ƒP•–]—g”Sg”2õö4
x][T]O$2ÐnX÷øUM‰V2WSUW4
Wè$û^î2XY4
UM‰V2WSUW4
U)&+#55*
VàáU)&+##V2XY4
UM‰V2W
SUW4
$
)&#* >?:@4AB52>bjG4k\4IbMIb:.64
Yoj-Jh:ôf’“_`h~Á–%(
' UM‰V2WSUW4
(
…3'2ÐnX÷øUM‰V2WSUW4
Wè$û^î2XY4
UM‰V2WSUW4
! ÕӃYžû^î2XY4
UM‰V2WSUW4
+#
# :.64Yoj-Jh:ôf’“_`h~Á
–%(
UM‰V2WSUW4
…3'2ÐnX÷øUM‰V2WSUW4
Wè$û^î2XY4
UM‰V2WSUW4
U!ZV2WSUW4
11!,*,C! \]NO^ÕӃYžõ
ö4'ø
UM‰V2WSUW4
w„[\{2ÐnXU]0V2WSUW4
' Z¼N$^_24
012¬­4
( M$x][T]OP;ãXY
012¬­4
D
1!,*,
D
1!
`;ÕӃYžû^î2XY4
UM‰V2WSUW4
! U4CV2WSUW4
•Ngž2ôõö4
:.64habfXYMx4kYoj-Jh¶ #À'–%(
)&+# DwžŸ#R
ÛÜ%Ý)&#* >?:@4AB
52>bjG4k\4IbMIbDwÞQ ¡f
¢£(
"+)`Ka%b
,-./01VW# >?:@4AB5.:CAl4]jh=%(·n
q346
VWl4VWÍWDwQ½¾f
–%(
n íg,-./{4;þX>?:@4AB5 1 §LMCH
.6G4H.IJc'#l4QÍWf–%(N y
.6G4H.IJ#*+
R,-./VW u #Á% # Rl4he8¶–%(
)*& ,
)*& 234./
$56
o ] y$
n í y .6G4H.IJ#*+
R,-./V
W u #Á% # Rl4he8¶–%(
)*& ,
"#
,"#
234
./
$56
)*& o ] y$
!"#
¼ q346
*+,-./01VWVWnQÁf–%(
u # R ,-./h y % .6G4H.IJ
%5 >?:@4AB5.:CAþXQZ[f
VWnQ
Áf–%(u # y % .6G4H.IJüýVW
nQÁf–%(u # ËÀUN % .6G4H.IJQU
N ,-./ËÀUNj4Y„…_†‡QÁf–%(
u # R ,-./h y % .6G4H.IJQ
Z[f
VWQÁf–%(µT R ,-./ R
! ! I4Y-:L-M % >?:@4
AB5LMCHQU
Q²³f
Z[f
–%(µVW# % 346>?:@4AB5.:CAh Rl4Q RÍW
%µh=%(gl4# XY {4; ´Dq4@€
>?
:@4AB5 1 §LMCH % .6G4H.I
Jhã–
–%(
!
"#
234
./
$56
!
,"#
o $% NOPQ•Ngž 2ãù4
0 y$ W
YN*
u #! ,-./h y % .6G4H.IJQ
Z[f
VWQÁf–%(µT ! ,-./ R
! ! I4Y-:L-M % >?:@4
AB5LMCHQU
Q²³f
Z[f
–%(µVW# R
>?:@4AB5.:CAl4Q ! ÍW%µh=%(gl4#
XY {4; ´Dq4@€
>?:@4AB5 1 §L
MCH % .6G4H.IJhã–
–%(
!
"#
7
,"#
7
o $% NOPQ•Ngž 2ãù4
0 ! $ WY
N*
u # ,-./ RVWQÁf–%(,-./dST – y
% .6G4H.IJQZ[f,-./ .IJQW×,-./ VC.6G4H
.6G4H.IJ yQU
–%( R%
,-./ R ! ! I4Y-:LM % >?:@4AB5LMCHQU
Q²³f
Z[f
–%(µVW# % 346>?:@4AB5.:CA
h Rl4Q RÍW%µh=%(ü- Rgl4# ,-./ R{4;>?:@4AB5 1 §LMCH
% .6G4H.IJhã–
–%(µT R
l4 uef#± R,-./ y.6G4H.IJ%(
R/l4 ugf#±,-./ R{4;>?:@4AB5
1 §LMCHLMCH.6G4H.IJ yQ㥖%(
74+
!
7
!
7
7
74+
89:4+
,
"#
o %&WYN*%&NOPQ 74+
!"
·¼ "#% M4zVWIh.˜™Qš›
f
–%(
!12Fcd+efgh #"$ &'"()*
&Ta1iJ3T3T,-.
#
)&#* >?:@4AB52>bjG4k\4IbMIb .6G4HY
opCqrs[#f–%(rsfN.6G4HYopCqhf.6G4
HYopCq#ijN^_)&#*%><
2>bjG4k\4
IbMIbh 23 >?4J‚ƒX#k
v4;v:;F0Fl
QLf.6G4HYopCqrst#q4@QK:ô%=h±'–‡¥(
µL ,-./´D{4;‰À†=h±'–%(
`4;‚ƒXhmLN# )&#*%><
2>bjG4k
\4IbMIbQˆ‰f
ijNa@:.#2fe8QÍW%=h
±'–%(
rsfN.6G4HYopCqhl4FQ^_.6G4HYopCqij
Û % .:CAŽ:MX.l4e8’“QL%=
h±'–%(žŸ Š
.IJl4]j9:;Q ¡f
¢£
(
CD5E+Fa1R
· % .6G4H.IJQU
"#
% M4zVW#,-./QGH%#=À˜™<=%(
z{X>?4
0àGst
$x][T]O%&$WYN*cd$
`ª†@2«'4
012¬­4

ÑÒ$WYN*$ߚ XM$ij2›œ4
)&#* ;-:@% ;-:@.6G4H>?4J‚ƒXh ! G8!Qˆ‰f
µQé+f–%(ˆ‰f
À^_;
-:@>?4J‚ƒXQ2f@4Hk#Lf–%(
>
:.64n…À^_)&#*%><
2>bjG4k
\4IbMIbQ:.64f–%(
!
)&+# :.64žŸ#R
*ÜfNÝ)&#*
>?:@4AB52>bjG4k\4IbMIb:.64Þ
Q ¡f
¢£(
*
,!7"5" # ]26o45^
]a@:.^XYGC6Qopf–%(
]8^./46XC
Ye8Q]E;^ST]
>^#mLf–%(
!!! # ]:2¦Œ4/^a.,6CYX:2Qg,
MC,f
]YoIb^Q3:6f
ST]`4;‚ƒX^/LQop
f–%(·#]a@:.54HB^Q,MC,f–%(
]~Á^Q,MC
,f
]qfa@:.Q~Á%^Qopf–%(
]Y-jX;YG:
;-:@^QǶ–%(
]8^Q,MC,f
ST
];-:@^
/LQopf–%(
]8^./46XCYe8Q]E;^S
T]
>^#mLf–%(
,!7"5" # ]26o45^
]q4.^XYGC6Qopf–%(
],*^./46
XCYe8Q];
*^ST]
>^#mLf–%(
!!! # ]./46^Qopf
]Yoj-J^Q3:6
f
]ÂÃr4^Q3:6f–%(·#]q4.^Q3:6f–
%(
],*^./46XCYe8Q];
*^ST
]
>^#mLf–%(
,-./|} 45
#
#
>6‚ƒX

#€Q:.64f–%(:.64[#<,>
*+,,;Aµq4@Q *+
, |}f–%S€ \)#]B^QAƒC,f–
%(
+
QBC6å‚f
{4;56Q7'–%(
l`sab
N\U’ÖSUW‰N\U’àá×Ø$Z‘2Ø
·
Ê"#%"!!-%&!! \]NO^ðOYÖ^•Ág”]
ƒÍsS
!"#2
NOPQ•Ngž$ËÌÁ#^s
S
NOPQ•Ngž2Ÿ‡4
"
ß 23 $ "$* < D2EF4
9 $ "$* < D2
G4t
$D$µE ! 2H·IßD ! 2J·àGs4
"$* < sZtDwmLDw˜™#R
.
IJ]ŒXQ ¡f
¢£(
% NOPQ•Ngžsab
0ÙÖ^$¿24v94
'
ßZ‘XM$Ä2œ
,!7"5",*
*8
!±
ßà%üq4.C,± !!!,
*, Q
:.64f–%(ßàq4.C,#R
žŸ Š
"#%"!!-%&!! Y-C6>4J9:;Q ¡f
¢
£(
>
)&#* ;-:@Q:.64f–%(
*
)&#* >?:@4AB52>bjG4k\4IbMIbQ:.
64f–%(
)&+# :.64žŸ#R
*ÜfNÝ)&#*
>?:@4AB52>bjG4k\4IbMIb:.64Þ
Q ¡f
¢£(
45
# % Q:.64f
q4@QK:ôf–%(
,-./|} 45
#
#
>6‚ƒX

#€Q:.64f–%(:.64[#<,>
*+,,;Aµq4@Q *+
, |}f–%S€ \)#]B^QAƒC,f–
%(
+
(
QBC6å‚f
q4@56Q7'–%(
t
$Z‘2Ÿ‡4Nu'øt
$ÙÖ^2 % N
OPQ•Ngžsôab4
t
$Z‘$¨©2 Ds Iy<\‘’“T^N\U
’$Ì]ƒ2Ÿ‡4ß]( y$WYN*Þ#¡Z‘ #
¢
%àᏐ\‘’“T^ 1 O$ÖSUQXc
d
¡] ‘UW#UÕsØ4
0ܵ¢
l4]je8#R
žŸ ]ŒXQ ¡f
¢£(
! #
;
2 #
š Xõ
ö4,:5# ÔՕ–]2w
ÑÒ$ "#
% •S•NgžXwu
067$)”NW2P;$š sc
pvþv|
ß×ØWYN*$P;$š s$$)”NW2þv|
ÑÒ$ "#% •S•Ngž$ߚ XM$ij2ÿv
4
,!"5" # ]26o45^
]a@:.^XYGC6Qopf–%(
]8^./46XC
Ye8Q]
>^ST]E;^#uf–%(
!!! # ]:2¦Œ4/^a.,6CYX:2Qg,
MC,f
]YoIb^Q3:6f
ST]`4;‚ƒX^/LQop
f–%(·#]a@:.54HB^Q,MC,f–%(
]~Á^Q,MC
,f·#]qfa@:.Q~Á%^Qopf–%(
]Y-jX;Y
G:;-:@^QǶ–%(
]8^Q,MC,f
ST;-:
@/LQopf–%(
]8^./46XCYe8Q]
>^
ST]E;^#uf–%(
>
,!"5" # ]26o45^
]q4.^XYGC6Qopf–%(
],*^./46XCYe8Q]
>^ST];
*^
#uf–%(
!!! # ]./46^Qopf
]Yoj-J^Q3:6
f
]ÂÃr4^Q3:6f–%(·#]q4.^Q3:6f–
%(
],*^./46XCYe8Q]
>^ST
];
*^#uf–%(
*
{4;QK:ôf–%(
t
$×ØZ‘X 2•“UOî!]4
ß×ØWYN*s ™O!"#2\]NO^4
$
g,-./# NOVCab.,pC6h±'–%&ª
"#% M4z.IJ:.64˜™#vU
¢£
ab.,>4C6;-:LcdÓ',
>6‚ƒX:
.64ÀÈ%€
(
#$
y± y " VW'
*+,-./78
9:;-:
*+,-./VW
q346
>?:@4AB5
.:CAD
!!!
8
#; XYGC6
:.64
! & >
E %%><
1
LMCH
! & &>
E %%><
L
MCH *+,./K*+
$%
!!!8
#;(
:.64
8;
(
,*
*8!
%
,-./VW
OP *+,-./e8
QZ[f
R
"# ,-./
" Q RST RU
VW
QZ[f
R
"# ,-./
,-./0178
9:;-:
,-./01VW
,-./01VWK
XCYjG4;'
x4L*+
OP "# ,-./
K*+
q346
,-./01VW
VWn
.:CAl4]j=m(
l4]j!
Dw!
EF "# ,-./OP
$%" K
*+
&
`4;ab.,;-:L
OPZ[.6G4H.IJ
;-:Lcde8(
OP *+,-./e8
www.dell.com
support.dell.com