HP StorageWorks x9720 Administrator's Manual

HP StorageWorks X9720 Network Storage System Administrator Guide
Abstract
This guide describes tasks related to cluster configuration and monitoring, system upgrade and recovery, hardware component replacement, and troubleshooting. It does not document X9000 file system features or standard Linux administrative tools and commands. For information about configuring and using X9000 Software file system features, see the HP StorageWorks X9000 File Serving Software File System User Guide.
This guide is intended for system administrators and technicians who are experienced with installing and administering networks, and with performing Linux operating and administrative tasks.
HP Part Number: AW549-96023 Published: April 2011 Edition: Seventh
© Copyright 2009, 2011 Hewlett-Packard Development Company, L.P.
Confidential computer software. Valid license from HP required for possession, use or copying. Consistent with FAR 12.211 and 12.212, Commercial Computer Software, Computer Software Documentation, and Technical Data for Commercial Items are licensed to the U.S. Government under vendor's standard commercial license.
The information contained herein is subject to change without notice. The only warranties for HP products and services are set forth in the express warranty statements accompanying such products and services. Nothing herein should be construed as constituting an additional warranty. HP shall not be liable for technical or editorial errors or omissions contained herein.
Acknowledgments
Microsoft® and Windows® are U.S. registered trademarks of Microsoft Corporation.
UNIX® is a registered trademark of The Open Group.
Warranty
WARRANTY STATEMENT: To obtain a copy of the warranty for this product, see the warranty information website:
http://www.hp.com/go/storagewarranty
Revision History
DescriptionSoftware
Version
DateEdition
Initial release of the X9720 Network Storage System.5.3.1December 2009First
Added network management and Support ticket.5.4April 2010Second
Added management console backup, migration to an agile management console configuration, software upgrade procedures, and system recovery procedures.
5.4.1August 2010Third
Revised upgrade procedure.5.4.1August 2010Forth
Added information about NDMP backups and configuring virtual interfaces, and updated cluster procedures.
5.5December 2010Fifth
Updated segment evacuation information.5.5March 2011Sixth
Revised upgrade procedure.5.6April 2011Seventh
Contents
1 Product description...................................................................................11
HP X9720 Network Storage System features...............................................................................11
System components.................................................................................................................11
HP X9000 Software features....................................................................................................11
High availability and redundancy.............................................................................................12
2 Getting started.........................................................................................13
Setting up the X9720 Network Storage System...........................................................................13
Installation steps................................................................................................................13
Additional configuration steps.............................................................................................13
Logging in to the X9720 Network Storage System.......................................................................14
Using the network..............................................................................................................14
Using the TFT keyboard/monitor..........................................................................................14
Using the serial link on the Onboard Administrator.................................................................14
Booting the system and individual server blades.........................................................................14
Management interfaces...........................................................................................................15
Using the GUI...................................................................................................................15
Customizing the GUI..........................................................................................................18
Adding user accounts for GUI access...................................................................................19
Using the CLI.....................................................................................................................19
Starting the array management software...............................................................................19
X9000 client interfaces.......................................................................................................19
X9000 Software manpages.....................................................................................................20
Changing passwords..............................................................................................................20
Configuring ports for a firewall.................................................................................................20
HP Insight Remote Support software..........................................................................................22
3 Configuring virtual interfaces for client access..............................................23
Network and VIF guidelines.....................................................................................................23
Creating a bonded VIF............................................................................................................23
Configuring standby backup nodes...........................................................................................23
Configuring NIC failover.........................................................................................................24
Configuring automated failover................................................................................................24
Example configuration.............................................................................................................24
Specifying VIFs in the client configuration...................................................................................24
Support for link state monitoring...............................................................................................25
4 Configuring failover..................................................................................26
Agile management consoles....................................................................................................26
Agile management console modes.......................................................................................26
Agile management consoles and failover..............................................................................26
Viewing information about management consoles..................................................................27
Cluster high availability...........................................................................................................27
Failover modes..................................................................................................................27
What happens during a failover..........................................................................................27
Setting up automated failover..............................................................................................28
Identifying standbys for file serving nodes.........................................................................28
Identifying power sources...............................................................................................28
Turning automated failover on and off..............................................................................30
Contents 3
Manually failing over a file serving node..............................................................................30
Failing back a file serving node...........................................................................................31
Using network interface monitoring......................................................................................31
Setting up HBA monitoring..................................................................................................33
Discovering HBAs..........................................................................................................33
Identifying standby-paired HBA ports...............................................................................34
Turning HBA monitoring on or off....................................................................................34
Deleting standby port pairings........................................................................................34
Deleting HBAs from the configuration database.................................................................34
Displaying HBA information............................................................................................34
Checking the High Availability configuration.........................................................................35
5 Configuring cluster event notification...........................................................37
Setting up email notification of cluster events..............................................................................37
Associating events and email addresses................................................................................37
Configuring email notification settings..................................................................................37
Turning email notifications on or off......................................................................................37
Dissociating events and email addresses...............................................................................37
Testing email addresses......................................................................................................38
Viewing email notification settings........................................................................................38
Setting up SNMP notifications..................................................................................................38
Configuring the SNMP agent...............................................................................................39
Configuring trapsink settings................................................................................................39
Associating events and trapsinks..........................................................................................40
Defining views...................................................................................................................40
Configuring groups and users..............................................................................................41
Deleting elements of the SNMP configuration........................................................................41
Listing SNMP configuration information.................................................................................41
6 Configuring system backups.......................................................................42
Backing up the management console configuration.....................................................................42
Using NDMP backup applications............................................................................................42
Configuring NDMP parameters on the cluster........................................................................43
NDMP process management...............................................................................................43
Viewing or canceling NDMP sessions..............................................................................43
Starting, stopping, or restarting an NDMP Server..............................................................44
Viewing or rescanning tape and media changer devices.........................................................44
NDMP events....................................................................................................................45
7 Creating hostgroups for X9000 clients.........................................................46
How hostgroups work..............................................................................................................46
Creating a hostgroup tree........................................................................................................46
Adding an X9000 client to a hostgroup.....................................................................................47
Adding a domain rule to a hostgroup........................................................................................47
Viewing hostgroups.................................................................................................................47
Deleting hostgroups................................................................................................................48
Other hostgroup operations.....................................................................................................48
8 Monitoring cluster operations.....................................................................49
Monitoring the X9720 Network Storage System status.................................................................49
Monitoring intervals...........................................................................................................49
Viewing storage monitoring output.......................................................................................49
4 Contents
Monitoring the status of file serving nodes..................................................................................49
Monitoring cluster events.........................................................................................................50
Viewing events..................................................................................................................50
Removing events from the events database table....................................................................51
Monitoring cluster health.........................................................................................................51
Health checks....................................................................................................................51
Health check reports..........................................................................................................51
Viewing logs..........................................................................................................................54
Viewing and clearing the Integrated Management Log (IML).........................................................54
Viewing operating statistics for file serving nodes........................................................................54
9 Maintaining the system.............................................................................56
Shutting down the system.........................................................................................................56
Shutting down the X9000 Software......................................................................................56
Powering off the X9720 system hardware..............................................................................56
Starting up the system.............................................................................................................57
Powering on the X9720 system hardware..............................................................................57
Starting the X9000 Software...............................................................................................57
Powering file serving nodes on or off.........................................................................................57
Performing a rolling reboot......................................................................................................57
Starting and stopping processes...............................................................................................58
Tuning file serving nodes and X9000 clients...............................................................................58
Migrating segments................................................................................................................60
Removing storage from the cluster.............................................................................................60
Maintaining networks..............................................................................................................62
Cluster and user network interfaces......................................................................................62
Adding user network interfaces............................................................................................62
Setting network interface options in the configuration database................................................63
Preferring network interfaces................................................................................................63
Unpreferring network interfaces...........................................................................................64
Making network changes....................................................................................................64
Changing the IP address for a Linux X9000 client..............................................................64
Changing the IP address for the cluster interface on a dedicated management console..........65
Changing the cluster interface.........................................................................................65
Managing routing table entries.......................................................................................65
Deleting a network interface...........................................................................................66
Viewing network interface information..................................................................................66
10 Migrating to an agile management console configuration............................67
Backing up the configuration....................................................................................................67
Performing the migration..........................................................................................................67
Converting the original management console node to a file serving node hosting the agile management
console.................................................................................................................................70
11 Upgrading the X9000 Software................................................................71
Automatic upgrades................................................................................................................71
Manual upgrades...................................................................................................................72
Preparing for the upgrade...................................................................................................72
Saving the node configuration.............................................................................................73
Performing the upgrade......................................................................................................73
Restoring the node configuration..........................................................................................74
Completing the upgrade.....................................................................................................74
Upgrading Linux X9000 clients.................................................................................................75
Contents 5
Upgrading Windows X9000 clients..........................................................................................75
Upgrading firmware on X9720 systems.....................................................................................76
Troubleshooting upgrade issues................................................................................................76
Automatic upgrade............................................................................................................76
Manual upgrade...............................................................................................................77
12 Licensing...............................................................................................78
Viewing license terms..............................................................................................................78
Retrieving a license key...........................................................................................................78
Using AutoPass to retrieve and install permanent license keys........................................................78
13 Upgrading the X9720 Network Storage System hardware............................79
Adding new server blades.......................................................................................................79
Adding capacity blocks...........................................................................................................81
Carton contents.................................................................................................................81
Where to install the capacity blocks.....................................................................................82
Base cabinet additional capacity blocks...........................................................................82
Expansion cabinet additional capacity blocks...................................................................82
Installation procedure.........................................................................................................83
Step 1—Install X9700c in the cabinet..............................................................................83
Step 2—Install X9700cx in the cabinet.............................................................................84
Step 3—Cable the capacity block...................................................................................84
Step 4—Cable the X9700c to SAS switches......................................................................86
Base cabinet............................................................................................................86
Expansion cabinet....................................................................................................87
Step 5—Connect the power cords...................................................................................87
Step 6—Power on the X9700c and X9700cx components..................................................87
Step 7—Discover the capacity block and validate firmware versions....................................88
Removing server blades...........................................................................................................88
Removing capacity blocks........................................................................................................88
14 Upgrading firmware................................................................................89
Firmware update summary.......................................................................................................89
Locating firmware...................................................................................................................89
Upgrading Onboard Administrator...........................................................................................90
Upgrading all Virtual Connect modules.....................................................................................90
Upgrading X9700c controller firmware......................................................................................91
Upgrading X9700cx I/O module and disk drive firmware............................................................92
Upgrading SAS switch module firmware....................................................................................92
15 Troubleshooting......................................................................................94
Managing support tickets........................................................................................................94
Creating, viewing, and deleting support tickets......................................................................94
Support ticket states............................................................................................................95
Updating the ticket database when nodes are added or removed............................................95
Configuring the support ticket feature....................................................................................95
Configuring shared ssh keys................................................................................................95
General troubleshooting steps..................................................................................................96
Escalating issues.....................................................................................................................96
Useful utilities and processes....................................................................................................96
Accessing the Onboard Administrator (OA) through the network..............................................96
Access the OA Web-based administration interface...........................................................96
6 Contents
Accessing the Onboard Administrator (OA) through the serial port...........................................96
Accessing the Onboard Administrator (OA) via service port....................................................97
Using hpacucli – Array Configuration Utility (ACU).................................................................97
The exds_stdiag utility........................................................................................................97
Syntax.........................................................................................................................98
Network testing tools..........................................................................................................98
exds_netdiag................................................................................................................99
Sample output..........................................................................................................99
exds_netperf.................................................................................................................99
POST error messages............................................................................................................100
LUN layout..........................................................................................................................100
X9720 monitoring................................................................................................................100
Identifying failed I/O modules on an X9700cx chassis..............................................................101
Failure indications............................................................................................................101
Identifying the failed component........................................................................................102
Re-seating an X9700c controller........................................................................................105
Viewing software version numbers..........................................................................................105
Troubleshooting specific issues................................................................................................106
Software services.............................................................................................................106
Failover..........................................................................................................................106
Windows X9000 clients...................................................................................................107
X9000 Software reinstall failed..........................................................................................107
Mode 1 or mode 6 bonding.............................................................................................107
X9000 RPC call to host failed............................................................................................108
Degrade server blade/Power PIC.......................................................................................108
ibrix_fs -c failed with "Bad magic number in super-block"......................................................108
LUN status is failed..........................................................................................................109
Apparent failure of HP P700m...........................................................................................109
X9700c enclosure front panel fault ID LED is amber..............................................................110
Spare disk drive not illuminated green when in use..............................................................110
Replacement disk drive LED is not illuminated green.............................................................110
X9700cx GSI LED is amber...............................................................................................110
X9700cx drive LEDs are amber after firmware is flashed.......................................................111
Configuring the Virtual Connect domain..................................................................................111
Synchronizing information on file serving nodes and the configuration database...........................112
16 Replacing components in the X9720 Network Storage System....................113
Customer replaceable components..........................................................................................113
Determining when to replace a component..............................................................................113
Hot-pluggable and non-hot-pluggable components....................................................................114
Returning the defective component..........................................................................................114
Parts-only warranty service.....................................................................................................114
Required tools......................................................................................................................114
Additional documentation......................................................................................................114
Replacing the c7000 blade enclosure and server blade parts.....................................................115
Replacing the blade enclosure...........................................................................................115
Replacing a server blade or system board of a server blade..................................................115
Replacing a server blade disk drive....................................................................................116
Replacing both disk drives.................................................................................................116
Replacing the Onboard Administrator (OA) module..............................................................116
Replacing the Ethernet Virtual Connect (VC) module (bay 1 or bay 2).....................................116
Replacing the SAS switch in Bay 3 or 4..............................................................................117
Replacing the P700m mezzanine card................................................................................118
Replacing capacity block parts...............................................................................................119
Contents 7
Replacing capacity block hard disk drive............................................................................119
Replacing the X9700c controller........................................................................................119
Replacing the X9700c controller battery..............................................................................120
Replacing the X9700c power supply..................................................................................121
Replacing the X9700c fan.................................................................................................121
Replacing the X9700c chassis...........................................................................................121
Replacing the X9700cx I/O module ..................................................................................122
Replacing the X9700cx power supply.................................................................................123
Replacing the X9700cx fan...............................................................................................123
Replacing a SAS cable.....................................................................................................123
17 Recovering the X9720 Network Storage System........................................125
Starting the recovery.............................................................................................................125
Configuring a file serving node..............................................................................................126
Configuring a file serving node using the original template....................................................126
Completing the restore on a file serving node......................................................................129
Configuring a file serving node manually............................................................................131
Configuring the management console on the dedicated (non-agile) Management Server blade.......139
Completing the restore on the dedicated (non-agile) Management Server................................147
Troubleshooting....................................................................................................................147
iLO remote console does not respond to keystrokes...............................................................147
18 Support and other resources...................................................................148
Contacting HP......................................................................................................................148
Related information...............................................................................................................148
HP websites.........................................................................................................................149
Rack stability........................................................................................................................149
Customer self repair..............................................................................................................149
Product warranties................................................................................................................150
Subscription service..............................................................................................................150
A Component and cabling diagrams...........................................................151
Base and expansion cabinets.................................................................................................151
Front view of a base cabinet..............................................................................................151
Back view of a base cabinet with one capacity block...........................................................152
Front view of a full base cabinet.........................................................................................153
Back view of a full base cabinet.........................................................................................154
Front view of an expansion cabinet ...................................................................................155
Back view of an expansion cabinet with four capacity blocks.................................................156
Performance blocks (c-Class Blade enclosure)............................................................................156
Front view of a c-Class Blade enclosure...............................................................................156
Rear view of a c-Class Blade enclosure...............................................................................157
Flex-10 networks...............................................................................................................157
Capacity blocks...................................................................................................................158
X9700c (array controller with 12 disk drives).......................................................................159
Front view of an X9700c..............................................................................................159
Rear view of an X9700c..............................................................................................159
X9700cx (dense JBOD with 70 disk drives)..........................................................................160
Front view of an X9700cx............................................................................................160
Rear view of an X9700cx.............................................................................................160
Cabling diagrams................................................................................................................161
Capacity block cabling—Base and expansion cabinets........................................................161
Virtual Connect Flex-10 Ethernet module cabling—Base cabinet.............................................161
8 Contents
SAS switch cabling—Base cabinet.....................................................................................163
SAS switch cabling—Expansion cabinet..............................................................................164
B Spare parts list ......................................................................................165
AW548A—Base Rack...........................................................................................................165
AW552A—X9700 Expansion Rack.........................................................................................165
AW549A—X9700 Server Chassis..........................................................................................166
AW550A—X9700 Blade Server ............................................................................................166
AW551A—X9700 Capacity Block (X9700c and X9700cx) .......................................................167
C Warnings and precautions......................................................................168
Electrostatic discharge information..........................................................................................168
Grounding methods..............................................................................................................168
Equipment symbols...............................................................................................................168
Weight warning...................................................................................................................169
Rack warnings and precautions..............................................................................................169
Device warnings and precautions...........................................................................................170
D Regulatory compliance and safety............................................................172
Regulatory compliance identification numbers..........................................................................172
Federal Communications Commission notice............................................................................172
Class A equipment...........................................................................................................172
Class B equipment...........................................................................................................172
Declaration of conformity for products marked with the FCC logo, United States only................173
Modifications..................................................................................................................173
Cables...........................................................................................................................173
Laser compliance..................................................................................................................173
International notices and statements........................................................................................174
Canadian notice (Avis Canadien)......................................................................................174
Class A equipment......................................................................................................174
Class B equipment......................................................................................................174
European Union notice.....................................................................................................174
BSMI notice....................................................................................................................174
Japanese notice...............................................................................................................174
Korean notice (A&B).........................................................................................................175
Safety.................................................................................................................................175
Battery Replacement notice...............................................................................................175
Taiwan Battery Recycling Notice...................................................................................175
Power cords....................................................................................................................175
Japanese Power Cord notice..............................................................................................176
Electrostatic discharge......................................................................................................176
Preventing electrostatic discharge..................................................................................176
Grounding methods.....................................................................................................176
Waste Electrical and Electronic Equipment directive...................................................................177
Czechoslovakian notice....................................................................................................177
Danish notice..................................................................................................................177
Dutch notice....................................................................................................................177
English notice..................................................................................................................178
Estonian notice................................................................................................................178
Finnish notice..................................................................................................................178
French notice...................................................................................................................178
German notice................................................................................................................179
Greek notice...................................................................................................................179
Contents 9
Hungarian notice.............................................................................................................179
Italian notice...................................................................................................................179
Latvian notice..................................................................................................................180
Lithuanian notice..............................................................................................................180
Polish notice....................................................................................................................180
Portuguese notice.............................................................................................................180
Slovakian notice..............................................................................................................181
Slovenian notice..............................................................................................................181
Spanish notice.................................................................................................................181
Swedish notice................................................................................................................181
Glossary..................................................................................................182
Index.......................................................................................................184
10 Contents
1 Product description
HP StorageWorks X9720 Network Storage System is a scalable, network-attached storage (NAS) product. The system combines HP X9000 File Serving Software with HP server and storage hardware to create a cluster of file serving nodes.
HP X9720 Network Storage System features
The X9720 Network Storage System provides the following features:
Segmented, scalable file system under a single namespace
NFS, CIFS, FTP, and HTTP support for accessing file system data
Centralized CLI and GUI for cluster management
Policy management
Continuous remote replication
Dual redundant paths to all storage components
Gigabytes-per-second of throughput
IMPORTANT: It is important to keep regular backups of the cluster configuration. See “Backing
up the management console configuration” (page 42) for more information.
System components
The X9720 Network Storage System includes the following components:
X9720 Network Storage System Base Rack, including
Two ProCurve 2810-24G management switches
Keyboard, video, and mouse (KVM)
X9720 Network Storage System performance chassis comprised of:
A c-Class blade enclosure◦ ◦ Two Flex-10 Virtual Connect modules Redundant SAS switch pair
Performance block comprised of a server blade and blade infrastructure
Capacity block (array) (minimum of one) comprised of:
X9700c (array controller chassis and 12 disk drives)◦ ◦ X9700cx (dense JBOD with 70 disk drives)
Software for manageability, segmented file system, and file serving
IMPORTANT: All software that is included with the X9720 Network Storage System is for the
sole purpose of operating the system. Do not add, remove, or change any software unless instructed to do so by HP-authorized personnel.
For more information about system components and cabling, see “Component and cabling
diagrams” (page 151).
HP X9000 Software features
HP X9000 Software is a scale-out, network-attached storage solution composed of a parallel file system for clusters, an integrated volume manager, high-availability features such as automatic
HP X9720 Network Storage System features 11
failover of multiple components, and a centralized management interface. X9000 Software can be deployed in environments scaling to thousands of nodes.
Based on a Segmented File System architecture, X9000 Software enables enterprises to integrate I/O and storage systems into a single clustered environment that can be shared across multiple applications and managed from a single central management console.
X9000 Software is designed to operate with high-performance computing applications that require high I/O bandwidth, high IOPS throughput, and scalable configurations. Examples of these applications include Internet streaming, rich media streaming, data mining, web search, manufacturing, financial modeling, life sciences modeling, and seismic processing.
Some of the key features and benefits are as follows:
Scalable configuration. You can add servers to scale performance and add storage devices
to scale capacity.
Single namespace. All directories and files are contained in the same namespace.
Multiple environments. Operates in both the SAN and DAS environments.
High availability. The high-availability software protects servers.
Tuning capability. The system can be tuned for large or small-block I/O.
Flexible configuration. Segments can be migrated dynamically for rebalancing and data
tiering.
High availability and redundancy
The segmented architecture is the basis for fault resilience—loss of access to one or more segments does not render the entire file system inaccessible. Individual segments can be taken offline temporarily for maintenance operations and then returned to the file system.
To ensure continuous data access, X9000 Software provides manual and automated failover protection at various points:
Server. A failed node is powered down and a designated standby server assumes all of its
segment management duties.
Segment. Ownership of each segment on a failed node is transferred to a designated standby
server.
Network interface. The IP address of a failed network interface is transferred to a standby
network interface until the original network interface is operational again.
Storage connection. For servers with HBA-protected Fibre Channel access, failure of the HBA
triggers failover of the node to a designated standby server.
12 Product description
2 Getting started
This chapter describes how to log into the system, how to boot the system and individual server blades, how to change passwords, and how to back up the management console configuration. It also describes the management interfaces provided with X9000 Software.
IMPORTANT: Do not modify any parameters of the operating system or kernel, or update any
part of the X9720 Network Storage System unless instructed to do so by HP; otherwise, the X9720 Network Storage System could fail to operate properly.
Setting up the X9720 Network Storage System
An HP service specialist sets up the X9720 Network Storage System at your site, including the following tasks:
Installation steps
Remove the product from the shipping cartons that you have placed in the location where the
product will be installed, confirm the contents of each carton against the list of included items and check for any physical damage to the exterior of the product, and connect the product to the power and network provided by you.
Review your server, network, and storage environment relevant to the HP Enterprise NAS
product implementation to validate that prerequisites have been met.
Validate that your file system performance, availability, and manageability requirements have
not changed since the service planning phase. Finalize the HP Enterprise NAS product implementation plan and software configuration.
Implement the documented and agreed-upon configuration based on the information you
provided on the pre-delivery checklist.
Document configuration details.
Additional configuration steps
When your system is up and running, you can perform any additional configuration of your cluster and file systems. The management console GUI and CLI are used to perform most operations. (Some of the features described here might have been configured for you as part of the system installation.)
Cluster. Configure the following as needed:
Virtual interfaces for client access.
Failover for file serving nodes, network interfaces, and HBAs.
Cluster event notification through email or SNMP.
Management console backups.
NDMP backups.
These cluster features are described later in this guide. File systems. Set up the following features as needed:
Additional file systems. Optionally, configure data tiering on the file systems to move files to
specific tiers based on file attributes.
NFS, CIFS, FTP, or HTTP. Configure the methods you will use to access file system data.
Quotas. Configure user, group, and directory tree quotas as needed.
Setting up the X9720 Network Storage System 13
Remote replication. Use this feature to replicate changes in a source file system on one cluster
to a target file system on either the same cluster or a second cluster.
Snapshots. Use this feature to capture a point-in-time copy of a file system.
File allocation. Use this feature to specify the manner in which segments are selected for storing
new files and directories.
For more information about these file system features, see the HP StorageWorks File Serving Software File System User Guide.
Logging in to the X9720 Network Storage System
Using the network
Use ssh to log in remotely from another host. You can log in to any server using any configured site network interface (eth1, eth2, or bond1).
With ssh and the root user, after you log in to any server, your .ssh/known_hosts file will work with any server in an X9720 Network Storage System.
The server blades in your original X9720 are configured to support password-less ssh between them; after you have connected to one, you can reach the others without specifying the root password again. If you wish to have the same support for additional server blades, or wish to access the X9720 itself without specifying a password, add the keys of the other servers to .ssh/ authorized keys on each server blade.
Using the TFT keyboard/monitor
If the site network is down, you can log in to the console as follows:
1. Pull out the keyboard monitor (See “Front view of a base cabinet” (page 151)).
2. Access the on-screen display (OSD) main dialog box by pressing the Print Scrn key or by pressing the Ctrl key twice within one second.
3. Double-click the first server name.
4. Log in as normal.
NOTE: By default, the first port is connected with the dongle to the front of blade 1 (that is, server
1). If server 1 is down, move the dongle to another blade.
Using the serial link on the Onboard Administrator
If you are connected to a terminal server, you can log in through the serial link on the Onboard Administrator.
Booting the system and individual server blades
Before booting the system, ensure that all of the system components other than the server blades—the capacity blocks and so on—are turned on. By default, server blades boot whenever power is applied to the X9720 Network Storage System performance chassis (c-Class Blade enclosure). If all server blades are powered off, you can boot the system as follows:
To boot the system:
1. Press the power button on server blade 1.
2. Log in as root to server 1.
14 Getting started
3. To power on the remaining server blades, run the command:
ibrix_server -P on -h <hostname>
NOTE: Alternatively, press the power button on all of the remaining servers. There is no
need to wait for the first server blade to boot.
Management interfaces
Cluster operations are managed through the X9000 Software management console, which provides both a GUI and a CLI. Most operations can be performed from either the GUI or the CLI. However, the following operations can be performed only from the CLI:
SNMP configuration (ibrix_snmpagent, ibrix_snmpgroup, ibrix_snmptrap,
ibrix_snmpuser, ibrix_snmpview)
Health checks (ibrix_haconfig, ibrix_health, ibrix_healthconfig)
Raw storage management (ibrix_pv, ibrix_vg, ibrix_lv)
Management console operations (ibrix_fm) and management console tuning
(ibrix_fm_tune)
File system checks (ibrix_fsck)
Kernel profiling (ibrix_profile)
NFS autoconnection (ibrix_autoconnect)
Cluster configuration (ibrix_clusterconfig)
Configuration database consistency (ibrix_dbck)
Shell task management (ibrix_shell)
Using the GUI
The GUI is a browser-based interface to the management console. See the release notes for the supported browsers and other software required to view charts on the dashboard.
If you are using HTTP to access the GUI, navigate to the following location, specifying port 80:
http://<management_console_IP>:80/fusion
If you are using HTTPS to access the GUI, navigate to the following location, specifying port 443:
https://<management_console_IP>:443/fusion
In these URLs, <management_console_IP> is the IP address of the management console user VIF.
The GUI prompts for your user name and password. The default administrative user is ibrix. Enter the password that was assigned to this user when the system was installed. (You can change the password using the Linux passwd command.) To allow other users to access the GUI, see
“Adding user accounts for GUI access” (page 19).
Management interfaces 15
The GUI dashboard opens in the same browser window. You can open multiple GUI windows as necessary. See the online help for information about all GUI displays and operations.
The GUI dashboard enables you to monitor the entire cluster. There are three parts to the dashboard: System Status, Cluster Overview, and the Navigator.
16 Getting started
System Status
The System Status section lists the number of cluster events that have occurred in the last 24 hours. There are three types of events:
Alerts. Disruptive events that can result in loss of access to file system data. Examples are a segment that is unavailable or a server that cannot be accessed.
Warnings. Potentially disruptive conditions where file system access is not lost, but if the situation is not addressed, it can escalate to an alert condition. Examples are a very high server CPU utilization level or a quota limit close to the maximum.
Information. Normal events that change the cluster. Examples are mounting a file system or creating a segment.
Cluster Overview
The Cluster Overview provides the following information:
Capacity
The amount of cluster storage space that is currently free or in use.
Filesystems
The current health status of the file systems in the cluster. The overview reports the number of file systems in each state (healthy, experiencing a warning, experiencing an alert, or unknown).
Segment Servers
The current health status of the file serving nodes in the cluster. The overview reports the number of nodes in each state (healthy, experiencing a warning, experiencing an alert, or unknown).
Services
Whether the specified file system services are currently running:
One or more tasks are running.
No tasks are running.
Statistics
Historical performance graphs for the following items:
Network I/O (MB/s)
Disk I/O (MB/s)
CPU usage (%)
Memory usage (%)
On each graph, the X-axis represents time and the Y-axis represents performance. Use the Statistics menu to select the servers to monitor (up to two), to change the maximum
value for the Y-axis, and to show or hide resource usage distribution for CPU and memory.
Recent Events
The most recent cluster events. Use the Recent Events menu to select the type of events to display.
You can also access certain menu items directly from the Cluster Overview. Mouse over the Capacity, Filesystems or Segment Server indicators to see the available options.
Management interfaces 17
Navigator
The Navigator appears on the left side of the window and displays the cluster hierarchy. You can use the Navigator to drill down in the cluster configuration to add, view, or change cluster objects such as file systems or storage, and to initiate or view tasks such as snapshots or replication. When you select an object, a details page shows a summary for that object. The lower Navigator allows you to view details for the selected object, or to initiate a task. In the following example, we selected Cluster Configuration in the Navigator, and the Summary shows configuration information. In the lower Navigator, we selected NDMP Backup > Active Sessions to see details about the sessions.
NOTE: When you perform an operation on the GUI, a spinning finger is displayed until the
operation is complete. However, if you use Windows Remote Desktop to access the management console, the spinning finger is not displayed.
Customizing the GUI
For most tables in the GUI, you can specify the columns that you want to display and the sort order of each column. When this feature is available, mousing over a column causes the label to change color and a pointer to appear. Click the pointer to see the available options. In the following example, you can sort the contents of the Mountpoint column in ascending or descending order, and you can select the columns that you want to appear in the display.
18 Getting started
Adding user accounts for GUI access
X9000 Software supports administrative and user roles. When users log in under the administrative role, they can configure the cluster and initiate operations such as remote replication or snapshots. When users log in under the user role, they can view the cluster configuration and status, but cannot make configuration changes or initiate operations. The default administrative user name is ibrix. The default regular username is ibrixuser.
Usernames for the administrative and user roles are defined in the /etc/group file. Administrative users are specified in the ibrix-admin group, and regular users are specified in the ibrix-user group. These groups are created when X9000 Software is installed. The following entries in the
/etc/group file show the default users in these groups:
ibrix-admin:x:501:root,ibrix ibrix-user:x:502:ibrix,ibrixUser,ibrixuser
You can add other users to these groups as needed, using Linux procedures.
Using the CLI
The administrative commands described in this guide must be executed on the management console host and require root privileges. The commands are located in $IBRIXHOMEbin. For complete information about the commands, see the HP StorageWorks X9000 File Serving Software CLI Reference Guide.
When using ssh to access the machine hosting the management console, specify the IP address of the management console user VIF.
Starting the array management software
Depending on the array type, you can launch the array management software from the management console GUI. In the Navigator, select Vendor Storage, select your array from the Vendor Storage page, and click Launch Storage Management.
X9000 client interfaces
X9000 clients can access the management console as follows:
Linux clients. Linux client commands can be used for tasks such as mounting or unmounting
file systems and displaying statistics. See the HP StorageWorks X9000 File Serving Software CLI Reference Guide for details about these commands.
Windows clients. The Windows client GUI can be used for tasks such as mounting or
unmounting file systems and registering Windows clients.
Using the Windows X9000 client GUI
The Windows X9000 client GUI is the client interface to the management console. To open the GUI, double-click the desktop icon or select the IBRIX Client program from the Start menu on the client. The client program contains tabs organized by function.
Management interfaces 19
NOTE: The Windows X9000 client application can be started only by users with Administrative
privileges.
Status. Shows the client’s management console registration status and mounted file systems,
and provides access to the IAD log for troubleshooting.
Registration. Registers the client with the management console, as described in the HP
StorageWorks File Serving Software Installation Guide.
Mount. Mounts a file system. Select the Cluster Name from the list (the cluster name is the
management console name), enter the name of the file system to mount, select a drive, and then click Mount. (If you are using Remote Desktop to access the client and the drive letter does not appear, log out and log back in.)
Umount. Unmounts a file system.
Tune Host. Tunable parameters include the NIC to prefer (the client uses the cluster interface
by default unless a different network interface is preferred for it), the communications protocol (UDP or TCP), and the number of server threads to use.
Active Directory Settings. Displays current Active Directory settings.
Online help is also available for the client GUI.
X9000 Software manpages
X9000 Software provides manpages for most of its commands. To view the manpages, set the MANPATH variable on the management console to include the path to the manpages and then export it. The manpages are in the $IBRIXHOME/man directory. For example, if $IBRIXHOME is /usr/local/ibrix (the default), you would set the MANPATH variable as follows on the management console and then export the variable.
MANPATH=$MANPATH:/usr/local/ibrix/man
Changing passwords
You may want to change the passwords on your system:
Hardware passwords. See the documentation for the specific hardware for more information.
Root password. Use the passwd(8) command on each server in turn.
X9000 Software user password. This password is created during installation and is used to
log on to the management console GUI. The default is ibrix. You can change the password on the management console using the Linux passwd command. You will be prompted to enter the new password.
# passwd ibrix
Configuring ports for a firewall
IMPORTANT: To avoid unintended consequences, HP recommends that you configure the firewall
during scheduled maintenance times.
When configuring a firewall, you should be aware of the following:
SELinux should be disabled.
By default, NFS uses random port numbers for operations such as mounting and locking.
These ports must be fixed so that they can be listed as exceptions in a firewall configuration
20 Getting started
file. For example, you will need to lock specific ports for rpc.statd, rpc.lockd, rpc.mountd, and rpc.quotad.
It is best to allow all ICMP types on all networks; however, you can limit ICMP to types 0, 3,
8, and 11 if necessary.
Be sure to open the ports listed in the following table.
DescriptionPort
SSH22/tcp
SSH for Onboard Administrator (OA); only for X9720 blades9022/tcp
NTP123/tcp, 123/upd
Multicast DNS, 224.0.0.2515353/udp
netperf tool12865/tcp
X9000 management console to file serving nodes80/tcp
443/tcp
X9000 management console and X9000 file system5432/tcp 8008/tcp 9002/tcp 9005/tcp 9008/tcp 9009/tcp 9200/tcp
Between file serving nodes and NFS clients (user network)
NFS
RPC
quota
lockmanager
lockmanager
mount daemon
stat
stat outgoing
reserved for use by a custom application (CMU) and can be disabled if not used
2049/tcp, 2049/udp 111/tcp, 111/udp 875/tcp, 875/udp 32803/tcp 32769/udp 892/tcp, 892/udp 662/tcp, 662/udp 2020/tcp, 2020/udp 4000:4003/tcp
Between file serving nodes and CIFS clients (user network)137/udp 138/udp 139/tcp 445/tcp
Between file serving nodes and X9000 clients (user network)9000:9002/tcp 9000:9200/udp
Between file serving nodes and FTP clients (user network)20/tcp, 20/udp 21/tcp, 21/udp
Between X9000 management console GUI and clients that need to access the GUI7777/tcp 8080/tcp
Dataprotector5555/tcp, 5555/udp
Internet Printing Protocol (IPP)631/tcp, 631/udp
Configuring ports for a firewall 21
HP Insight Remote Support software
HP Insight Remote Support supplements your monitoring, 24x7 to ensure maximum system availability by providing intelligent event diagnosis, and automatic, secure submission of hardware event notifications to HP, which will initiate a fast and accurate resolution, based on your product’s service level. Notifications may be sent to your authorized HP Channel Partner for on-site service, if configured and available in your country. The software is available in two variants:
HP Insight Remote Support Standard: This software supports server and storage devices and
is optimized for environments with 1-50 servers. Ideal for customers who can benefit from proactive notification, but do not need proactive service delivery and integration with a management platform.
HP Insight Remote Support Advanced: This software provides comprehensive remote monitoring
and proactive service support for nearly all HP servers, storage, network, and SAN environments, plus selected non-HP servers that have a support obligation with HP. It is integrated with HP Systems Insight Manager. A dedicated server is recommended to host both HP Systems Insight Manager and HP Insight Remote Support Advanced.
Details for both versions are available at:
http://www.hp.com/go/insightremotesupport
The required components for HP Insight Remote Support are preinstalled on the file serving nodes. You will need to install the Central Management Server (CMS) on a separate Windows system. See the X9000 Series release notes for more information.
22 Getting started
3 Configuring virtual interfaces for client access
X9000 Software uses a cluster network interface to carry management console traffic and traffic between file serving nodes. This network is configured as bond0 when the cluster is installed. For clusters with an agile management console configuration, a virtual interface is also created for the cluster network interface to provide failover support for the console.
Although the cluster network interface can carry traffic between file serving nodes and clients, HP recommends that you configure one or more user network interfaces for this purpose. Typically, bond1 is created for the first user network when the cluster is configured.
To provide high availability for a user network, you should configure a bonded virtual interface (VIF) for the network and then set up failover for the VIF. This method prevents interruptions to client traffic. If necessary, the file serving node hosting the VIF can fail over to its standby backup node, and clients can continue to access the file system through the backup node.
Network and VIF guidelines
To provide high availability, the user interfaces used for client access should be configured as bonded virtual interfaces (VIFs). Note the following:
Nodes needing to communicate for file system coverage or for failover must be on the same
network interface. Also, nodes set up as a failover pair must be connected to the same network interface.
Use a Gigabit Ethernet port (or faster) for user networks.
NFS, CIFS, FTP, and HTTP clients can use the same user VIF. The servers providing the VIF
should be configured in backup pairs, and the NICs on those servers should also be configured for failover.
For X9000 Linux and Windows clients, the servers hosting the VIF should be configured in
backup pairs. However, X9000 clients do not support backup NICs. Instead, X9000 clients should connect to the parent bond of the user VIF or to a different VIF.
Creating a bonded VIF
Use the following procedure to create a bonded VIF (bond1:1 in this example):
1. If high availability (automated failover) is configured on the servers, disable it. Run the following command on the management console:
# ibrix_server –m -U
2. Identify the bond1:1 VIF:
# ibrix_nic –a -n bond1:1 –h node1,node2,node3,node4
3. Assign an IP address to the bond1:1 VIFs on each node. In the command, -I specifies the IP address, -M specifies the netmask, and -B specifies the broadcast address:
# ibrix_nic –c –n bond1:1 –h node1 –I 16.123.200.201 –M 255.255.255.0 -B 16.123.200.255 # ibrix_nic –c –n bond1:1 –h node2 –I 16.123.200.202 –M 255.255.255.0 -B 16.123.200.255 # ibrix_nic –c –n bond1:1 –h node3 –I 16.123.200.203 –M 255.255.255.0 -B 16.123.200.255 # ibrix_nic –c –n bond1:1 –h node4 –I 16.123.200.204 –M 255.255.255.0 -B 16.123.200.255
Configuring standby backup nodes
Assign standby backup nodes for the bond1:1 interface. The backup nodes should be configured in pairs. For example, node1 is the backup for node2, and node2 is the backup for node1.
Network and VIF guidelines 23
1. Identify the VIF:
# ibrix_nic –a -n bond1:2 –h node1,node2,node3,node4
2. Set up a standby server for each VIF:
# ibric_nic –b –H node1/bond1:1,node2/bond1:2 # ibric_nic –b –H node2/bond1:1,node1/bond1:2 # ibric_nic –b –H node3/bond1:1,node4/bond1:2 # ibric_nic –b –H node4/bond1:1,node3/bond1:2
Configuring NIC failover
NIC monitoring should be configured on VIFs that will be used by NFS, CIFS, FTP, or HTTP. Use the same backup pairs that you used when configuring standby servers. For example:
# ibric_nic –m -h node1 -A node2/bond1:1 # ibric_nic –m -h node2 -A node1/bond1:1 # ibric_nic –m -h node3 -A node4/bond1:1 # ibric_nic –m -h node4 -A node3/bond1:1
Configuring automated failover
To enable automated failover for your file serving nodes, execute the following command:
ibrix_server m [-h SERVERNAME]
Example configuration
This example uses two nodes, ib50-81 and ib50-82. These nodes are backups for each other, forming a backup pair.
[root@ib50-80 ~]# ibrix_server -l Segment Servers =============== SERVER_NAME BACKUP STATE HA ID GROUP
----------- ------- ------------ --- ------------------------------------ ----­ib50-81 ib50-82 Up on 132cf61a-d25b-40f8-890e-e97363ae0d0b servers ib50-82 ib50-81 Up on 7d258451-4455-484d-bf80-75c94d17121d servers
All VIFs on ib50-81 have backup (standby) VIFs on ib50-82. Similarly, all VIFs on ib50-82 have backup (standby) VIFs on ib50-81. NFS, CIFS, FTP, and HTTP clients can connect to bond1:1 on either host. If necessary, the selected server will fail over to bond1:2 on the opposite host. X9000 clients could connect to bond1 on either host, as these clients do not support or require NIC failover. (The following sample output shows only the relevant fields.)
[root@ib50-80 ~]# ibrix_nic -l HOST IFNAME TYPE STATE IP_ADDRESS BACKUP_HOST BACKUP_IF
------- ------ ------- ------------------- ------------- ----------- --------­ib50-81 bond1:1 User Up, LinkUp 16.226.50.220 ib50-82 bond1:1 ib50-81 bond0 Cluster Up, LinkUp 172.16.0.81 ib50-81 bond1:2 User Inactive, Standby ib50-81 bond1 User Up, LinkUp 16.226.50.81 ib50-82 bond0 Cluster Up, LinkUp 172.16.0.82 ib50-82 bond1 User Up, LinkUp 16.226.50.82 ib50-82 bond1:2 User Inactive, Standby ib50-82 bond1:1 User Up, LinkUp 16.226.50.228 ib50-81 bond1:1
Specifying VIFs in the client configuration
When you configure your clients, you may need to specify the VIF that should be used for client access.
NFS/CIFS. Specify the VIF IP address of the servers (for example, bond1:0) to establish connection. You can also configure DNS round robin to ensure NFS or CIFS client-to-server distribution. In both cases, the NFS/CIFS clients will cache the initial IP they used to connect to the respective share, usually until the next reboot.
24 Configuring virtual interfaces for client access
FTP. When you add an FTP share on the Add FTP Shares dialog box or with the ibrix_ftpshare command, specify the VIF as the IP address that clients should use to access the share.
HTTP. When you create a virtual host on the Create Vhost dialog box or with the ibrix_httpvhost command, specify the VIF as the IP address that clients should use to access shares associated with the Vhost.
X9000 clients. Use the following command to prefer the appropriate user network. Execute the command once for each destination host that the client should contact using the specified interface.
ibrix_client -n -h SRCHOST -A DESTNOST/IFNAME
For example:
ibrix_client -n -h client12.mycompany.com -A ib50-81.mycompany.com/bond1
NOTE: Because the backup NIC cannot be used as a preferred network interface for X9000
clients, add one or more user network interfaces to ensure that HA and client communication work together.
Support for link state monitoring
Do not configure link state monitoring for user network interfaces or VIFs that will be used for CIFS or NFS. Link state monitoring is supported only for use with iSCSI storage network interfaces, such as those provided with X9300 Gateway systems.
Support for link state monitoring 25
4 Configuring failover
This chapter describes how to configure failover for agile management consoles, file serving nodes, network interfaces, and HBAs.
Agile management consoles
The management console maintains the cluster configuration and provides graphical and command-line user interfaces for managing and monitoring the cluster. Typically, one active management console and one passive management console are installed when the cluster is installed. This is called an agile management console configuration.
NOTE: Optionally, the management console can be installed on a dedicated Management
Server. This section describes the agile management console configuration.
Agile management console modes
An agile management console can be in one of the following modes:
active. In this mode, the management console controls console operations. All cluster
administration and configuration commands must be run from the active management console.
passive. In this mode, the management console monitors the health of the active management
console. If the active management console fails, the passive management console becomes the active console.
maintenance. In this mode, the management console does not participate in console operations.
Maintenance mode should be used for operations such as manual failover of the active management console, X9000 Software upgrades, and blade replacements.
Agile management consoles and failover
Using an agile management console configuration provides high availability for management console services. If the active management console fails, the cluster virtual interface will go down. When the passive management console detects that the cluster virtual interface is down, it will become the active console. This management console rebuilds the cluster virtual interface, starts management console services locally, transitions into active mode, and take over management console operation.
Failover of the active management console affects the following features:
User networks. The virtual interface used by clients will also fail over. Users may notice a brief
reconnect while the newly active management console takes over management of the virtual interface.
Support tickets. The existing support ticket information is not moved to the newly active
management console. Support Ticket operations are always handled by the active management console and the final output of the operations is stored there.
Management console GUI. You will need to reconnect to the management console VIF after
the failover.
Failing over the management console manually
To fail over the active management console manually, place the console into maintenance mode. Enter the following command on the node hosting the console:
ibrix_fm -m maintenance
The command takes effect immediately.
26 Configuring failover
The failed-over management console remains in maintenance mode until it is moved to passive mode using the following command:
ibrix_fm -m passive
A management console cannot be moved from maintenance mode to active mode.
Viewing information about management consoles
To view mode information, use the following command:
ibrix_fm –i
NOTE: If the management console was not installed in an agile configuration, the output will
report FusionServer: fusion manager name not set! (active, quorum is not configured).
When a management console is installed, it is registered in the management console configuration. To view a list of all registered management consoles, use the following command:
ibrix_fm –f
Cluster high availability
X9000 Software High Availability keeps your data accessible at all times. Failover protection can be configured for file serving nodes, network interfaces, individual segments, and HBAs. Through physical and logical configuration policies, you can set up a flexible and scalable high availability solution. X9000 clients experience no changes in service and are unaware of the failover events.
Failover modes
High Availability has two failover modes: the default manual failover and the optional automated failover. A manual failover uses the ibrix_server command or the management console GUI to fail over a file serving node to its standby. The server can be powered down or remain up during the procedure. Manual failover also includes failover of any network interfaces having defined standbys. You can perform a manual failover at any time, regardless of whether automated failover is in effect.
Automated failover allows the management console to initiate failover when it detects that standby-protected components have failed. A basic automated failover setup protects all file serving nodes. A comprehensive setup also includes network interface monitoring to protect user network interfaces and HBA monitoring to protect access from file serving nodes to storage via an HBA.
When automated failover is enabled, the management console listens for heartbeat messages that the file serving nodes broadcast at one-minute intervals. The management console automatically initiates failover when it fails to receive five consecutive heartbeats or, if HBA monitoring is enabled, when a heartbeat message indicates that a monitored HBA or pair of HBAs has failed.
If network interface monitoring is enabled, automated failover occurs when the management console receives a heartbeat message indicating that a monitored network might be down and then the console cannot reach that interface.
If a file serving node fails over, you will need to manually fail back the node.
What happens during a failover
The following events occur during automated or manual failover of a file serving node to its standby:
1. The management console verifies that the standby is powered on and accessible.
2. The management console migrates ownership of the node’s segments to the standby and
notifies all file serving nodes and X9000 clients about the migration. This is a persistent change.
3. If network interface monitoring has been set up, the management console activates the standby
user network interface and transfers the IP address of the node’s user network interface to it.
Cluster high availability 27
To determine the progress of a failover, view the Status tab on the GUI or execute the ibrix_server -l command. While the management console is migrating segment ownership, the operational status of the node is Up-InFailover or Down-InFailover, depending on whether the node was powered up or down when failover was initiated. When failover is complete, the operational status changes to Up-FailedOver or Down-FailedOver. For more information about operational states, see “Monitoring the status of file serving nodes” (page 49).
Both automated and manual failovers trigger an event that is reported on the GUI.
Setting up automated failover
The recommended minimum setup for automated failover protection is as follows:
1. Identify standbys for file serving nodes or specific segments. You must implement either
server-level or segment-level standby protection; you cannot implement both.
2. Identify power sources for file serving nodes. For APC power sources, associate file serving
nodes to power source slots.
3. Turn on automated failover.
If your cluster includes one or more user network interfaces carrying NFS/CIFS client traffic, HP recommends that you identify standby network interfaces and set up network interface monitoring.
If your file serving nodes are connected to storage via HBAs, HP recommends that you set up HBA monitoring.
Identifying standbys for file serving nodes
file serving nodes can be configured to provide standby service for one another in the following configurations:
1 x 1. Set up standby pairs, where each server in a pair is the standby for the other.
1 x N. Assign the same standby to a certain number of primaries.
Contact HP Support for recommendations based on your environment. The following restrictions apply to all types of standby configurations:
The management console must have access to both the primary server and its standby.
The same file system must be mounted on both the primary server and its standby.
A server identified as a standby must be able to see all segments that might fail over to it.
In a SAN environment, a primary server and its standby must use the same storage infrastructure
to access a segment’s physical volumes (for example, a multiported RAID array).
To identify a standby for a file serving node, use the following command:
<installdirectory>/bin/ibrix_server -b -h HOSTNAME1,HOSTNAME2
For example, to identify node s2.hp.com as the standby for all segments on node s1.hp.com:
<installdirectory>/bin/ibrix_server -b -h s1.hp.com,s2.hp.com
For performance reasons, you might want to fail over specific segments to a standby instead of failing over all segments on a node to a standby. Use this command to identify the segments:
<installdirectory>/bin/ibrix_fs -b -f FSNAME -s LVLIST -h HOSTNAME
For example, to identify node s1.hp.com as the standby for segments ilv_1, ilv_2, and ilv_3 in file system ifs1:
<installdirectory>/bin/ibrix_fs -b -f ifs1 -s ilv_1,ilv_2,ilv_3 -h s1.hp.com
Identifying power sources
To implement automated failover, perform a forced manual failover, or remotely power a file serving node up or down, you must set up programmable power sources for the nodes and their standbys. Using programmable power sources prevents a “split-brain scenario” between a failing
28 Configuring failover
file serving node and its standby, allowing the failing server to be centrally powered down by the management console in the case of automated failover, and manually in the case of a forced manual failover.
X9000 Software works with iLO, IPMI, OpenIPMI, and OpenIPMI2 integrated power sources and with APC power sources.
Preliminary configuration
Certain configuration steps are required when setting up power sources:
All types. If you plan to implement automated failover, ensure that the management console
has LAN access to the power sources.
Integrated power sources. Install the environment and any drivers and utilities, as specified
by the vendor documentation. If you plan to protect access to the power sources, set up the UID and password to be used.
APC. Enable SNMP access. Set the Community Name to ibrix and the Access Type to
write+. If write+ does not work with your configuration, set the Access Type to write.
Identifying power sources
All power sources must be identified to the configuration database before they can be used. Integrated power sources. To identify an integrated power source, use the following command:
<installdirectory>/bin/ibrix_powersrc -a -t {ipmi|openipmi|openipmi2|ilo}
-h HOSTNAME -I IPADDR -u USERNAME -p PASSWORD
For example, to identify an iLO power source at IP address 192.168.3.170 for node ss01:
<installdirectory>/bin/ibrix_powersrc -a -t ilo -h ss01 -I 192.168.3.170
-u Administrator -p password
APC power source. To identify an APC power source, use the following command:
<installdirectory>/bin/ibrix_powersrc -a -t {apc|apc_msp} -h POWERSRCNAME -n NUMSLOTS
-I IPADDR
For example, to identify an eight-port APC power source named ps1 at IP address 192.168.3.150:
<installdirectory>/bin/ibrix_powersrc -a -t apc -h ps1 -n 8 -I 192.168.3.150
For APC power sources, you must also associate file serving nodes to power source slots. (This step is unnecessary for integrated power sources because the nodes are connected by default to slot 1.) Use the following command:
<installdirectory>/bin/ibrix_hostpower -a -i SLOTID -s POWERSOURCE -h HOSTNAME
For example, to identify that node s1.hp.com is connected to slot 1 on APC power source ps1:
<installdirectory>/bin/ibrix_hostpower -a -i 1 -s ps1 -h s1.hp.com
Updating the configuration database with power source changes
If you move a file serving node to a different power source slot, unplug it from a power source slot, or change its IP address or password, you must update the configuration database with the changes. To do this, use the following command. The user name and password options are needed only for remotely managed power sources. Include the -s option to have the management console skip BMC.
<installdirectory>/bin/ibrix_powersrc -m [-I IPADDR] [-u USERNAME] [-p PASSWORD] [-s] -h POWERSRCLIST
The following command changes the IP address for power source ps1:
<installdirectory>/bin/ibrix_powersrc -m -I 192.168.3.153 -h ps1
To change the APC slot association for a file serving node, use the following command:
<installdirectory>/bin/ibrix_hostpower -m -i FROM_SLOT_ID,TO_SLOT_ID -s POWERSOURCE
-h HOSTNAME
Cluster high availability 29
For example, to identify that node s1.hp.com has been moved from slot 3 to slot 4 on APC power source ps1:
<installdirectory>/bin/ibrix_hostpower -m -i 3,4 -s ps1 -h s1.hp.com
Dissociating a file serving node from a power source
You can dissociate a file serving node from an integrated power source by dissociating it from slot 1 (its default association) on the power source. Use the following command:
<installdirectory>/bin/ibrix_hostpower -d -s POWERSOURCE -h HOSTNAME
To dissociate a file serving node from an APC power source on the specified slot, use the following command. To dissociate the node from all slots on the power source, omit the -i option.
<installdirectory>/bin/ibrix_hostpower -d [-s POWERSOURCE [-i SLOT]] -h HOSTNAME
For example, to dissociate file serving node s1.hp.com from slot 3 on APC power source ps1:
<installdirectory>/bin/ibrix_hostpower -d -s ps1 -i 3 -h s1.hp.com
Deleting power sources from the configuration database
To conserve storage, delete power sources that are no longer in use from the configuration database. If you are deleting multiple power sources, use commas to separate them.
<installdirectory>/bin/ibrix_powersrc -d -h POWERSRCLIST
Turning automated failover on and off
Automated failover is turned off by default. When automated failover is turned on, the management console starts monitoring heartbeat messages from file serving nodes. You can turn automated failover on and off for all file serving nodes or for selected nodes.
To turn on automated failover, use the following command:
<installdirectory>/bin/ibrix_server -m [-h SERVERNAME]
To turn off automated failover, include the -U option:
<installdirectory>/bin/ibrix_server -m -U [-h SERVERNAME]
To turn automated failover on or off for a single file serving node, include the -h SERVERNAME option.
Manually failing over a file serving node
To set up a cluster for manual failover, first identify server-level or segment-level standbys for each file serving node, as described in “Identifying standbys for file serving nodes” (page 28).
Manual failover does not require the use of programmable power supplies. However, if you have installed and identified power supplies for file serving nodes, you can power down a server before manually failing it over. You can fail over a file serving node manually, even when automated failover is turned on.
A file serving node can be failed over from the GUI or the CLI. On the CLI, complete the following steps:
1. Run ibrix_server -f, specifying the node to be failed over in the HOSTNAME option. If appropriate, include the -p option to power down the node before segments are migrated:
<installdirectory>/bin/ibrix_server -f [-p] -h HOSTNAME
2. Determine whether the failover was successful:
<installdirectory>/bin/ibrix_server -l
The contents of the STATE field indicate the status of the failover. If the field persistently shows Down-InFailover or Up-InFailover, the failover did not complete; contact HP Support for assistance. For information about the values that can appear in the STATE field, see “What happens during
a failover” (page 27).
30 Configuring failover
Loading...
+ 157 hidden pages