Copyright 2012, 2017 Hewlett Packard Enterprise Development LP
Notices
The information contained herein is subject to change without notice. The only warranties for Hewlett
Packard Enterprise products and services are set forth in the express warranty statements accompanying
such products and services. Nothing herein should be construed as constituting an additional warranty.
Hewlett Packard Enterprise shall not be liable for technical or editorial errors or omissions contained
herein.
Confidential computer software. Valid license from Hewlett Packard Enterprise required for possession,
use, or copying. Consistent with FAR 12.211 and 12.212, Commercial Computer Software, Computer
Software Documentation, and Technical Data for Commercial Items are licensed to the U.S. Government
under vendor's standard commercial license.
Links to third-party websites take you outside the Hewlett Packard Enterprise website. Hewlett Packard
Enterprise has no control over and is not responsible for information outside the Hewlett Packard
Enterprise website.
Acknowledgments
Intel®, Itanium®, Pentium®, Intel Inside®, and the Intel Inside logo are trademarks of Intel Corporation in
the United States and other countries.
Microsoft® and Windows® are either registered trademarks or trademarks of Microsoft Corporation in the
United States and/or other countries.
UNIX® is a registered trademark of The Open Group.
Revision history
Changes in this edition:
Added new part numbers for Adding SAS disk drive LEDs : HBA mode table under Troubleshooting"
chapter
The HPE Integrity BL860c i4 Server Blade is a dense, low-cost, Intel Itanium processor server blade.
Using a Blade Link hardware assembly, multiple BL860c i4 Server Blades can be conjoined to create
dual-blade, four socket and quad-blade, eight socket variants.
NameNumber of Conjoined Server
BL860c i412
BL870c i424
BL890c i448
The three blade configurations support the HP-UX operating system and are designed for deployment in
HPE c-Class enclosures, specifically the 10U c7000 and the 6U c3000 Enclosures. The nPartition
configuration feature is enabled from the factory, enabling the BL870c i4 and BL890c i4 to be partitioned
into one more electrically isolated partitions. See the nPartitions Administrator Users Guide for more
information.
NOTE:
For purposes of this guide, make sure that the c-Class server blade enclosure is powered on and
running properly and that the OA and iLO 3 is operational.
Server blade overview
ProductCPU cores
(octo)
Number of Processor Sockets
Blades
DIMM slotsmax memoryPCIe I/O
Mezzanine
card capacity
SAS Hard Disk
Drives
BL860c i41624364GB with
BL870c i43248768GB with
BL890c i464961.5TB with
8 Overview
32
16GB DIMMs
64
16GB DIMMs
128
16GB DIMMs
Page 9
Server blade components
1CPU08ICH mezzanine connector
2CPU0 power connector9Mezzanine connector 3
(type 1 or 2)
3Mezzanine connector 1 (type1)10 CPU1 power connector
4Mezzanine connector 2 (type
1 or 2)
5System board thumbscrew12 SAS backplane
6System board thumbscrew13 Pull tab
7Battery (CR2032)
11 CPU1
NOTE:
The iLO 3 password is
located on the pull
tab.
Server blade components9
Page 10
Site preparation
The BL860c i4 does not have cooling or power systems. Cooling and power is provided by the c-Class
enclosure.
IMPORTANT:
To avoid hardware damage, allow the thermal mass of the product to equalize to the temperature
and humidity of the installation facility after removing the shipping materials. A minimum of one hour
per 10°C (50°F) of temperature difference between the shipping facility and installation facility is
required.
Server blade dimensions and weight
Table 1: Server blade dimensions and weight for the BL860c i4
Dimensionsvalue
Height36.63 cm (14.42 in.)
Width5.14 cm (2.025 in.)
Depth48.51 cm (19.1 in.)
WeightUnloaded: 8.6 kg (19 lb)
Table 2: Blade Link dimensions and weight
Blade Link typeHeightWidthWeight
BL1 (BL860 i4)44 mm (1.73 in)51 mm (2 in).5 lb (.22 kg)
BL2 (BL870 i4)44 mm (1.73 in)106 mm (4.17 in)1 lb (.45 kg)
BL4 (BL890 i4)44 mm (1.73 in)212 mm (8.34 in)2 lb (.90 kg)
Enclosure information
All three blade configurations are supported in HPE c7000 and c3000 Enclosures.
This information is for both c3000 and c7000 Enclosures.
SpecificationValue
Temperature range
Operating10°C to 35°C (50°F to 95°F)
Non-operating-30°C to 60°C (-22°F to 140°F)
Wet bulb temperature
Operating28ºC (82.4ºF)
Non-operating38.7ºC (101.7ºF)
Relative humidity (noncondensing)
Operating20% to 80%
Non-operating5% to 95%
1
Storage maximum humidity of 95% is based on a maximum temperature of 45°C (113°F). Altitude maximum for
storage corresponds to a pressure minimum of 70 KPa.
For more information on the c-Class enclosures see:
For more site preparation information, go to http://www.hpe.com/info/Blades-docs, select HPE
Integrity BL860c i4 Server Blade in the list of servers, and then select the Generalized Site
Preparation Guidelines.
1
Sample Site Inspection Checklist
Table 3: Customer and Hewlett Packard Enterprise Information
Customer Information
Name:Phone number:
Street address:City or Town:
State or province:Country
Zip or postal code:
Primary customer contact:Phone number:
Enclosure environmental specifications11
Table Continued
Page 12
Customer Information
Secondary customer contact:Phone number:
Traffic coordinator:Phone number:
Hewlett Packard Enterprise information
Sales representativeOrder number:
Representative making surveyDate:
Scheduled delivery date
Table 4: Site Inspection Checklist
Check either Yes or No. If No, include comment number or date.Comment or
Date
Computer Room
NumberArea or conditionYesNo
1.Is there a completed floor plan?
2.Is adequate space available for maintenance needs?
Front 36 inches (91.4 cm) minimum and rear 36 inches
(91.4 cm) minimum are recommended clearances.
3.Is access to the site or computer room restricted?
4.Is the computer room structurally complete? Expected
date of completion?
5.Is a raised floor installed and in good condition?
6.Is the raised floor adequate for equipment loading?
7.Are channels or cutouts available for cable routing?
8.Is a network line available?
9.Is a telephone line available?
10.Are customer-supplied peripheral cables and LAN cables
available and of the proper type?
11.Are floor tiles in good condition and properly braced?
12.Is floor tile underside shiny or painted? If painted, judge
12 Site preparation
the need for particulate test.
Table Continued
Page 13
Check either Yes or No. If No, include comment number or date.Comment or
Date
Power and Lighting
NumberArea or ConditionYesNo
13.Are lighting levels adequate for maintenance?
14.Are AC outlets available for servicing needs (for example,
laptop)?
15.Does the input voltage correspond to equipment
specifications?
15a.Is dual source power used? If so, identify types and
evaluate grounding.
16.Does the input frequency correspond to equipment
specifications?
17.Are lightning arrestors installed inside the building?
18.Is power conditioning equipment installed?
19.Is a dedicated branch circuit available for equipment?
20.Is the dedicated branch circuit less than 75 feet (22.86
m)?
21.Are the input circuit breakers adequate for equipment
loads?
Safety
NumberArea or ConditionYesNo
22.Is an emergency power shutoff switch available?
23.Is a telephone available for emergency purposes?
24.Does the computer room have a fire protection system?
25.Does the computer room have antistatic flooring installed?
26.Do any equipment servicing hazards exist (loose ground
wires, poor lighting, and so on)?
Cooling
NumberArea or ConditionYesNo
Table Continued
Site preparation13
Page 14
Check either Yes or No. If No, include comment number or date.Comment or
Date
27.Can cooling be maintained between 5°C (41 °F) and 35°C
(95 °F) (up to 1,525 m/5,000 ft)? Derate 1°C/305 m (34 °F/
1,000 ft) above 1,525 m/5,000 ft and up to 3,048 m/10,000
ft.
28.Can temperature changes be held to 5°C (9 °F) per hour
with tape media? Can temperature changes be held to
20°C (36 °F) per hour without tape media?
29.Can humidity level be maintained at 40% to 55% at 35°C
(95 °F) noncondensing?
30.Are air-conditioning filters installed and clean?
Storage
NumberArea or ConditionYesNo
31.Are cabinets available for tape and disc media?
32.Is shelving available for documentation?
Training
NumberArea or Condition
33.Are personnel enrolled in the System Administrator’s
Course?
34.Is on-site training required?
Power subsystem
The power subsystem is located on the system board. The BL860c i4 Server Blade receives 12 Volts
directly from the enclosure. The voltage is immediately passed through an E-fuse circuit, which will
immediately cut power to the blade if a short circuit fault or over current condition is detected. The E-fuse
can also be intentionally power cycled through the manageability subsystem. The 12V is distributed to
various points on the blade and is converted to lower voltages through power converters for use by
integrated circuits and loads on the blade.
14 Power subsystem
Page 15
ESD handling information
CAUTION:
Wear an ESD wrist strap when handling internal server components. Acceptable ESD wrist straps
include:
•The wrist strap that is included in the ESD kit with circuit checker (part number 9300-1609).
•The wrist strap that is included in the ESD kit without circuit checker (part number 9300-1608).
If the above options are unavailable, the throw away (one use only) strap that ships with some
Hewlett Packard Enterprise memory products can also be used, with increased risk of electrostatic
damage.
When removing and replacing server components, use care to prevent injury and equipment damage.
Many assemblies are sensitive to damage by electrostatic discharge.
Follow the safety precautions listed to ensure safe handling of components, to prevent injury, and to
prevent damage to the server blade:
•When removing or installing a server blade or server blade component, review the instructions
provided in this guide.
•Do not wear loose clothing that might snag or catch on the server or on other items.
•Do not wear clothing subject to static charge build-up, such as wool or synthetic materials.
•If installing an internal assembly, wear an antistatic wrist strap, and use a grounding mat such as those
included in the Electrically Conductive Field Service Grounding Kit.
•Handle components by the edges only. Do not touch any metal-edge connectors or electrical
components on accessory boards.
Unpacking and inspecting the server blade
Be sure that you have adequately prepared your environment for your new server blade, received the
components that you ordered, and verified that the server and the containers are in good condition after
shipment.
Verifying site preparation
Verifying site preparation is an essential factor of a successful server blade installation, and includes the
following tasks:
•Gather LAN information. Determine the two IP addresses for the iLO 3 MP LAN and the server blade
LAN.
•Establish a method to connect to the server blade console. For more information on console
connection methods, see Using iLO 3 on page 29 for more information.
•Verify electrical requirements. Be sure that grounding specifications and power requirements are met.
•Confirm environmental requirements.
Inspect the shipping containers for damage
Hewlett Packard Enterprise shipping containers protect their contents under normal shipping conditions.
After the equipment arrives, carefully inspect each carton for signs of shipping damage. Shipping damage
constitutes moderate to severe damage such as punctures in the corrugated carton, crushed boxes, or
large dents. Normal wear or slight damage to the carton is not considered shipping damage. If you find
shipping damage to the carton, contact your Hewlett Packard Enterprise customer service representative
immediately.
ESD handling information15
Page 16
Unpacking the server blade
Procedure
1. Use the instructions printed on the outside top flap of the carton.
2. Remove inner accessory cartons and the top foam cushions.
IMPORTANT:
Inspect each carton for shipping damage as you unpack the server blade.
3. Place the server blade on an antistatic pad.
Verifying the inventory
The sales order packing slip lists the equipment shipped from Hewlett Packard Enterprise. Use this
packing slip to verify that the equipment has arrived.
NOTE:
To identify each item by part number, see the sales order packing slip.
Returning damaged equipment
If the equipment is damaged, immediately contact your Hewlett Packard Enterprise customer service
representative. The service representative initiates appropriate action through the transport carrier or the
factory and assists you in returning the equipment.
16 Unpacking the server blade
Page 17
Installing the server blade into the enclosure
Installation sequence and checklist
StepDescriptionCompleted
1Perform site preparation (see Site preparation on page 10 for more
information).
2Unpack and inspect the server shipping container and then inventory the
contents using the packing slip.
3Install additional components shipped with the server. For these procedures, see
the documentation that with the component or the user service guide.
4Install and power on the server blade.
5Configure iLO 3 MP access.
6Access iLO 3 MP.
7Access UEFI from iLO 3 MP.
8Download latest firmware and update using HP Smart Update Manager.
9Install and boot the OS.
NOTE:
For more information regarding HPE Integrity Server Blade upgrades, see Upgrading a conjoined
configuration on page 59 for more information.
Installing and powering on the server blade
Preparing the enclosure
HPE BladeSystem enclosures ship with device bay dividers to support half-height devices. To install a full
height device, remove the blanks and the corresponding device bay divider.
CAUTION:
To prevent improper cooling and thermal damage, do not operate the server blade or the enclosure
unless all hard drive and device bays are populated with either a component or a blank.
Procedure
1. Remove the device bay blank.
Installing the server blade into the enclosure17
Page 18
2. Remove the three adjacent blanks.
Removing a c7000 device bay divider
Procedure
1. Slide the device bay shelf locking tab to the left to open it.
2. Push the device bay shelf back until it stops, lift the right side slightly to disengage the two tabs from
the divider wall, and then rotate the right edge downward (clockwise).
18 Removing a c7000 device bay divider
Page 19
3. Lift the left side of the device bay shelf to disengage the three tabs from the divider wall, and then
remove it from the enclosure.
Removing a c3000 device bay mini-divider or device bay divider
Procedure
1. Slide the locking tab down.
Removing a c3000 device bay mini-divider or device bay divider19
Page 20
2. Remove the mini-divider or divider:
a. c3000 mini-divider:
Push the divider toward the back of the enclosure until the divider drops out of the enclosure.
b. c3000 divider
I.Push the divider toward the back of the enclosure until it stops.
II.Slide the divider to the left to disengage the tabs from the wall.
III.Rotate the divider clockwise.
IV.Remove the divider from the enclosure.
20Installing the server blade into the enclosure
Page 21
Installing interconnect modules
For specific steps to install interconnect modules, see the documentation that ships with the interconnect
module.
Interconnect bay numbering and device mapping
•HPE BladeSystem c7000 Enclosure
•HPE BladeSystem c3000 Enclosure
Installing interconnect modules21
Page 22
To support network connections for specific signals, install an interconnect module in the bay
corresponding to the embedded NIC or mezzanine signals.
Server blade signalc7000 interconnect bay c3000 interconnect
bay
NIC 1 (Embedded)11
NIC 2 (Embedded)21
NIC 3 (Embedded)11
NIC 4 (Embedded)21
Mezzanine 13 and 42
Mezzanine 25 and 63 and 4
7 and 83 and 4
Mezzanine 35 and 63 and 4
7 and 83 and 4
For detailed port mapping information, see the BladeSystem enclosure installation poster or the
BladeSystem enclosure setup and installation guide for your product on the Hewlett Packard Enterprise
website (http://www.hpe.com/info/Blades-docs).
Interconnect bay
labels
Installing the server blade into the enclosure
NOTE:
When installing additional blades into an enclosure, additional power supplies might also be needed
to meet power requirements. For more information, see the BladeSystem enclosure setup and
installation guide for your product on the Hewlett Packard Enterprise website (http://www.hpe.com/
info/Blades-docs).
22 Installing the server blade into the enclosure
Page 23
Procedure
1. Remove the connector covers if they are present.
NOTE:
Before installing and initializing the server blade, install any server blade options, such as an
additional processor, hard drive, or mezzanine card.
2. Prepare the server blade for installation.
3. Install the server blade.
Installing the server blade into the enclosure23
Page 24
The server blade should come up to standby power. The server blade is at standby power if the blade
power LED is amber.
Server blade power states
The server blade has three power states: standby power, full power, and off. Install the server blade into
the enclosure to achieve the standby power state. Server blades are set to power on to standby power
when installed in a server blade enclosure. Verify the power state by viewing the LEDs on the front panel,
and using Table 5.
Table 5: Power States
Power StatesServer Blade
Installed in
Enclosure?
Standby powerYesNoYesNo
Full powerYesYesYesYes
OffNoNoNoNo
Powering on the server blade
Use one of the following methods to power on the server blade:
NOTE:
To power on blades in a conjoined configuration, only power on the Monarch blade. See Blade Link
bay location rules for rules on the definition of the Monarch blade.
•Use a virtual power button selection through iLO 3.
•Press and release the Monarch power button.
When the server blade goes from the standby mode to the full power mode, the blade power LED
changes from amber to green.
Front Panel Power
Button Activated?
Standby Power
Applied?
DC Power
Applied?
24 Server blade power states
Page 25
For more information about iLO 3, see Using iLO 3 on page 29.
Powering off the server blade
Before powering down the server blade for any upgrade or maintenance procedures, perform a backup of
critical server data and programs.
Use one of the following methods to power off the server blade:
NOTE:
To power off blades in a conjoined configuration, only power off the Monarch blade.
•Use a virtual power button selection through the iLO 3 GUI (Power Management, Power & Reset) or
the iLO 3 TUI commands.
This method initiates a controlled remote shutdown of applications and the OS before the server blade
enter standby mode.
•Press and release the Monarch power button.
This method initiates a controlled shutdown of applications and the OS before the server blade enter
standby mode.
•Press and hold the Monarch power button for more than 4 seconds to force the server blade to enter
standby mode.
This method forces the server blade to enter standby mode without properly exiting applications and
the OS. It provides an emergency shutdown method in the event of a hung application.
Installing the Blade Link for BL860c i4, BL870c i4 or
BL890c i4 configurations
IMPORTANT:
Without an attached Blade Link, the server blades will not power on.
NOTE:
Before installing the Blade Link for BL870c i4 or BL890c i4, make sure the following statements are
true:
•All blades have the same CPU SKUs
•All blades have the same hardware revision (only use BL860c i4, BL870c i4, or BL890c i4 Server
Blades)
•All blades have CPU0 installed
•All blades have the same firmware revision set
•All blades follow the memory loading rules for your configuration, see DIMMs on page 48
•The enclosure OA firmware is compatible with the blade firmware
•The Monarch blade has an ICH mezzanine card installed
•The proper Blade Link is being used for your configuration
To check on the blade hardware revisions and CPU SKUs, go to the Command Menu in the iLO 3
TUI and enter the DF command. This dumps the FRU content of the blades.
Powering off the server blade25
Page 26
NOTE:
If you will be upgrading an initial installation, see the user service guide for more information on
server blade upgrades.
Table 6: Blade Link bay location rules
Class Number of
conjoined blades
BL11 (standard for
BL860c i4)
Supported
enclosures
Blade location
rules
Partner blade
support?
Partner blade halfheight bay number /
Server blade fullheight bay number
c7000No specific bay
location rules for
blades
YesBottom half-height
adjacent bay, paired with
the server blade in fullheight bays 1&2, 3&4,
5&6, or 7&8
c3000Half-height bay 8, paired
with the server blade in
full-height bay 3.
CAUTION:
The bay minidivider must be
installed in the
c3000 enclosure
to ensure the
partner blade is
inserted correctly.
Failure to install
the bay minidivider might
result in damage
to the blade or
enclosure when
installing the
partner blade.
1
BL22 (BL870c i4)c7000Bays 1&2, 3&4,
c3000Bays 1&2, 3&4
26Installing the server blade into the enclosure
NoN/A
5&6, or 7&8 with
Monarch blade in
odd bay
with Monarch
blade in odd bay
Table Continued
Page 27
Class Number of
conjoined blades
Supported
enclosures
Blade location
rules
Partner blade
support?
Partner blade halfheight bay number /
Server blade fullheight bay number
BL2E2 (BL870c i4)c7000 onlyBays 2&3, 4&5
or 6&7 with
Monarch blade in
even bay using
full-height
numbering
2 (BL870c i4)c3000 onlyBays 2&3 with
Monarch blade in
even bay using
full-height
numbering.
BL44 (BL890c i4)c7000 onlyBays 1&2&3&4
or 5&6&7&8,
with Monarch
blade defaulting
to slot 1 or slot 5,
respectively
4 (BL890c i4)c3000 onlyBays 1&2&3&4
with Monarch
blade defaulting
to slot 1
YesBottom half-height bay 9
paired with full-height
bays 2&3, bottom halfheight bay 11 paired
with full-height bays
4&5, bottom half-height
bay 13 paired with fullheight bays 6&7
NoN/A
NoN/A
NoN/A
1
Upgrading a conjoined configuration on page 59
To install the Blade Link:
Procedure
1.Log on to the OA.
2.Install the first blade into the lowest bay number, this blade becomes the Monarch blade (Installing
3.Wait 10 seconds. The IP address of the installed blade appears in the OA.
For information on installing the c3000 bay mini-divider, see the HPE BladeSystem c3000 Enclosure Setup and
Installation Guide.
CAUTION:
Using the incorrect Blade Link can cause damage to the Blade Link and to the connectors on both
the Blade Link and the server blades.
IMPORTANT:
Failure to follow bay location rules can prevent server blade power on.
NOTE:
The manufacturing part numbers for the Blade Link is located on a sticker on the PCA.
the server blade into the enclosure on page 22).
Installing the server blade into the enclosure27
Page 28
4.Insert each adjacent blade, waiting 10 seconds between blades.
NOTE:
The blades will go into stand-by.
5.Using the OA, verify that the rest of the blades that will be conjoined have an IP address and are
powered off.
6.Remove the plastic protectors from the connectors on the back of the Blade Link.
7.Push in the blue release latch on the handle to release the handle.
8.Pull the handle all the way out.
9.Align the guide pins on the back of the Blade Link to the holes on the front of the server blades. As
you insert the pins into the holes, ensure the face on the Blade Link is evenly aligned parallel to the
face of the server blades.
10. Press firmly on the left and right sides of the Blade Link face until the handle naturally starts to close.
CAUTION:
If not properly aligned, you can damage the Blade Link.
11. Close the handle when it has engaged.
12. Log into iLO 3 on the Monarch blade. For more information, see the HPE Integrity iLO3 Operations
Guide.
13. In iLO 3, go to the Command Menu and execute xd -r to reboot all of the iLO 3s in the conjoined
set.
14. Run the conjoin checks
Integrity BL870c i4 and BL890c i4 systems go through a process called “conjoining” when the Blade
Link is attached. The system cannot boot until that process is completed properly.
a. Execute the following CM commands in the iLO 3 TUI to show data from all blades. This
information can be used to determine if the blades are successfully conjoined:
I.DF — Lists the FRUs on all of the blades (2 or 4).
II.SR — Shows a table of each blades firmware revisions.
III.Blade — Shows information about the OA and the bays used.
b. Check to see if the OA shows a properly conjoined system from its GUI.
28Installing the server blade into the enclosure
Page 29
IMPORTANT:
The secondary UUID and other system variables are stored on the Monarch blade. If you do
not put the Monarch blade in the leftmost slot, your system variables will not match. If you ever
change your iLO 3 configuration (such as adding users) that data is also stored on the Monarch
blade.
NOTE:
Auxiliary blades are not slot dependent after being installed and configured, however when the
conjoined systems ship, they come with A, B, C, D stickers located under the Blade Links.
While auxiliary blades are not slot dependent after being installed and configured, Hewlett
Packard Enterprise recommends using the shipped order to ensure proper auxiliary blade
function.
15. Still in the iLO 3 Command Menu, power on the Monarch blade with the PC -on -nc command.
Powering on the Monarch blade will power the entire conjoined system on.
16. Boot the Monarch blade. Booting the Monarch blade boots the entire conjoined system.
Using iLO 3
The iLO 3 subsystem is a standard component of selected server blades that monitors blade health and
provides remote server manageability. The iLO 3 subsystem includes an intelligent microprocessor,
secure memory, and a dedicated network interface. This design makes iLO 3 independent of the host
server and operating system. The iLO 3 subsystem provides remote access to any authorized network
client, sends alerts, and provides other server management functions.
Using iLO 3, you can:
•Remotely power on, power off, or reboot the host server.
•Subscribe to we-man alerts from iLO 3 regardless of the state of the host server.
•Access advanced troubleshooting features through the iLO 3 interface.
•Access Remote Console and vMedia functionality.
For more information about iLO 3 basic features, see the iLO 3 documentation on the Hewlett Packard
Enterprise website (
UEFI is an architecture that provides an interface between the server blade OS and the server blade
firmware. UEFI provides a standard environment for booting an OS and running preboot applications.
Use this procedure to access UEFI or the OS from the iLO 3 MP. Your security parameters were set
regarding remote access.
Procedure
1. Retrieve the factory iLO 3 password from the iLO 3 Network pull tag located on the right side of the
Monarch blade.
2. From the MP Main Menu, enter the co command to access the Console.
NOTE:
Terminal windows should be set to a window size of 80 columns x 25 rows for optimal viewing of
the console at UEFI.
3. After memory test and CPU late self test the following message appears:
Using iLO 329
Page 30
Press Ctrl-C now to bypass loading option ROM UEFI drivers.
4. The prompt will timeout if Ctrl-C is not pressed within a few seconds. If Ctrl-C is pressed, you will be
presented with two options: After selecting an option, boot will proceed.
NOTE:
If no option is selected, normal boot will proceed after ten seconds.
•Bypass loading from I/O slots.
•Bypass loading from I/O slots and core I/O.
The Bypass loading from I/O slots and core I/O option may be useful if a bad core I/O UEFI driver
is preventing system boot. USB drives can still be used at the UEFI shell to update core I/O drivers.
CAUTION:
Hitting Ctrl-C before the prompt will not work and may even disable this feature, be sure wait
for the prompt before hitting Ctrl-C.
NOTE:
It can take several minutes for this prompt to appear, and the window of time when Ctrl-C can
be pressed is very short. For typical boots, Hewlett Packard Enterprise recommends that you
let the prompt time out.
5. Depending on how the server blade was configured from the factory, and if the OS is installed at the
time of purchase, you are taken to:
a. UEFI shell prompt
b. OS login prompt
6. If the server blade has a factory-installed OS, you can interrupt the boot process to configure your
specific UEFI parameters.
7. If you are at the UEFI shell prompt, go to UEFI Front Page on page 30.
8. If you are at the OS login prompt, go to OS login prompt on page 33.
UEFI Front Page
If you are at the UEFI shell prompt, enter exit to get to the UEFI Front Page.
30 UEFI Front Page
Page 31
To view boot options, or launch a specific boot option, press B or b to launch the Boot Manager.
Installing the server blade into the enclosure31
Page 32
To configure specific devices, press D or d to launch the Device Manager. This is an advanced feature
and should only be performed when directed.
To perform maintenance on the system such as adding, deleting, or reordering boot options, press M or
m to launch the Boot Maintenance Manager.
32Installing the server blade into the enclosure
Page 33
To perform more advanced operations, press S or s to launch the UEFI Shell.
To view the iLO 3 LAN configuration, press I or i to launch the iLO 3 Setup Tool.
Saving UEFI configuration settings
There are other UEFI settings you can configure at this time. For more UEFI configuration options, see
RAID configuration and other utilities on page 133.
Booting and installing the operating system
From the UEFI Front Page prompt, you can boot and install in either of two manners:
•If your OS is loaded onto your server blade, see Operating system is loaded onto the server blade
on page 33.
•If the OS is not installed onto your server blade, see Operating system is not loaded onto the
server blade on page 33.
Operating system is loaded onto the server blade
If the OS is loaded on your server blade, normally UEFI will automatically boot to the OS. If the UEFI
Front Page is loaded, press ENTER to start auto boot, or B or b to select a specific boot option for your
OS.
Use your standard OS logon procedures, or see your OS documentation to log on to your OS.
Operating system is not loaded onto the server blade
There are two options on how to load the OS if it is not loaded onto your server blade.
•To load the OS from a DVD, see Installing the OS from an external USB DVD device or tape
device on page 35.
•To load the OS using HP Ignite-UX, see Installing the OS using Ignite-UX on page 36.
OS login prompt
If your server blade is at the OS login prompt after you establish a connection to the server blade, use
your standard OS log in procedures, or see your OS documentation for the next steps.
Saving UEFI configuration settings33
Page 34
Installing the latest firmware using HP Smart Update
Manager
The HP Smart Update Manager (HP SUM) utility enables you to deploy firmware components from either
an easy-to-use interface or a command line. It has an integrated hardware discovery engine that
discovers the installed hardware and the current versions of firmware in use on target servers. This
prevents extraneous network traffic by only sending the required components to the target. HP SUM also
has logic to install updates in the correct order and ensure all dependencies are met before deployment
of a firmware update. It also contains logic to prevent version-based dependencies from destroying an
installation and ensures updates are handled in a manner that reduces any downtime required for the
update process. HP SUM does not require an agent for remote installations.
Key features of HP SUM are:
•GUI and CLI–command line interface
•Dependency checking, which ensures appropriate installation order and dependency checking
between components
•Intelligent deployment deploys only required updates
•Support for updating firmware on network-based targets, such as the OA, iLO (through the Network
Management Port), and VC Ethernet modules
•Improved deployment performance
•Remote command-line deployment
•Windows X86 or Linux X86 support
HP SUM is included in the firmware bundles download from
the BL860c i4, BL870c i4, and BL890c i4.
For more information about HP SUM, see the HP Smart Update Manager User Guide (http://
www.hpe.com/info/hpsum/documentation).
http://www.hpe.com, and is supported on
34 Installing the latest firmware using HP Smart Update Manager
Page 35
Operating system procedures
Operating systems supported on the server blade
HP-UX 11i v3 HWE 1209
Installing the operating system onto the server blade
The following procedures describe generalized operating system installation. For more details, see the
operating system documentation.
Installing the OS from an external USB DVD device or tape device
NOTE:
Tapeboot requires BL8x0c i4 system firmware bundle 42.06 or later and a partner tape blade, or an
additional 51378-B21 Integrity Smart Array P711m HBA running 6.22 firmware or later to boot from
an Ultrium 6250 tape drive.
Procedure
1. If using an external USB DVD device:
a. Connect the Integrity SUV cable to the front of the Monarch server blade.
b. Connect the USB DVD cable to one of the USB ports on the SUV cable.
NOTE:
Some DVD drives might also require a separate power connection.
c. Turn on the external USB DVD device.
2. Insert the OS media into the USB DVD device or tape device.
3. Power on the server blade and boot to UEFI. If the server blade is already powered on, then reboot to
UEFI using the reset command at the UEFI prompt.
4. From the UEFI Front Page, press S or s to launch the UEFI Shell.
NOTE:
If the device is already selected or you already know the device name, then skip the following
step.
If you are using a tape device, when the UEFI shell comes up, you should see a message similar to
the following on the console:
•If the correct media is installed, it will boot from tape when you enter the index number.
•If there is no media in the SAS tape drive and you select 1, the following message appears:
tapeboot: Could not load tapeboot image
6. The OS now starts loading onto the server blade. Follow the on-screen instructions to install the OS
fully.
7. Continue with
Configuring system boot options on page 37
Installing the OS using Ignite-UX
Ignite-UX is an HP-UX administration toolset that enables:
•Simultaneous installation of HP-UX on multiple clients
•The creation and use of custom installations
•The creation of recovery media
•The remote recovery of clients
To install the OS onto the server blade using Ignite-UX, go to http://www.hpe.com/info/Blades-docs.
36 Installing the OS using Ignite-UX
Page 37
Installing the OS using vMedia
NOTE:
Installing the OS using vMedia might be significantly slower than installing using other methods.
vMedia enables connections of a DVD physical device or image file from the local client system to the
remote server. The virtual device or image file can be used to boot the server with an operating system
that supports USB devices.
vMedia depends on a reliable network with good bandwidth. This is especially important when you are
performing tasks such as large file transfers or OS installations.
For more information regarding loading the OS with vMedia, see the vMedia Chapter of the HPE Integrity
After the OS is loaded, make sure to save your nonvolatile memory settings to preserve boot entries
in case of blade failure.
Configuring system boot options
•Boot Manager
Contains the list of boot options available. Ordinarily the boot options list includes the UEFI Internal
Shell and one or more operating system loaders.
To manage the boot options list for each server, use the UEFI Shell, the Boot Maintenance Manager,
or operating system utilities.
•Autoboot setting
The autoboot setting determines whether a server automatically loads the first item in the boot
options list or remains at the UEFI Front Page menu. With autoboot enabled, UEFI loads the first item
in the boot options list after a designated timeout period.
Configure the autoboot setting for an Integrity server using either the autoboot UEFI Shell command
or the Set Time Out Value menu item from the Boot Maintenance Manager.
Examples of autoboot commands for HP-UX:
◦Disable autoboot from the UEFI Shell by issuing autoboot off
◦Enable autoboot with the default timeout value by issuing autoboot on
◦Enable autoboot with a timeout of 60 seconds by issuing the autoboot 60
◦Set autoboot from HP-UX using setboot
◦Enable autoboot from HP-UX using setboot -b on
◦Disable autoboot from HP-UX using setboot -b off
For more information on the autoboot command, enter help autoboot.
Booting and shutting down HP-UX
To boot HP-UX, use one of the following procedures:
•To boot HP-UX normally, see HP-UX standard boot on page 38. HP-UX boots in multi-user mode.
•To boot HP-UX in single-user mode, see Booting HP-UX in single-user mode on page 40.
•To boot HP-UX in LVM-maintenance mode, seeBooting HP-UX in LVM-maintenance mode on page
40.
Installing the OS using vMedia37
Page 38
Adding HP-UX to the boot options list
You can add the \EFI\HPUX\HPUX.EFI loader to the boot options list from the UEFI Shell or the Boot
Maintenance Manager.
NOTE:
On Integrity server blades, the operating system installer automatically adds an entry to the boot
options list.
NOTE:
To add an HP-UX boot option when logged in to HP-UX, use the setboot command. For more
information, see the setbootvolume(1M) manpage.
To add HP-UX to the list:
Procedure
1. Access the UEFI Shell environment.
a. Log in to iLO 3 for Integrity and enter the CO command to access the system console.
When accessing the console, confirm that you are at the UEFI Front Page.
If you are at another UEFI menu, then choose the Exit option or press X or x to exit the menu. Exit
until you return to the screen that lists the keys that can be pressed to launch various Managers.
b. Press S or s to launch the UEFI shell.
2. Access the UEFI System Partition (fsX: where X is the file system number) for the device from which
you want to boot HP-UX.
For example, enter fs2: to access the UEFI System Partition for the bootable file system number2.
The UEFI Shell prompt changes to reflect the file system currently accessed.
The full path for the HP-UX loader is \EFI\HPUX\HPUX.EFI and it should be on the device you are
accessing.
3. At the UEFI Shell environment, use the bcfg command to manage the boot options list.
The bcfg command includes the following options for managing the boot options list:
a. bcfg boot dump – Display all items in the boot options list for the server.
b. bcfg boot rm # – Remove the item number specified by # from the boot options list.
c. bcfg boot mv #a#b – Move the item number specified by #a to the position specified by #b in
the boot options list.
d. bcfg boot add #file.efi "Description" – Add a new boot option to the position in the
boot options list specified by #. The new boot option references file.efi and is listed with the title
specified by Description.
For example, bcfg boot add 1 \EFI\HPUX\HPUX.EFI "HP-UX 11i v3" adds an HP-UX
11i v3 item as the first.
For more information, see the help bcfg command.
4. Exit the console and iLO 3 MP interfaces.
Press Ctrl–B to exit the system console and return to the iLO 3 MP Main Menu. To exit the MP, press
X or x at the Main Menu.
HP-UX standard boot
Use either of the following procedures to boot HP-UX:
38 Adding HP-UX to the boot options list
Page 39
•Booting HP-UX from the UEFI Boot Manager on page 39
•Booting HP-UX from the UEFI Shell on page 39
Booting HP-UX from the UEFI Boot Manager
Procedure
1. From the UEFI Boot Manager menu, choose an item from the boot options list to boot HP-UX.
2. Access the UEFI Boot Manager menu for the server on which you want to boot HP-UX.
3. Log in to iLO 3 MP and enter the CO command to choose the system console.
4. Confirm you are at the UEFI Front Page. If you are at another UEFI menu, then choose the Exit
option or press X or x to exit the menu. Exit until you return to the screen that lists the keys that can be
pressed to launch various Managers. Press B or b to launch the Boot Manager.
5. At the UEFI Boot Manager menu, choose an item from the boot options list.
Each item in the boot options list references a specific boot device and provides a specific set of boot
options or arguments you use when booting the device.
6. Press Enter to initiate booting using your chosen boot option.
7. Exit the console and iLO 3 MP interfaces.
8. Press Ctrl–B to exit the system console and return to the MP Main Menu. To exit the MP Main Menu,
press X or x.
Booting HP-UX from the UEFI Shell
Procedure
1. Access the UEFI Shell.
2. From the UEFI Front Page, press S or s to launch the UEFI shell.
3. Use the map command to list the file systems (fs0, fs1, and so on) that are known and have been
4. To select a file system to use, enter its mapped name followed by a colon (:). For example, to operate
5. Enter HPUX at the UEFI Shell command prompt to launch the HPUX.EFI loader from the currently
6. Allow the HPUX.EFI loader to proceed with the boot command specified in the AUTO file, or manually
mapped.
with the boot device that is mapped as fs0, enter fs0: at the UEFI Shell prompt.
selected boot device.
If needed, specify the full path of loader by entering \EFI\HPUX\HPUX at the UEFI Shell command
prompt.
specify the boot command.
By default, the HPUX.EFI loader boots using the loader commands found in the \EFI\HPUX\AUTO
file on the UEFI System Partition of the selected boot device. The AUTO file typically contains the bootvmunix command.
To interact with the HPUX.EFI loader, interrupt the boot process (for example, type a space) within the
time-out period provided by the loader. To exit the loader, use the exit command, which returns you to
UEFI.
Booting HP-UX from the UEFI Boot Manager39
Page 40
Booting HP-UX in single-user mode
Procedure
1. Use steps 1–5 from Booting HP-UX from the UEFI Shell on page 39 to access the UEFI shell and
launch the HPUX.EFI loader.
2. Access the HP-UX Boot Loader prompt (HPUX>) by pressing any key within the 10 seconds given for
interrupting the HP-UX boot process. Use the HPUX.EFI loader to boot HP-UX in single-user mode in
step 3.
After you press a key, the HPUX.EFI interface (the HP-UX Boot Loader prompt, HPUX>) launches. For
help using the HPUX.EFI loader, enter the help command. To return to the UEFI Shell, enter exit.
3. At the HPUX.EFI interface (the HP-UX Boot loader prompt, HPUX>) enter the boot -is vmunix
command to boot HP-UX (the /stand/vmunix kernel) in single-user (-is) mode.
Booting HP-UX in LVM-maintenance mode
The procedure for booting HP-UX into LVM Maintenance Mode is the same as for booting into single user
mode (Booting HP-UX in single-user mode on page 40), except use the -lm boot option instead of the
-is boot option:
HPUX> boot -lm vmunix
Shutting down HP-UX
For more information, see the shutdownvolume(1M) manpage.
Procedure
1. Log in to HP-UX running on the server that you want to shut down or log in to iLO 3 MP for the server
2. Issue the shutdown command with the appropriate command-line options.
and use the Console menu to access the system console. Accessing the console through iLO 3 MP
enables you to maintain console access to the server after HP-UX has shut down.
The command-line options you specify determines the way in which HP-UX shuts down and whether
the server is rebooted.
Use the following list to choose an HP-UX shutdown option for your server:
a. Shut down HP-UX and halt (power off) the server using the shutdown -h command.
Reboot a halted server by powering on the server using the PC command at the iLO 3 MP
Command menu.
b. Shut down HP-UX and reboot the server by issuing the shutdown -r command.
40 Booting HP-UX in single-user mode
Page 41
Optional components
If your server blade has no additional components to install, go to Installing and powering on the
server blade on page 17.
Partner blades
The following partner blades are supported:
•Ultrium 448c Tape Blade
•SB920c Tape Blade
•SB1760c Tape Blade
•SB3000c Tape Blade
•SB40c Storage Blade
•D2200sb Storage Blade
IMPORTANT:
In c7000 enclosures, partner blades are supported with BL860c i2 servers and BL870c i2 servers
with BL2E blade links.
In c3000 enclosures, partner blades are supported with BL860c i2 servers.
Partner blades are not supported with BL890c i2 servers.
Partner blade slotting rules are dependent on the conjoined blade configuration. For more
information on partner bay blade locations, see Installing the Blade Link for BL860c i4, BL870c
i4 or BL890c i4 configurations on page 25.
NOTE:
SAS tape boot is now supported with tape blades on HPE Integrity BL860c i4, BL870c i4, and
BL890c i4 server blades.
NOTE:
PCIe Gen-1 (driven from the ICH10 on the monarch blade’s ICH mezzanine card) provides the
interface to the partner blades. A separate mezzanine pass-thru card is not required.
Hot-plug SAS disk drives
The server blade supports up to two hot-plug SAS drives.
CAUTION:
To prevent improper cooling and thermal damage, do not operate the server blade or the enclosure
unless all hard drive and device bays are populated with either a component or a blank.
IMPORTANT:
The disk drive does not seat properly when 180° out of alignment. Verify the orientation before
insertion.
Optional components41
Page 42
Procedure
1. Remove the hard drive blank.
NOTE:
For a list of supported disk drives for the server blade, see: Server blade components list on page
109.
2. Prepare the hard drive.
3. Slide the drive into the cage until it is fully seated.
4. Close the lever to lock the drive into place.
42Optional components
Page 43
Installing internal components
Removing the access panel
Procedure
1. Lift the access panel latch.
2. Slide the access panel backwards approximately 2 cm (0.75 in).
3. Remove the access panel by lifting it straight up and off the server blade.
After the access panel is off, you can do the following:
•Add an additional processor (Processor and heatsink module on page 43).
•Add additional memory DIMMs (DIMMs on page 48).
•Add additional mezzanine cards (Mezzanine cards on page 54).
Processor and heatsink module
Processor load order
Observe the following guidelines when installing additional processors:
•In a BL860c i4, CPU0 is installed before CPU1.
•In a BL870c i4 or BL890c i4, each blade must have CPU0 installed.
•When adding additional CPUs in a conjoined configuration:
◦Load both CPU0 and CPU1 in the Monarch blade first.
◦Load additional CPUs in sequence, from lowest slot-numbered blade to highest.
CAUTION:
The pins on the processor socket are very fragile. Any damage to them might require replacing the
server blade base unit.
CAUTION:
To avoid damage to the processor, verify that the plastic tabs on the processor are pulled fully out
before installation.
Installing internal components43
Page 44
The processor and heatsink ship as two separate units and are coupled together during installation into
the server blade.
Procedure
1. Remove the CPU airflow baffle.
CAUTION:
To avoid damage to the processor, handle the processor only by the edges. Do not touch the bottom
of the processor, especially the contact area.
CAUTION:
To prevent possible server blade malfunction and damage to the equipment, multiprocessor
configurations must contain processors with the same part number.
2. Transfer the duplicate part/serial numbers label from the processor module to the processor heatsink.
a. Remove the duplicate tear-away label that lists the part and serial numbers from the processor
module.
b. Place the label on the top of the heatsink.
3. Install the processor over the load posts.
NOTE:
Ensure pin 1, indicated on the empty socket with an embossed triangle, matches the pin 1
marker on the processor module, the chamfered corner of its attached voltage regulator
heatsink.
44Optional components
Page 45
4. Remove the heatsink cover.
CAUTION:
During installation, after removing the protective cover from the heatsink:
•Do not touch or come into contact with the thermal interface material.
•Immediately install the heatsink.
CAUTION:
To avoid damage to the server blade and processor, ensure the processor heatsink locking
handle is fully back against the stops, rotated about 120° back. Also verify that the plastic tabs
on the processor heatsink are pulled fully out before installation.
5. Install the heatsink over the load posts.
CAUTION:
Do not lower the heatsink locking handle before pushing the plastic locking tabs into place.
Optional components45
Page 46
CAUTION:
To prevent thermal instability and damage to the server blade, do not separate the processor
module from the processor's heatsink after they have been coupled.
6. Secure the heatsink to the processor
a. Slide both plastic locking tabs into place. (See callout 1 in the following figure).
b. Grasp and rotate the latch downward. (See callout 2 in the following figure.)
WARNING:
The heatsink locking lever can constitute a pinch hazard, keep your hands on top of the lever
during installation to avoid personal injury.
46Optional components
Page 47
NOTE:
Positive engagement clicking should occur during mating of the processor heat sink and
processor module onto the socket to ensure proper seating.
7. Connect the power cord. (See callout 3 in the previous figure).
8. Tie wrap the processor cable to the right tie point on the processor assembly.
CAUTION:
When the CPU is installed, dress all slack in the power cable to the connector end of the cable.
Failure to do so could result in pinched or damaged CPU power cables.
IMPORTANT:
If you are adding an additional processor to your server blade, the DIMMs in the server blade must
be reconfigured to support both CPUs. For more information, see DIMM pair load order.
Optional components47
Page 48
DIMMs
DIMM installation guidelines
Observe the following guidelines when installing memory:
•Use only Hewlett Packard Enterprise low-profile (1.2 in.) DIMMs.
•In a BL860c i4 Server Blade, memory is loaded in identical pairs.
•In a BL870c i4 or BL890c i4 Server Blade, memory is loaded in quads that are systematically loaded
IMPORTANT:
DIMMs from other sources may adversely affect data integrity.
between blades and processors.
◦The BL870c i4 has a single interleaving domain that allows mixing two DIMM types within the
domain. The highest DIMMs get loaded first in quads, alternating between blades and processor
sockets; then the lower DIMMs get loaded in quads, alternating between blades and processor
sockets .
◦The BL890c i4 system has dual interleaving domains, with each interleave domain consisting of
two adjacent blades. In this system different DIMM types cannot be mixed in the same interleaving
domain. Higher capacity DIMMs are loaded in quads alternating between blades 1 and 2 and their
processor sockets; then the lower capacity DIMMs are loaded in quads alternating between blades
3 and 4 and their processor sockets. This means there can never be more than 12 quads of either
of the two DIMM types.
IMPORTANT:
If loading a single DIMM type in an 8 socket system, the DIMMs have to be loaded alternating
between blades and processor sockets since starting to load one interleaving domain first,
then the second can result in an unbalanced memory configuration.
•Different densities in different pairs or quads are supported, provided the following rules are kept.
Only two DIMM sizes can be mixed in a BL870c i4 or BL890c i4 Server Blade. Mixing more than two
DIMM pair capacities is only supported on the BL860c i4.
◦In a BL870c i4 Server Blade, the two DIMM sizes must alternate between blades. Load all of one
DIMM size, then repeat with the second DIMM size.
◦In a BL890c i4 Server Blade, blades 1 and 2 must have one DIMM size and blades 3 and 4 must
have the other DIMM size. If only one DIMM size is used on BL980c i4, alternate between all 4
blades.
IMPORTANT:
If loading a single DIMM type in a BL980c i4, the DIMMs have to be loaded alternating
between blades and processor sockets, since starting to load one interleaving domain first,
then the second, can result in an unbalanced memory configuration.
NOTE:
Integrity Server Blade upgrades follow the same DIMM loading rules. See Upgrading a
conjoined configuration on page 59 for more information on server blade upgrades.
Performance
The maximum number of usable DIMM slots, and therefore the maximum amount of memory for any
particular configuration is tied to the number of processors used in the configuration. In addition, the
conjoined blade products all have minimum processor recommendations (oner per conjoined blade) for
best performance.
48 DIMMs
Page 49
There are 12 DIMMs associated with each processor. Each processor has DIMM pairs A through F. A
processor must installed to access its associated memory. A processor can be installed without memory
as long as one of the processors in the conjoined system has memory.
For best performance, the minimum recommendation is a quad of DIMMs for each processor installed in
a conjoined blade product. For best performance (not required), load only the same type (size) of DIMMs
throughout the system. Uniform memory (selected in the memconfig interleaving options) is supported
with a total memory capacity equal to a "power of two" (loading 4, 8, 16, 32, or 64 DIMMs at a time).
Therefore, for best performance, if using the uniform memory selections, load a quad (even better two
quads) of DIMMs for each processor. To obtain the fastest memory access, load a total memory capacity
equal to a power of four (4, 16, 64).
IMPORTANT:
The system will boot with memory loaded on only one processor. However, from a performance
perspective, this is not a recommended configuration.
To enable Double DRAM chip sparing, for all the DIMMs loaded in the system, use only 4GB, 8GB, or
16GB DIMMs.
The 10GigE LOMs that have jumbo frames enabled require 1GB of main memory for each port (8GB/
server blade). Anything below this minimum results in insufficient memory and a hang either during the
HP-UX boot or during the I/O configuration for the Integrity BL8x0c i4 systems. For example, the minimum
memory configuration for the BL860c i4 is 8GB, for the BL870c i4 is 16GB, and for the BL890c i4 is
32GB.
Minimum memory requirements assume the Flex-10/FlexFabric adapters are not installed. Increase the
minimum required memory by 2GB for each Flex-10/FlexFabric mezzanine card that is installed into the
system.
DIMM pair load order
Figure 1: DIMM quad load order
Optional components49
Page 50
Figure 2: Mixed DIMM load order
Table 7: DIMM pair load order
CPU0CPU1
CPU0 only
Both CPUs loaded
1st3A4A——
2nd9B10B——
3rd1C6C——
4th7D12D——
5th2E5E——
6th8F11F——
1st3A4A——
2nd——1A7A
3rd9B10B——
4th——6B10B
5th1C6C——
6th——3C9C
50Optional components
7th7D12D——
8th——4D12D
Table Continued
Page 51
9th2E5E——
10th——2E8E
11th8F11F——
12th——5F11F
DIMM quad load order rules
CPU0CPU1
Figure 3: DIMM quad load order
Figure 4: Mixed DIMM load order
Optional components51
Page 52
Table 8: DIMM quad load order
CPU0CPU1
Both
CPUs
loaded
NOTE:
•If more than two DIMM types are installed in the BL870c i4 or BL890c i4 systems, the customer
•If DIMMs are loaded in pairs in the BL870c i4 or BL890c i4 systems, the customer will receive a
Once the memory loading rules have been satisfied, all memory installed in the system will be
recognized.
1st3A4A9B10B————
2nd————1A7A6B10B
3rd1C6C7D12D————
4th————3C9C4D12D
5th2E5E8F11F————
6th————2E8E5F11F
will receive a warning indicating that optimum interleaving is not possible and memory may be
de-allocated. Mixing more than two DIMM types is only supported on the BL860c i4.
warning indicating that the memory is not configured according to the Hewlett Packard
Enterprise loading rules and some memory may be de-allocated.
Installing DIMMS
Procedure
1. Power off the server blade.
2. For the conjoined products only, remove the Blade Link.
3. Remove the server blade from the enclosure.
4. Remove the access panel.
5. Remove the DIMM baffle (DIMM baffle on page 120).
6. Locate the DIMM slots on the server blade system board.
52Optional components
Page 53
NOTE:
The server blade ships with at least two DIMMs installed in slots 3A and 4A.
7. Ensure the DIMM slot latches are open.
8. Align the DIMM's notch with the slot's notch.
CAUTION:
Use only Hewlett Packard Enterprise low profile DIMMs. DIMMs from other sources might
adversely affect data integrity.
DIMMs do not seat fully if turned the wrong way.
DIMMs in a pair or quad must be identical.
9. Insert a DIMM in a slot and push down firmly until the latches click shut, first one latch and then the
other.
IMPORTANT:
To ensure proper function of the memory baffle, all DIMM latched must be in the CLOSED
position.
Optional components53
Page 54
Mezzanine cards
Optional mezzanine cards enable additional network connectivity and provide Fibre Channel support. For
mezzanine card locations, see Server blade components on page 9.
Optional mezzanine cards are classified as Type I mezzanine cards and Type II mezzanine cards. The
card type determines where it can be installed in the server blade:
•Install Type I mezzanine cards on any mezzanine connector.
•Install Type II mezzanine cards only on Mezzanine 2 connector or Mezzanine 3 connector.
To install a card:
Procedure
1. Power off the server blade and remove it from the server blade enclosure (Preparing the server
blade for servicing on page 112).
2. Remove the access panel (Removing the access panel on page 43).
3. Remove the mezzanine connector cover.
54 Mezzanine cards
Page 55
4. Align the mezzanine connector on the option card with the mezzanine connector on the system board.
5. Press down on the connector to seat the card.
CAUTION:
To prevent damage to the server blade, apply pressure over the mezzanine connector when
installing the mezzanine card. Do not apply pressure to the edges of the card.
HPE Smart Array P711m Controller
The Smart Array P711m is a PCIe card supporting direct attach and shared SAS Storage. This card is
supported in slots 1, 2, and 3 on the BL860c i4, BL870c i4, and BL890c i4 server blades.
HPE Smart Array P711m Controller55
Page 56
NOTE:
When a mixed configuration of BL860c i4, BL870c i4, and BL890c i4 server blades and other cClass server blades is in the same enclosure, the preferred location for the P711m cards is slot 3.
Supercap pack mounting kit
The Smart Array P711m Controller includes a cabled Supercap Pack which mounts inside the server
using the orderable AM341A Mounting Kit for Low profile battery (Hewlett Packard Enterprise part
number AD399-2132A KIT, RAID BATTERY HOLDER). This kit contains a black metal mounting bracket
and a mezzanine card blank. The mounting bracket clips onto a mezzanine card in slot 1. If no card is
used in slot 1, the mezzanine card blank included in the kit must be installed to mount the bracket.
For installing the bracket on to the server blade, the mounting bracket has:
•A mezzanine post loop that slides over the abutting mezzanine 2 post
•A thumbscrew clip that clips over the mezzanine 1 card or mezzanine blank thumbscrew located at the
rear of the blade
•A mezzanine foot that clips over the edge of the mezzanine 1 card or mezzanine card blank that is
installed closest to the center of the blade
For installing the Supercap Pack onto the bracket, the mounting bracket has:
•A C-hook to restrain the cable
•Two restraining clips to restrain the Supercap Pack
1Mezzanine foot3Cable restraining clips
2Mezzanine post loop4C-hook
Installing the Supercap mounting bracket
Procedure
1. Power off the server blade and remove it from the server blade enclosure (Preparing the server
blade for servicing on page 112).
2. Remove the access panel (Access panel on page 117).
3. Remove the DIMM baffle (DIMM baffle on page 120).
4. Temporarily remove any mezzanine card installed in slot 2 (Mezzanine cards on page 127).
56 Supercap pack mounting kit
Page 57
NOTE:
Mezzanine slot 1 must contain the P711m Controller, another mezzanine card, or the mezzanine
card blank. If you are installing the P711m card in mezzanine slot 1, do so now following the
instructions “Installing the P711m Controller Board.” If you are not installing the P711m card in
mezzanine slot 1, and no other mezzanine card is present in slot 1, then install the mezzanine
card blank at this time.
5. Clip the mounting bracket onto the mezzanine card or mezzanine card blank.
6. Press the bracket thumbscrew clip over the thumbscrew closest to the mezzanine post.
7. Push the remaining bracket foot over the other edge of the mezzanine card or mezzanine card blank.
CAUTION:
Push only enough to anchor the bracket, keeping the bracket level.
CAUTION:
To avoid damage to the power cable ensure the battery bracket does not pinch the power cable
to the power connectors.
8. Replace any mezzanine cards that you had to remove from mezzanine slot 2 (Mezzanine cards on
page 127).
To remove the component, reverse the procedure.
Installing the P711m controller board
This card is supported in slots 1, 2, and 3 on the BL860c i4, BL870c i4, and BL890c i4 server blades.
Installing the P711m controller board57
Page 58
1Slot 12Slot 23 Slot 3
NOTE:
When a mixed configuration of BL860c i4, BL870c i4, and BL890c i4 server blades and other cClass server blades are in the same enclosure, the preferred location for the P711m cards is slot 3.
Use the procedure for installing standard mezzanine cards to install the SAS controller board, see
Mezzanine cards on page 54 for more information.
Installing the Supercap Pack
The Supercap Pack snaps onto the mounting bracket to secure the Supercap Pack inside the server
blade.
NOTE:
The mounting bracket is installed either on top of a mezzanine card in slot 1, or the mezzanine
blank in slot 1, depending upon the customer's configuration. If no card is used in slot 1, the
mezzanine card blank must be installed for the bracket to clip onto.
NOTE:
The Supercap Pack includes two plastic tabs that must be removed for proper fit in the BL860c i4,
BL870c i4, and BL890c i4 server blades. One tab is located on the rounded end of the Supercap
Pack, the other tab is directly above the cable as it exits the plastic housing of the Supercap Pack.
To remove these tabs, squeeze them firmly toward each other (significant force is required).
Optionally, these two tabs can be clipped off even with the rest of the Supercap Pack.
58 Installing the Supercap Pack
Page 59
1. Position the Supercap Pack with the print side down and push the end of the cable that is connected
to the Supercap Pack under the mounting bracket’s C-hook, routing the cable so its supercap end is
on the side of the hook that is away from slot 3.
2. Push the other end of the Supercap Pack down between the two restraining clips.
3. Consolidate and secure remaining cable length using tie wrap or similar retention method.
To remove the component, reverse the procedure.
Replacing the access panel
Procedure
1. Place the access panel onto the server blade by lining up the keyways on the panel to the posts on the
server blade chassis.
2. Slide the access panel toward the front of the server blade until the panel lock button snaps into place.
Upgrading a conjoined configuration
This section contains information required to complete an upgrade of a BL860c i4, BL870c i4, or BL890c
i4 server blade. This document expands on the information provided in the Read This First document
which is included in the upgrade kit.
Replacing the access panel59
Page 60
Procedure summary
Procedure
1. Original blade server check
The existing blade servers that will be upgraded are checked and prepared for the upgrade.
2. Adding resources
The blades for the upgrade are placed together, internal component changes are made if necessary,
blade firmware is updated if necessary, the Upgrade Blade Link is installed, System Information
Parameters are exchanged between the server blades and the new Blade Link, and then the final
system partitioning can be set.
3. Final check
The upgraded sever partitions are booted to the UEFI shell and variables restored if necessary, the
upgraded server is checked for normal operation at a low level, and finally the Operating System is
booted and the system checked for normal operation.
4. Warranty registration
Warranty and support are arranged to be transferred from the server being upgraded to the upgraded
server.
Upgrade kit contents
•The HPE Integrity BL860c i4, BL870c i4 & BL890c i4 Server Blade Upgrades Read Me First.
•One of the following Upgrade Blade Links:
◦AM395A HPE BL8x0c i4 Upgrade BL2 Blade Link
◦AM396A HPE Integrity BL870c i4 Upgrade BL2-c7E Blade Link
◦AM398A HPE BL8x0c i4 Upgrade BL4 Blade Link
NOTE:
Each Blade Link will include a new Product Number label to attach to the system label carrier
card; the actual part number will depend on the type of upgrade ordered.
•AM394A HPE Integrity BL8x0c i4 Upgrade Blade. One or more Upgrade Blades may be included
depending on the type of upgrade ordered. These should be pre-installed with ordered processors,
memory, and other components.
60 Procedure summary
Page 61
Before getting started
IMPORTANT:
•Do not install or use the upgrade Blade Link until explicitly called for in the upgrade procedure.
For more information, see Blade link and system information parameters on page 75.
•The server blade firmware may need to be updated on the existing server blades as part of the
upgrade process, see Determining your current firmware revisions on page 62 for more
information.
•Processor and memory DIMM loading and distribution follows applicable rules across the blades.
For more information, see Installing internal components on page 43 and Rules for serverblade internal components on page 66.
•Adding new Upgrade Blades to an enclosure may require additional power supplies in the
enclosure to meet the new power requirements. This situation would be reported by the
enclosure Onboard Administrator. For more information, see the BladeSystem setup and
information guide for your enclosure at http://www.hpe.com/info/Blades-docs.
•If more than one existing blade server will be combined in the upgrade, one of them must be
designated as the “Monarch Server”. This is the original server which is officially being upgraded,
The Monarch blade from this server will become the Monarch blade in the upgraded server. Also,
the Monarch Server UUID and Serial Number will be transferred to the upgraded server.
•The OS boot disk must be attached to the designated Monarch Server.
•When VC is used, some rearrangement of the VC server profile may be necessary. See
Possible changes due to VC profile mapping on the upgraded blade server on page 77.
•Because the socket count of the upgraded server is likely to be greater than that of the server
being upgraded, a new OS license will probably be required.
•At the end of the upgrade, the Blade Links removed from the original blade servers cannot be
reused.
Supported operating systems
HP-UX 11i v3 OE Update for September 2012
IMPORTANT:
If you will be using HP-UX, review the HP-UX errata documentation that is listed at http://
www.hpe.com/info/Blades-docs.
Minimum firmware versions
A minimum set of firmware is required for the blades and blade enclosures involved in the upgrade.
Firmware on the server blades, the enclosure HPE Onboard Administrator, and the HPE Virtual Connect
Manager (if Virtual Connect interconnect modules are installed) must be updated to at least the minimum
firmware level. Every blade in the upgrade must be at the same firmware level.
IMPORTANT:
Any required updates must be checked for and completed during the upgrade process.
Support & Drivers —> enter
product name —> select
operating system —> “Cross
operating system”
Support & Drivers —> enter
product name
Go to http://www.hpe.com —>
Support & Drivers —> enter
Interconnect Module product
name
Procedure
1. Log into the system iLO 3 MP.
2. From the Main Menu, enter the cm command to access the Command Menu.
3. Enter the sr command to view the system FW revisions.
It is possible to access the iLO TUI before iLO has completed its initialization at power up. The
indicator to know that iLO has completed initialization is the ability to access the nPartitions
command via the TUI.
Hewlett-Packard Integrated Lights-Out 3 for Integrity
(C) Copyright 1999-2010 Hewlett-Packard Development Company, L.P.
MP Host Name: ilo002264fee2de
iLO MP Firmware Revision 01.41.01
*************************************************************************
MP ACCESS IS NOT SECURE
A default MP user Administrator/Admin is currently configured
and remote access is enabled. Modify the default password or
delete the default user (see UC command)
OR
Disable all types of remote access (see SA command)
4. Compare the first four Revisions lines with the version information listed in the release notes at
www.hpe.com for the current server blade, HP Smart Update Manager firmware bundle. If any of the
firmware revisions are less than those in the release notes, then the server firmware must be updated.
http://
OA
Use one of the following methods to determine the current OA firmware version:
•OA GUI — Connect to the OA GUI, expand the Active Onboard Administrator section of the left
sidebar menu, and select Firmware Update. The Firmware Version is listed under FirmwareInformation.
Optional components63
Page 64
•OA CLI — Log in to the OA CLI through the serial console port or LAN and enter show oa info at
the command prompt.
OA8> show oa info
Onboard Administrator #1 information:
Product Name : BladeSystem c7000 Onboard Administrator with KVM
Part Number : 456204-B21
Spare Part No.: 503826-001
Serial Number : OB12BP7320
UUID : 09OB12BP7320
Manufacturer : HP
Firmware Ver. : 3.60 Jun 11 2012
Hw Board Type : 0
Hw Version : B1
VCM
Use one of the following methods to determine the current VCM firmware version:
OA GUI — Select the Interconnect Bay 1 device and then select the Information tab. Under
Information there is a Firmware Version line which indicates the current VC FW version installed on the
primary VC interconnect module.
64Optional components
Page 65
VC Support Utility — Launch VC Support Utility – Interactive and enter version.
HP BladeSystem c-Class Virtual Connect Support Utility
Version 1.7.0 (Build 95)
Build Date: Oct 13 2010 07:03:49
Copyright (C) 2007-2010 Hewlett-Packard Development Company, L.P.
All Rights Reserved
Please enter action ("help" for list): version
Please enter Onboard Administrator IP Address: 10.3.0.8
Please enter Onboard Administrator Username: Administrator
Please enter Onboard Administrator Password: *****
The target configuration is integrated into a Virtual Connect Domain. Please
enter the Virtual Connect Domain administrative user credentials to continue.
User Name: Administrator
Password: *****
=======================================================================
Enclosure Bay Module Version
=======================================================================
USE827CYVC 1 HP VC Flex-10 Enet Module 3.15 2010-10-09T07:18:16Z
Total execution time: 00:01:31
Press Return/Enter to exit...
Rules for server blade internal components
You may need to remove or redistribute internal components across server blades to complete the
upgrade. The following sections cover some key facts regarding internal components in upgraded servers
and directs the reader to locations for additional information.
Processor loading and matching rules
There are two processor sockets on each blade: socket 0 and socket 1. Socket 0 of each blade in the
server must be filled. Any additional processors must be loaded in socket 1 of the blade in the lowest
numbered enclosure bay which has an empty socket 1.
Every processor must be of the same type and must have the following:
•The same core count.
•The same core frequency.
•The same cache size.
This can be checked using the following methods:
•Examine the part number marked on each processor and each blade to make sure they are the same
on each component.
•Examine the core count, core frequency, and cache size of each processor to ensure that every
processor in the upgrade is the same. This can be done by logging into the Monarch iLO 3 MP,
choosing the Command Menu, then entering df to display FRU information. Each processor of each
blade (by enclosure bay number) can be chosen individually and the information displayed will include
the information needed for comparison.
For more information, see Processor and heatsink module on page 43, and CPU and heatsinkmodule on page 123.
DIMM matching and loading rules
The rules regarding DIMM load order change, especially when upgrading from a BL860c i4 to either a
BL870c i4 or BL890c i4. For more information on DIMM loading rules, see DIMM quad load order rules.
NOTE:
Memory loading rules apply across all blades in a sever depending on the server type (BL870c i4 or
BL890c i4) regardless of nPartition configuration.
There are also minimum system memory and performance guidelines must be followed. See the Memory
subsystem information for HPE Integrity Server Blades (BL860c i4, BL870c i4, and BL890c i4) Technical
White Paper (http://www.hpe.com/support/Integrity_i2_Blades_System_Errata).
66Optional components
Page 67
Optional mezzanine cards
The HPE NC532m Dual Port 10GbE BL-c Adapter is restricted to two cards per blade on a BL890c i4.
There are no additional rules regarding the number of cards installed or the order of their installation
across the blades in an upgraded server blade. Ensure all mezzanine cards follow the mezzanine port to
interconnect module mapping detailed in Installing interconnect modules on page 21.
Upgrading the original server
Existing server blade check
IMPORTANT:
Do not remove Blade Links or other components at this time.
Procedure
1. If not already booted, verify that the designated Monarch Server (in each partition) is able to boot itsOS. For information on launching the OS from the UEFI shell, refer to UEFI Front Page on page 30.
2. If the Monarch Server is a HP Virtual Machines (VM) host with active guests or a Virtual Services
Platform (VSP) with active Virtual Partitions (vPars), then either stop those guests or vPars now or
migrate them to a temporary host server. Refer to the HP Integrity VM or vPar Administrator Guides
regarding migration.
3. Backup any important OS data on the Monarch Server and any existing servers, as needed.
4. If the Monarch Server employs Virtual Connect and a partition will grow in size in the upgrade, thenyou may wish to use the Portable Image tool to assist with the upgrade. Refer to Preserving VCassigned MAC addresses in HP-UX by enabling Portable Image on page 82 and if you choose to
use PI, then execute the steps under “Pre-Upgrade Procedure for PI” in that section now.
5. Boot each partition of the Monarch Server, and every other existing blade server to be combined in the
upgrade, to the UEFI shell and make sure no problems are seen.
6. Any existing secondary nPartition (that is, any partition which does not include the server Monarch
blade, those designated as 'B', 'C', or 'D' partitions) which will be carried over into the upgraded server
as secondary nPartitions, must have their UEFI variables backed up now so they can be restored later.
See the nPar Administrator Users Guide for BL870c i4 & BL890c i4 for instructions on how to do this.
The primary nPartition in the Monarch Server (the partition which includes the server Monarch blade,
the one designated as 'A') will not lose UEFI variable data.
7. Shutdown each existing blade server which will be combined into the upgraded server. See Powering
off the server blade on page 25.
8. If the enclosure uses Virtual Connect, then unassign any VC server profiles assigned to enclosure
bays occupied by existing servers and any other enclosure bays which will be occupied by the
upgraded server. See the HPE Virtual Connect for c-Class BladeSystem User Guide for more
information on VCM server profiles
9. If this system is managed using HPE Systems Insight Manager (HPE SIM), then it's recommended
that all nodes associated with the original servers be deleted now. This would include the BLDomain
node, if present. Rediscovery to incorporate the upgraded server will be executed after the upgrade
has been completed.
Adding resources
Procedure
1.If not already in place, put the Monarch Server into its enclosure bay location for the upgrade. The
2.Place any other existing servers and Upgrade Blades to be combined in the upgrade into enclosure
Monarch blade of this server must be in the Monarch blade position of the upgraded server.
bays adjacent to the Monarch Server.
Upgrading the original server67
Page 68
IMPORTANT:
Existing servers other than the Monarch Server and also any Upgrade Blades must be placed
into enclosure bays with higher numbers than those occupied by the Monarch Server.
3.Transfer the Monarch Server label carrier card to the Upgrade Blade Link:
a. Remove the label carrier card from behind the trap door of the Monarch Server Blade Link. See
Blade Link on page 130 for the trap door location. The label carrier card will look like the image
below:
b. Remove the product number portion of the Field Upgrade Only label on the Upgrade Blade Link.
This label will look similar to the image below:
Below is an example showing upgrade Blade Link sticker locations for the BL870c i4.
c. Place the product number portion from the Field Upgrade Only label on top of the old product
number and barcode on the label carrier card from the Monarch Server.
IMPORTANT:
Do not cover the serial number on the label carrier card.
d. Remove the position information label (the left hand label in the above picture) from the front of
the upgrade Blade Link bezel.
Temporarily remove the “Field Upgrade Only” label if it blocks the Blade Link trap door.
68Optional components
Page 69
e. Place the label carrier card behind the trap door of the Upgrade Blade Link, but do not install the
Upgrade Blade link at this time.
f. Reapply the “Field Upgrade Only” label to the upgrade Blade Link, making sure not to cover the
Blade Link handle.
4.Remove the Blade Links from the existing servers that will be included in the upgrade ( Blade Link
for BL870c i4 or BL890c i4 configurations on page 113).
IMPORTANT:
Mark the Monarch Server Blade Link and keep track of it until the upgrade is successfully
completed. Do not reuse the old Blade Links after the upgrade has been completed.
5.Make any necessary blade internal component changes. The following hardware components have
loading rules that can change after upgrading:
•processors
•DIMMs
•optional mezzanine cards
•ICH mezzanine boards
See Rules for server blade internal components on page 66 for more information.
6.Fully insert any unseated blades into the enclosure. Make sure all server blades that will be included
in the upgraded server are now fully seated into the enclosure with no Blade Links installed.
7.Install the upgrade Blade Link by completing steps 1 through 11 in Installing the Blade Link for
BL860c i4, BL870c i4 or BL890c i4 configurations on page 25.
8.Wait a couple of minutes for the blades to conjoin and then enter the show server list
command at the Onboard Administrator command line interface. The output should show a status of
“OK” for each blade in the upgraded server, or “OTHER” if there’s a current VC domain defined on
the enclosure. The Monarch blade should have an iLO 3 IP address assigned.
9.Log into the iLO 3 MP interface and enter the cm command at the Main Menu to reach the Command
Menu prompt.
10. Enter the sr command to view the system revision information. A sample output for a BL890c i4 is
shown below. The BL890c i4 will show four blades composing the server blade, a BL870c i4 will
show two. If the output does not show all of the blades expected, then there is a problem which must
be resolved before proceeding. Reseating the Blade Link or individual blades may correct server
blade conjoining problems.
This is just an example so actual version numbers may be different from what is shown.
11. If the sr output indicates that the firmware is not consistent across all of the blades, which should
only be due to a new Upgrade Blade with different firmware version being added, then complete a
second server firmware update. Update the system to the same firmware version used previously for
the existing servers. An example of sr output indicating a firmware inconsistency is shown below.
This is just an example so actual version numbers may be different from what is shown.
If an update is required, it may need to be completed using the HP SUM Force Install option for both
the iLO 3 and system firmware bundles. HP SUM should detect the mismatch and set the Force
Install option by default. When finished, log back into the iLO 3 MP of the Monarch blade and enter
sr from the Command Menu to verify that the firmware levels on all conjoined blades are now
consistent and up to date.
12. Exchange the System Information Parameters by following the steps below. For more information
regarding the information contained in the System Information Parameters, see
system information parameters on page 75.
a. Log into the Monarch iLO 3 MP, enter the Command Menu, and execute the command sysset.
This will list the system information parameters stored in the Upgrade Blade Link (primary) and
Blade link and
70Optional components
Page 71
those stored on the server blade (secondary). Below is an example of the output when upgrading
from a BL860c i4 to a BL870c i4.
[ilo002264fee1be] CM:hpiLO-> sysset
SYSSET
Current System Information Parameters:
Manufacturer : HP
Product Name (prodname) : Integrity BL870c i2
Secondary Product Name : Integrity BL860c i2
Product Number (prodnum) : AM358A
Secondary Product Number : AD399A
Serial number (serial) : Uninitialized
Secondary Serial Number : USA0003
UUID (uuid) : FFFFFFFF-FFFF-FFFF-FFFF-FFFFFFFFFFFF
Secondary UUID : 89742B3C-AD78-11DF-BB2A-5ABBD7A51B74
Product ID : 0x702
An Upgrade Blade Link has been detected.
Primary and secondary system parameters values do not match.
Copy secondary UUID and serial number to primary? (Y/[N]):
The product name and product number from the Upgrade Blade Link reflect a BL870c i4, while
the Secondary product name and number (from the Monarch Server) reflect the original BL860c
i4.
b. The system detects the Upgrade Blade Link and prompts you to make the required system
information exchanges between Upgrade Blade Link and the server. Enter y to approve the
exchange. You will then be informed that a reset of the iLO 3 MP is necessary for the changes to
take effect.
Copy secondary UUID and serial number to primary? (Y/[N]): y
-> Secondary system parameters values have been copied to primary
system parameters.
->System Information parameters have been updated.
iLO MP needs to be reset in order for these changes to take effect
c. Reset the iLO 3 MP by entering xd –r –nc at the Command Menu prompt.
[ilo002264fee1be] CM:hpiLO-> xd -r -nc
XD -r -nc
-> MP reset requested
MP is now being reset...
-> Command successful.
d. After the iLO 3 MP reset (less than one minute), log back into the iLO 3 MP, go to the Command
Menu, and enter sysset again to verify that the system information parameters have been set for
correctly for both Primary and Secondary. The values between Primary and Secondary should
now match.
[ilo002264fee1be] CM:hpiLO-> sysset
SYSSET
Current System Information Parameters:
Optional components71
Page 72
Manufacturer : HP
Product Name (prodname) : Integrity BL860c i4
Secondary Product Name : Integrity BL860c i4
Product Number (prodnum) : AM377A
Secondary Product Number : AM377A
Serial number (serial) : SGH2101LRY
Secondary serial number : SGH2101LRY
UUID (uuid) : 59F30FC7-8A4A-11E1-AF3F-DC6163A31240
Secondary UUID : 59F30FC7-8A4A-11E1-AF3F-DC6163A31240
Product ID : 0x711
13. Remove the Field Upgrade Only sticker from the bezel of the Upgrade Blade Link. This is no longer a
Field Upgrade Blade Link because the Serial Number and UUID have now been programmed on the
Blade Link.
14. As mentioned earlier, the nPartition configuration of the upgraded server defaults to a single partition.
An example of what might be seen entering the npar command from the iLO 3 MP Command Menu
is shown below for an upgrade to a BL890c i4.
[ilo80c16e9877f4] CM:hpiLO-> npar
NPAR
Current nPartition Configuration:
Bay 1 2 3 4
-------------------------------------------------------------------- nPartition : A A A A
Domain MP : yes - - Power Off : yes - - nPartition Unlocked : - - - nPartition Menu:
Current -> AAAA - One 4-Blade System (default)
AACC - Two 2-Blade nPartitions
AACD - One 2-Blade nPartition, and Two 1-Blade nPartitions
ABCC - Two 1-Blade nPartitions, and One 2-Blade nPartition
ABCD - Four 1-Blade nPartitions
Enter new value, or [Q] to Quit:
15. When a blade is changed from one nPartition to another, the system variable and configuration
information is cleared. This provides a convenient method for ensuring that all Auxiliary blades in the
upgraded server are cleared of stale or inconsistent information. Set the nPartition configuration to
the highest number of partitions possible for the upgraded server. For a BL870c i4 that will be two
(AB) while for a BL890c i4 that will be four (ABCD). The iLO 3 MP will automatically reset for the
change to take effect. An example output is shown below for a BL890c i4. Note that since the
Monarch blade never changes partitions (it always remains in the 'A' partition), that blade will not
have its system environment variables cleared.
New nPartition Configuration (* marks modified systems):
Bay 1 2 3 4
-------------------------------------------------------------------- nPartition : * A * B * C * D
-> nPartition Configuration has been updated.
-> iLO will now reset on all modified systems...
Note: System environment variables may be cleared on all modified
72Optional components
Page 73
systems
for the nPartition configuration to take effect. This configuration
change
may take several minutes to appear on other manageability user
interfaces.
-> iLO reset succeeded.
-> Command successful.
[ilo80c16e9877f4] CM:hpiLO->
16. Wait a minute or so then log back into the iLO 3 MP and use the npar command from the Command
Menu to set the nPartition configuration required for the upgraded server. If no nPartitioning is
required, then set the configuration back to a single, domain-wide partition. See the nParAdministrator Users Guide for BL870c i4 & BL890c i4 for information on configuring nPartitions.
Below is an example of the output which could be seen setting a BL890c i4 back to a single
nPartition.
New nPartition Configuration (* marks modified systems):
Bay 1 2 3 4
--------------------------------------------------------------------nPartition : * A * A * A * A
-> nPartition Configuration has been updated.
-> iLO will now reset on all modified systems...
Note: System environment variables may be cleared on all modified
systems
for the nPartition configuration to take effect. This configuration
change
may take several minutes to appear on other manageability user
interfaces.
-> iLO reset succeeded.
-> Command successful.
[ilo80c16e9877f4] CM:hpiLO->
NOTE:
nPartitions allows some flexibility in timing the incorporation of new blade resources which
allows the original Monarch Server to be brought back up quickly after the upgrade. This is
referred to as "Quick Boot." See The Quick Boot option on page 77 and the HPE Integrityserver blade upgrade kits whitepaper for more information about this option.
17. After a minute or so log back into the iLO 3 MP and enter npar -nc from the Command Menu to
verify that the nPar configuration is now correct.
18. If the enclosure uses Virtual Connect, then the VCM server profiles for the partitions of the upgraded
server should be assigned now. The Monarch Server VCM server profile should be assigned to the
partition which includes the Monarch blade. Profiles which were previously assigned to existing
servers may need some modifications after being assigned to the upgrade so that network
assignments and other elements of the VC server profile match what is required. See Possiblechanges due to VC profile mapping on the upgraded blade server on page 77 for more
information. If the HP-UX Portable Image tool is used (see Preserving VC-assigned MAC
Optional components73
Page 74
Final hardware check
Procedure
1. Log into the Monarch iLO 3 MP of each partition in the upgraded server and power on each partition
2. If UEFI variables need to be restored on secondary partitions, then take care of that now and reboot
3. If there are any embedded I/O device or mezzanine card I/O FW updates needed, then take care of
4. If necessary, switch the onboard SmartArray SAS controller on new upgrade blades to RAID mode
addresses in HP-UX by enabling Portable Image on page 82), then it’s best not to make
changes until after the entire upgrade is completed and any changes then should be made very
carefully. Once assigned, verify that there are no errors indicated by VCM or by the OA. Any VCM
Domain Status or OA System Status alerts or warnings should be resolved before proceeding.
Sometimes, VCM problems can be corrected by unassigning then reassigning the VC server profile.
by entering pc –on at the Command Menu prompt. Let each partition boot to the UEFI shell. Do not
autoboot any currently installed Operating Systems.
the partition to the UEFI shell. See the nPar Administrator Users Guide for BL870c i4 & BL890c i4 for
information on restoring the system environment variables.
those now.
(Configuring a Smart Array Controller on page 133).
NOTE:
All new Upgrade Blades come with the onboard SmartArray SAS controller in HBA mode.
5. Complete the following checks to be sure that the upgraded server blade is functioning properly. Any
problems found in the steps immediately below should be resolved before proceeding.
a. Check the iLO 3 MP SEL for any problems. See Errors and error logs on page 103 for more
information on resolving any errors.
IMPORTANT:
If any errors are found in the SEL, do not continue until resolving them.
b. Execute the UEFI shell command lanaddress and verify that the expected number of LAN ports
are shown and that there is a link indication for every port that is connected to a network or network
switch.
NOTE:
You should see an M on each applicable line.
c. Execute the UEFI shell command map –r to make sure that all of the expected disks are mapped.
d. Execute the UEFI shell command cpuconfig to make sure that all of the expected processors are
there, indicate active, and that CPU threads are set as expected.
e. Execute the UEFI shell command info mem –dimms to make sure that all of the expected DIMMs
are there and indicate active with the correct memory interleave setting.
6. If the upgraded server is to be incorporated into HPE SIM, and HPE SIM has not yet discovered the
upgraded server, then execute rediscovery of the enclosure from SIM now. Ensure that the upgraded
server is correctly incorporated and complete any configuration steps necessary.
Booting the operating system
The server blade has now been physically upgraded and it may be ready to boot to the OS. Certain
upgrade scenarios require OS reinstallation:
Procedure
1. Boot the OS. See Operating system procedures on page 35 for more information.
74Optional components
Page 75
IMPORTANT:
A new Operating System license may be required for the upgraded server. See Operating
System Licenses on page 76 for more information.
2. If the HP-UX Portable Image tool was used, execute the steps under “Post-Upgrade Procedure for PI”
in the Preserving VC-assigned MAC addresses in HP-UX by enabling Portable Image on page
82 section.
3. Verify that the system, especially networking and storage, function as expected. If anything must be
adjusted, then make those adjustments now.
4. If the Monarch Server is an HPVM host server or a Virtual Services Platform (VSP), then the VM
guests or vPars can now be migrated back or started if they were stopped. Refer to the HP Integrity
VM or vPar Administrator Guides document for information regarding guest and vPar migration. Test
each VM guest or vPar to make sure each is functioning properly.
5. The upgrade has been completed. New resources can now be configured and used as needed.
Warranty registration
After completing the upgrade, contact Hewlett Packard Enterprise with the server serial number and new
product number. The warranty of the designated Monarch Server will be transferred to the upgraded
server blade and the service time will be reset. Proof of purchase must be provided and the Hewlett
Packard Enterprise warranty organization must be notified.
The proof of purchase should be a scanned copy or fax of a sales receipt or invoice from Hewlett Packard
Enterprise or an authorized reseller. It can also be an order confirmation, shipping receipt, or reseller
based purchase/shipping confirmation that contains the following:
•The purchase or shipping date.
•The Monarch Server Product Number or Product Name/product description.
•The Serial Number, but it is not required (some invoices do not have serial numbers).
•The vendor’s name (the name of the business where the unit was purchased).
IMPORTANT:
Be sure that all information is clearly readable.
More information regarding server blade warranties can be found at: http://www.hpe.com/info/Blades-
docs.
Support
A support Flex Care Pack with at least the same level of care as the server blade to be upgraded (the
Monarch Server) should have been purchased when the Upgrade Kit was ordered. Your Hewlett Packard
Enterprise sales representative will apply the support package to the upgraded server blade and provide
credit for the unused portion of the package purchased for the original Monarch Server. The new Flex
Care Pack must be registered with the serial number of the original Monarch Server (now attached to the
upgraded server) included to ensure support entitlement. For more information, go to http://cpc.hpe.com/. You can also contact your Hewlett Packard Enterprise sales representative or contact
Hewlett Packard Enterprise sales through the following link: http://www.hpe.com/contact.
Blade link and system information parameters
Every BL860c i4, BL870c i4 and BL890c i4 server blade has a set of information called the System
Information Parameters which helps to identify the server blade and is used for certain applications and
for the server blade warranty. This information includes:
•product name
•product number
Support75
Page 76
•UUID
•serial number
The System Information Parameters are stored in nonvolatile memory onboard the Blade Link and server
blades. The System Information Parameters can be viewed from the iLO 3 MP interface Command Menu
prompt by entering sysset.
[ilo002264fed1da] CM:hpiLO-> sysset
SYSSET
Current System Information Parameters:
Manufacturer : HP
product name (prodname) : Integrity BL860c i4
Secondary Product Name : Integrity BL860c i4
Product Number (prodnum) : AM377A
Secondary Product Number : AM377A
Serial number (serial) : SGH2101LRY
Secondary serial number : SGH2101LRY
UUID (uuid) : 59F30FC7-8A4A-11E1-AF3F-DC6163A31240
Secondary UUID : 59F30FC7-8A4A-11E1-AF3F-DC6163A31240
Product ID : 0x711
There are two sets of parameters shown:
•Primary values are stored on the Blade Link.
•Secondary values are those stored on the server blades.
When an existing blade server is upgraded, the system must retain the existing UUID and Serial Number,
but the Product Name and Product Number will need to be changed to reflect the upgraded server type.
The Upgrade Blade Link will come already installed with a Product Name and Number which reflects the
upgrade configuration. This information will need to be copied to the server blades, and the UUID and
Serial Number of the server blades will need to be copied to the Upgrade Blade Link. This exchange of
System Information Parameters is executed by firmware but it's started by the user entering the sysset
command.
The server UUID, serial number, and product number are printed on the label carrier card behind the
Blade Link trapdoor which is described in
ship with a label carrier card, but has a trap door for one. It also has a Field Upgrade Only label on the
front bezel which contains a peel-off sticker with the new Product Number.
Operating System Licenses
HP-UX requires a License to Use (LTU) for each processor (socket) installed in a server and HP-UX uses
tiered licensing levels based on the maximum processor socket count of the server it runs on. The tier
level will rise when the upgraded server contains more sockets than the original Monarch Server. The
“tiers” map to the blade servers as follows:
•BL860c i4 – 2 Socket Tier
•BL870c i4 – 4 Socket Tier
•BL890c i4 – 8 Socket Tier
LTUs must be purchased for each processor (socket) installed in the upgraded server at the tier level of
the upgraded server type for the installed HP-UX OE. Your Hewlett Packard Enterprise sales
representative must be contacted to get these new licenses. Full credit will be received when the original
Monarch Server licenses are surrendered. The credit will reflect the current purchase price of the original
licenses, not the original purchase price.
Blade Link on page 130. The Upgrade Blade Link does not
For more information contact your Hewlett Packard Enterprise sales representative. Hewlett Packard
Enterprise sales can also be contacted at http://www.hpe.com/contact.
76 Operating System Licenses
Page 77
The HP-UX license is delivered physically or electronically by certificate. To contact your Hewlett Packard
Enterprise sales representative for more information about software licensing agreements, go to http://
www.hpe.com/contact.
The Quick Boot option
The ability to scale up a blade server by adding more blade level resources is inherent to the BL8x0c
architecture. This ability to scale up when needed provides flexibility and control over capital expenditures
for your compute environment. The introduction of nPartition capability in the BL8x0c i4 blade server adds
another level of flexibility. New blade resources can be added to a blade server and then the original
server can be quickly booted within its own partition with new resources allocated to a separate partition.
The new resources can be configured and used separately or they can be added to the original server
partition as time allows or need requires. This provides flexibility in managing system downtime.
For example, a BL870c i4 original server can be upgraded to a BL890c i4 by adding two Upgrade Blades.
Let's say that the down-time of the BL870c i4 needs to be minimized with the new blade resources
intended for a separate partition or to be incorporated into the original server at a later time. The BL870c
i4 can be brought down, physically upgraded by the addition of two Upgrade Blades and Upgrade Blade
Link, and then partitioned in such a way that the two blades of the original server are placed into an AA
partition while the two new blades are placed into a CC partition. This is shown in the figure below.
The original server blades in the AA partition can be rebooted immediately functioning just as they did
before the upgrade. With the AA partition up and running, the new resources in the CC partition can be
configured as an independent server partition without impact to the AA partition. Or the new blade
resources can be incorporated into the AA partition at a later date when the need arises or time allows for
it.
Possible changes due to VC profile mapping on the upgraded blade server
Virtual Connect Manager supports assigning a VC server profile to each bladed server. Profiles can
include VC-assigned:
•UUID
•serial number
•Ethernet port MAC addresses with network assignments
•FC HBA connection WWNs with FC SAN assignments.
•FCoE connection WWNs with FC SAN assignments
The Quick Boot option77
Page 78
The VC-assigned MAC addresses, FC WWNs, FCoE WWNs, UUID and Serial Number can come from
blocks of values supplied by VCM or from user-defined blocks of values. Alternatively, the factoryassigned default values for all of these can be used. The following information pertains to all of these
options.
When performing an upgrade it is critical to understand how the server profile connection entries will be
mapped to ports on the upgraded server as this may impact the OS. What happens for each of the three
I/O assignment categories (Ethernet network assignments, FCoE SAN assignments, and FC SAN
assignments) is different for each category so each is described separately below. Stated briefly, the list
of network and SAN assignments in the original server profile can map differently to upgraded server
ports due to the increased number of blades. The discussion which follows contains a lot of detail, but the
best way to check for and fix any problems with the server profile applied to the upgraded server may be
to open the profile in VCM after it's assigned, see how the connection entries mapped to the upgraded
server, and make adjustments if necessary. The VC server profile should always be checked after being
applied to the upgrade to make sure that profile connection entries are mapped to ports as expected and
to make sure there are no profile errors or warnings which must be resolved. See the HPE VirtualConnect for c-Class BladeSystem User Guide for more information on VCM server profiles. The only
caution is that if you use the PI tool (more on that follows), then you must let HP-UX boot with the profile
untouched and disable PI before making changes to the profile.
FC SAN connections
The FC SAN connection entries in the VCM server profile should map to the same ports whether
assigned to the original or the upgraded server. VCM will map entries first to all ports on blade 1, then
blade 2, etc. So no adjustments to the FC SAN entries in the profile should be necessary.
The FC SAN connection entries in a profile should be thought of as sets of entries. If there are N Virtual
Connect FC Interconnect modules in an enclosure (where N is rounded up to an even number) then there
are N FC SAN connection entries per set. The first N entries are mapped to the FC HBA ports on the
Monarch blade, the second set of N entries are mapped to the FC HBA ports on the first Auxiliary blade,
the next set of N entries to the next Auxiliary blade, until all of the FC connection entries have been
mapped to ports or until there are no more FC HBA ports available to map entries to. Below are a few
examples of upgrading a BL860c i4 to a BL870c i4 in an enclosure with four Virtual Connect FC
Interconnect modules where the BL860c i4 becomes the Monarch blade in the upgraded server.
•If the profile assigned to the BL860c i4 server had four or fewer FC SAN connection entries then those
entries are still mapped to the FC HBA ports on the original BL860c i4 blade (since it is the Monarch in
the new server). There will be no FC SAN connection entries mapped to any ports on the Auxiliary
blade in the upgraded server.
•If the profile assigned to the BL860c i4 server had more than four FC SAN connection entries then the
first four entries are still mapped to the FC HBA ports on the original BL860c i4 blade (since it is the
Monarch in the new server). The additional FC SAN connection entries (up to four), that had not been
mapped to any ports prior to the upgrade, will be mapped to FC HBA ports on the Auxiliary blade in
the upgraded server.
FCoE SAN connections
The mapping of FCoE connection entries is much the same as the FC connection entry mapping, but with
a small twist. Each blade in the BL8x0c i4 family can have up to four embedded FCoE ports active which
would appear in the VCM server profile mapping as LOM:1-b, LOM:2-b, LOM:3-b, and LOM:4-b. VCM will
map FCoE connection entries from a profile to the Monarch blade LOM1 and LOM2 first, then to any
FCoE mezzanine cards on the Monarch. If there are still entries in the profile, they will be mapped to the
first Auxiliary blade LOM1 and LOM2 followed by any FCoE mezzanine card ports on that blade. After all
blades have had connection entries mapped to LOM1, LOM2, and the FCoE mezzanine cards, if there
are still entries left in the profile, they will be mapped to the Monarch blade LOM3 and LOM4, then to the
first Auxiliary blade LOM3 and LOM4, and etc. until all FCoE connection entries have been mapped or the
ports in the server have been exhausted. The twist, then, is only in the mapping of FCoE connection
entries to LOM3 and LOM4 ports. If a profile assigned to the original server has no connection entries
mapped to LOM3 and LOM4, then the FCoE entries will remain mapped to the same ports on the same
78Optional components
Page 79
blades after the upgrade. However, if there are FCoE connection entries mapped to LOM3 and LOM4
then those entries will get mapped to LOM1 and LOM2 ports on the new blades in the upgraded server.
Below is an example upgrading a BL860c i4 to a BL870c i4 where VC-assigned WWNs are used and the
VCM server profile contains five FCoE connection entries. Notice that the first four entries map to the
same ports on both the original and the upgraded server, but the fifth entry, mapped to LOM3 on the
original server, is mapped to LOM1 of the Auxiliary blade in the upgrade.
Profile entryAssigned SANVCM assigned
WWN
1SAN-150:06:0B:
00:00:C3:26:10
2SAN-250:06:0B:
00:00:C3:26:12
3SAN-350:06:0B:
00:00:C3:26:14
4SAN-450:06:0B:
00:00:C3:26:16
5SAN-550:06:0B:
00:00:C3:26:18
The WWN is assigned in this example by VCM (as opposed to using HW default WWNs) so the WWN
does not change, but the HW path as seen by HP-UX is different so this would be seen as a new device.
If the OS boot disk were attached on SAN-5, then the boot path in UEFI Boot Manager would no longer
be valid. If it's okay that the HW path has changed, then HP-UX could be configured to use the new HW
path and a new HP-UX boot path entry could be made in UEFI Boot Manager, if necessary. If the HW
path must remain the same, then four connection entries can be added to the profile before assigning the
profile to the upgraded server. This is shown below.
Mapping to
BL860c i4 ports
Monarch LOM:1-bMonarch LOM:1-b
Monarch LOM:2-bMonarch LOM:2-b
Monarch MEZZ1:1Monarch MEZZ1:1
Monarch MEZZ1:2Monarch MEZZ1:2
Monarch LOM:3-bAuxiliary LOM:1-b
Mapping to
BL870c i4 ports
Profile entryAssigned SANVCM assigned WWNMapping to BL870c i4
ports
1SAN-150:06:0B:
00:00:C3:26:10
2SAN-250:06:0B:
00:00:C3:26:12
3SAN-350:06:0B:
00:00:C3:26:14
4SAN-450:06:0B:
00:00:C3:26:16
5unassigned50:06:0B:
00:00:C3:26:18
6unassigned50:06:0B:
00:00:C3:26:1A
Monarch LOM:1-b
Monarch LOM:2-b
Monarch MEZZ1:1-b
Monarch MEZZ1:2-b
Auxiliary LOM:1-b
Auxiliary LOM:2-b
Table Continued
Optional components79
Page 80
Profile entryAssigned SANVCM assigned WWNMapping to BL870c i4
ports
7unassigned50:06:0B:
00:00:C3:26:1C
8unassigned50:06:0B:
00:00:C3:26:1E
9SAN-550:06:0B:
00:00:C3:26:20
The Monarch blade LOM3 now has the correct SAN assignment so the HW path from the original server
has been restored, but the WWN is changed so a change to the SAN infrastructure may be necessary so
that the required storage devices associate with the new WWN. Also note that this enables FCoE on port
'b' of the Auxiliary blade LOM and MEZZ FlexFabric devices leaving only three ports on each device
available for Ethernet connections. LOM3 and LOM4 belong to the same physical LOM device on the
blade so because both 'b' ports on a LOM device must be configured with the same protocol, the Monarch
LOM:4-b will be enabled as an FCoE port and not available as an Ethernet port since LOM:3-b has an
FCoE entry mapped to it. The Auxiliary blade LOM3 and LOM4 'b' ports, however, remain available as
Ethernet ports because neither of those ports have an FCoE connection entry mapped to them.
Auxiliary MEZZ1:1-b
Auxiliary MEZZ1:2-b
Monarch LOM:3-b
Ethernet network connections
The Ethernet connection entries in a profile are handled differently from SAN entries in order to balance
the use of Flex-10 ports (even if no Flex-10 ports are used, the same balanced approach is used). Virtual
Connect manager distributes the Ethernet connection entries across all of the blades in a server, and on
each blade it distributes connection entries across all of the Ethernet physical ports. This is done so that
as few Flex-10 sub-ports are used as needed, which results in maximum band-width available to each
Flex-10 sub-port. This distribution of Ethernet connection entries across all Ethernet ports is done for
single-blade servers, too. The net result is that the mapping of Ethernet connection entries in a profile
assigned to a server will likely change when that server is upgraded. Consider the following examples.
•If the upgrade is from a BL860c i4 to a BL870c i4 then every other Ethernet connection entry starting
with the first entry (for example, the first, third, fifth, etc. entries) will be mapped to ports on the
Monarch blade, and every other entry starting with the second entry (for example, the second, fourth,
sixth, etc. entries) will be mapped to ports on the Auxiliary blade.
•If the upgrade is from a BL860c i4 to a BL890c i4, then every fourth Ethernet connection entry starting
with the first entry (for example, the first, fifth, ninth, etc. entries) will be mapped to ports on the
Monarch blade; every fourth entry starting with the second entry (for example, the second, sixth, tenth,
etc. entries) will be mapped to the first Auxiliary blade; every fourth entry starting with the third entry
(for example, the third, seventh, eleventh, etc. entries) will be mapped to the second Auxiliary blade;
and finally every fourth entry starting with the fourth entry will be mapped to ports on the 3rd Auxiliary
blade.
As an example, consider a profile with 8 Ethernet connection entries where each entry defines a
connection to a different network (this makes it easy to see what happens when an upgrade is done).
Assume that the profile is assigned to a BL860c i4 that is upgraded to a BL870c i4. Assume that VC
Flex-10 Interconnect modules exist in bays 1 and 2, thus the FlexFabric LOM ports on the blades are
utilized as Flex-10 ports. The following table shows how these entries are mapped to ports on the original
server and to the blades in the upgraded server. The examples here assume that the LOM FlexFabric
devices are not in FCoE mode. That is, there are no FCoE connection entries in the VCM server profile.
If it is important that each of the networks be mapped to the same ports on the Monarch blade after the
upgrade, then the profile will need to be edited to add seven Ethernet connection entries, and then the
entries must be edited to obtain the desired network connections aligned with the ports on the Monarch
blade. The resulting set of Ethernet connection entries in the profile would be as follows.
Mapping to
BL860c i4 ports
Mapping to
BL870c i4 ports
Profile entryAssigned networkVCM assigned MAC
address
1LAN-100-17-A4-77-90-10Monarch LOM1–a
2unassigned00-17-A4-77-90-12Auxiliary LOM1–a
3LAN-200-17-A4-77-90-14Monarch LOM2–a
4unassigned00-17-A4-77-90-16Auxiliary LOM2–a
5LAN-300-17-A4-77-90-18Monarch LOM3–a
6unassigned00-17-A4-77-90-1AAuxiliary LOM3–a
7LAN-400-17-A4-77-90-1CMonarch LOM4–a
8unassigned00-17-A4-77-90-1EAuxiliary LOM4–a
9LAN-500-17-A4-77-90-20Monarch LOM1–b
10unassigned00-17-A4-77-90-22Auxiliary LOM1–b
11LAN-600-17-A4-77-90-24Monarch LOM2–b
Mapping to BL870c i4
ports
12unassigned00-17-A4-77-90-26Auxiliary LOM2–b
13LAN-700-17-A4-77-90-28Monarch LOM3–b
Table Continued
Optional components81
Page 82
Profile entryAssigned networkVCM assigned MAC
address
14unassigned00-17-A4-77-90-2AAuxiliary LOM3–b
15LAN-800-17-A4-77-90-2CMonarch LOM4–b
NOTE:
While the network mapping has been adjusted to preserve the original mapping of networks to
ports, it is not possible to adjust the MAC address assignment to preserve the same mapping of
MAC addresses to ports. If your OS is HP-UX, see the section Preserving VC-assigned MACaddresses in HP-UX by enabling Portable Image on page 82 for a method which can be used
to preserve VC-assigned MAC addresses.
Mapping to BL870c i4
ports
Preserving VC-assigned MAC addresses in HP-UX by enabling Portable Image
If your OS is HP-UX where the VC server profile does not use the hardware default MAC addresses and it
is important that Ethernet connections configured in the OS retain their VC-assigned MAC addresses,
then there is a package called “Portable Image” (PI) which can be used. The HP-UX Portable Image
product is installed and enabled on the blade server before the upgrade. After the upgrade, PI will
automatically modify the HP-UX network instance numbers of the new hardware environment so that they
match the pre-upgrade environment. This takes advantage of the VCM interconnect stacking links to
allow all of the Monarch Server network connections to function correctly with no corrections needed in
the original VC server profile.
If you choose to use PI to assist the upgrade, there will be a pre-upgrade step to follow on the designated
Monarch Server and then another step to follow after the upgrade. These two steps are detailed below
but they must not be executed until they’re explicitly called for in the upgrade procedure steps.
Pre-Upgrade Procedure for PI
These instructions are to be executed on the designated Monarch Server with the OS booted.
1. You need to obtain the PI package if it’s not already installed.
a. Check for it being already installed by executing the following command:
swlist|grep HPPortableImage
b. If nothing is returned, then the package must be installed now. Go to http://www.hpe.com/
support/softwaredepot and search for “Portable Image”. Follow the link for HP-UX Portable
Image in the product list and follow the instructions there for installing the depot.
2. Clear out any old PI data by executing the following command: /opt/network/bin/hpuxpitool -
r all
3. Enable PI by executing the following command: /opt/network/bin/hpuxpitool -e
You should see the following:
* Future operations will ask whether to update the backup.
* The requested changes have been applied to the currently
running configuration.
Tunable Value Expression Changes
gio_portable_image (before) 0 Default Immed
(now) 1 1
4. The PI state must be saved before starting the upgrade. This is done by explicitly executing the PI
save command:
/opt/network/bin/hpuxpitool -s
82 Preserving VC-assigned MAC addresses in HP-UX by enabling Portable Image
Page 83
5. Dump the saved PI state to a text file for possible use later. Obtain the dump by executing the
following command:
/opt/network/bin/hpuxpitool –g
This will list out each network instance along with its pre-upgrade MAC address.
6. Shutdown the OS using a shutdown command.
Post-Upgrade Procedure for PI
1. Dump the new PI state to a file for possible use later. Obtain the dump by executing the following
command:
/opt/network/bin/hpuxpitool –g
This will list out each network instance along with its post-upgrade MAC address.
2. Disable PI by executing the following command:
/opt/network/bin/hpuxpitool -d
You should see the following:
* Future operations will ask whether to update the backup.
* The requested changes have been applied to the currently
running configuration.
Tunable Value Expression Changes
gio_portable_image (before) 1 1 Immed
(now) 0 0
3. The networking should be checked to make sure that the original system network instances function
correctly. If the HP-UX boot reports a LAN interface configuration failure, then check the /etc/rc.log file
for the failed instance number which may indicate something like the following example:
ERROR: lan18 interface: ifconfig: no such interface
"/sbin/rc2.d/S340net start" FAILED
Get the MAC address for the failed instance number from the pre-upgrade PI dump and look for that
MAC address in the post-upgrade PI state dump. From that entry, note the new instance number and
edit the HP-UX networking configuration for the failed instance number so that it uses the new
instance number.
Optional components83
Page 84
Troubleshooting
Cause
This chapter provides strategies, procedures, and tools for troubleshooting server blade error and fault
conditions.
Methodology
General troubleshooting methodology
Procedure
1. Review the following list of symptoms:
a. Front Panel LED blinking
b. System messages on the OA
c. System Alert present on system console
d. Server blade will not power-up
e. Server blade will not boot
f. Error/Event Message received
g. MCA occurred
2. Narrow down the observed issue to the specific troubleshooting procedure required. Isolate the failure
to a specific part of the server blade to perform more detailed troubleshooting. For example:
a. Issue - Front Panel LED blinking
NOTE:
The front panel health LED flashes amber with a warning indication, or flashes red with a fault
indication.
•Look for a system alert on the OA or system console.
•Analyze the alert by using the SEL, to identify the last error logged by the server blade. Use the
iLO 3 MP commands to view the SEL, through the MP menu interface.
MP MAIN MENU:
CO: Console
VFP: Virtual Front Panel
CM: Command Menu
CL: Console Log
SL: Show Event Logs
HE: Main Help Menu
DB: Debug Shell
X: Exit Connection
[ilo3cd92b2fca24] MFG! hpiLO-> sl
Event Log Viewer Menu:
Log Name Entries % Full Latest Timestamped Entry
--------E - System Event 489 18 % 03 Jan 2001 00:11:17
F - Forward Progress 13251 68 % 03 Jan 2001 00:11:17
I - iLO Event 136 27 % 03 Jan 2001 03:30:04
C - Clear All Logs
84 Troubleshooting
Page 85
L - Live Events
Enter menu item or [Ctrl-B] to Quit: e
Log Name Entries % Full Latest Timestamped Entry
--------E - System Event 489 18 % 03 Jan 2001 00:11:17
Event Log Navigation Help:
+ View next block (forward in time, e.g. from 3 to 4)
- View previous block (backward in time, e.g. from 3 to 2)
<CR> Continue to the next or previous block
D Dump the entire log
F First entry
L Last entry
J Jump to entry number
H View mode configuration - Hex
K View mode configuration - Keyword
T View mode configuration - Text
S View mode configuration - Cause/Action
A Alert Level Filter options
U Alert Level Unfiltered
? Display this Help menu
Q Quit and return to the Event Log Viewer Menu
Ctrl-B Exit command, and return to the MP Main Menu
# Location |Alert | Encoded Field | Data Field | Keyword/Timestamp
--------480 SFW 1,0,2,0 *7 F480003709E1039B 000000000000000F BOOT_HALT_CELL
02 Jan 2001 22:29:12
479 SFW 1,0,2,0 *7 E48000D109E10399 FFFFFF01FFFFFF94 MEM_NO_MEM_FOUND
02 Jan 2001 22:29:12
478 SFW 1,0,2,0 2 40801FB109E10397 0000000000000000
MEM_NON_OPTIMAL_CONFIG
02 Jan 2001 22:29:12
477 SFW 1,0,2,0 *5 BA801F9D09E10395 FFFFFF0100100B74
MEM_DIMM_INFO_HW_DEALLOCATED
02 Jan 2001 22:29:11
476 SFW 1,0,2,0 *5 BA801F9D09E10393 FFFFFF0100090B74
MEM_DIMM_INFO_HW_DEALLOCATED
02 Jan 2001 22:29:11
475 SFW 1,0,2,0 *3 688024FE09E10391 000000000100000F MEM_DDR_TRN_ERR
02 Jan 2001 22:29:11
474 SFW 1,0,2,0 *3 688024FE09E1038F 0000000001010100 MEM_DDR_TRN_ERR
02 Jan 2001 22:29:11
MFG! SL:hpiLO (+,-,<CR>CR>,D,F,L,J,H,K,T,S,A,U,?,Q or Ctrl-B to Quit)>
3. You should have a good idea about which area of the server blade requires further analysis. For
example, if the symptom was “server blade will not power-up”, the initial troubleshooting procedure
might have indicated a issue with the DC power rail not coming up after the power was turned on.
4. Your goal is to identify the failed FRU and replace it. You must now perform the specific removal and
replacement procedure, and verification steps, see
109 for more details.
Removing and replacing components on page
NOTE:
If multiple FRUs are identified as part of the solution, fix all identified failed FRUs.
5. You might have to perform specific recovery procedures to finish the repair.
If a failure occurs the front panel LEDs and the SEL helps you identify the issue or FRU:
•The front panel LEDs and LAN LEDs of the server blade change color and blink to help identify
specific issues, and display LAN activity.
Troubleshooting85
Page 86
For information on LED locations and states, see Front panel LEDs on page 94.
•The SEL provides detailed information about the errors identified by the LEDs.
For server alerts of levels 3-5, the attention condition on the server LED can only be cleared by cycling
DC power.
If the LEDs and SEL do not give you enough information for you to identify the issue you are
experiencing, Hewlett Packard Enterprise also provides diagnostic tools with each operating system (see
Troubleshooting tools on page 92 for more details).
NOTE:
Always examine the iLO 3 MP SEL in the case of a blinking yellow or red front panel health LED,
before replacing any hardware.
Executing recommended troubleshooting methodology
The recommended methodology for troubleshooting a server blade error or fault is as follows:
Procedure
1. Consult the system console for any messages, emails, and so on, pertaining to a server blade error or
fault.
2. View the front panel LEDs (power and health), locally or remotely through the iLO 3 MP vfp
command.
3. Read the symptom/condition information in the left column of
4. Perform the actions specified in the Action column.
Table 10.
For more details, see the appropriate subsection of this chapter, where this information is provided in the
Action column. The Action you are directed to perform might be to access and read one or more error
logs (the event log and/or the FPL).
You can follow the recommended troubleshooting methodology, and use Table 10 and Table 11 or go
directly to the subsection of this chapter which corresponds to your own entry point. Table 9 provides the
corresponding subsection or location title for these different entry points (for example, to start by
examining the logs, go directly to Errors and error logs on page 103).
Table 9: Troubleshooting Entry Points
Entry PointSubsection or Location
Front panel LEDsSee Basic and advanced troubleshooting tables on page 87,
Troubleshooting tools on page 92, and Front panel LEDs on
page 94.
Virtual Front Panel LEDs in the
iLO 3 TUI
SEL and FPLsSee Errors and error logs on page 103.
Offline and Online DiagnosticsSee Troubleshooting tools on page 92.
System Event AnalyzerSee http://www.hpe.com/support/WEBES_Overview for more
See Virtual Front Panel LEDs in the iLO 3 TUI on page 98.
Use the following troubleshooting tables to determine the symptoms or condition of a suspect server
blade. The state of the front panel LEDs can be viewed locally.
NOTE:
Virtual front panel LEDs in the iLO 3 TUI are not the same as the front panel LEDs, see Virtual
Front Panel LEDs in the iLO 3 TUI on page 98 for more information about Virtual front panel
LED troubleshooting.
Table 10: Basic Low End Troubleshooting
StepConditionAction
1Server blade appears “dead” -- no front
panel LEDs are on, and iLO 3 heartbeat
isn ot blinking green.
2Server blade does not power on after front
panel power button is momentarily
pressed (less than four seconds).
Nothing is logged for this condition.
1. For new server installations, review the
installation procedures.
2. Be sure that the enclosure power cords are
connected to both the power supplies and to the
AC receptacles.
3. Be sure that AC power, at the proper AC
voltage levels, is available to the receptacles.
4. If the power button’s integrated LED on front
panel remains off, then reseat the server blade.
5. As a last resort, replace the server blade. This
issue is fixed when the front panel LED states are
as follows: blade health is steady green.
A fatal fault has been detected and logged,
attempting to power on the server.
1. Examine enclosure power supply LEDs. If they
are not steady green, then replace power supply.
2. If the enclosure power supply LED is green,
then you might need an additional power supply
to supply sufficient power to run the blades in the
enclosure.
3. Examine the iLO 3 MP subsystem logs for
events related to DC power rails.
4. An improper SBL or e-keying error can also
cause power on to fail.
This issue is fixed when the front panel LEDs are
as follows: Health is steady green and power is
steady green.
Table Continued
Basic and advanced troubleshooting tables87
Page 88
StepConditionAction
3aBlade health LED is off and iLO 3 LED is
flashing amber or off.
3bBlade health LED is flashing amber or red
and iLO 3 MP is running.
4aCannot see iLO 3 MP prompt on system
console -- blade server power is on. iLO 3
MP heartbeat LED is flashing green.
A fatal fault has been detected and logged while
booting or running system firmware (SFW or iLO
3 firmware).
1. Cannot access the iLO 3 MP at this time.
2. Must reseat or replace the server blade. This
issue is fixed when the iLO 3 MP logs can be
read and both front panel health LED and server
power LED states show: Flashing green or steady
green, and steady green, respectively.
A fatal error has been detected and logged while
booting or running system firmware. Examine the
iLO 3 MP logs for events related to switched DC
power or cooling fans or configuration. This issue
is fixed when both front panel health LED and
server power LED states show: Flashing green or
steady green, and steady green, respectively.
Nothing can be logged for this condition. The
blade health LED state indicates that the server
blade is either booting or running system FW, or
booting or running OS.
1. Check the MP LAN connection. If the
connection is functioning, check the serial
console connection.
4bCannot see UEFI menu on system
console. iLO 3 MP is running.
2. Look for loose, damaged, or disconnected
signal cables between the system console
device, and serial port connector on the front
panel.
3. Be sure that the RS232C configuration
matches between the server blade and the local
console (see Troubleshooting the serverinterface (system console) on page 107 for
more details).
4. As a last resort, replace the server blade. This
issue is fixed when the iLO 3 MP menu appears
on the system console.
1. Examine the iLO 3 MP logs for entries related
to processors, processor power modules, and
shared memory, and core I/O devices (see
Errors and error logs on page 103 for more
details).
2. Check the console POST message to use if
any unusual messages are reported.
3. As a last resort, replace the server blade. This
issue is fixed when the UEFI menu appears on
the system console.
88Troubleshooting
Table Continued
Page 89
StepConditionAction
4cCannot find a boot disk. The iLO 3 MP is
running.
5Cannot see OS prompt on system
console. iLO 3 MP is running.
Nothing can be logged for this condition.
1. Check the disk drive LED.
2. Examine the Smart Array POST message.
3. If FC storage is used for boot, check the FC
card status with the offline utility (drvcfg).
4. Reinsert the boot disk into the drive bay.
5. Search for the boot disk path using the UEFI
shell (map-r) command.
6. Examine the iLO 3 MP logs for entries related
to processors, processor power modules, and
shared memory, and core I/O devices (see
Errors and error logs on page 103 for more
details).
7. As a last resort, replace the server blade. This
issue is fixed when all boot paths are found.
Nothing can be logged for this condition.
1. Examine the console message.
2. Verify the console configuration with the UEFI
conconfig command.
3. Examine the iLO 3 MP logs for entries related
to I/O devices (see Errors and error logs on
page 103 for more details). This issue is fixed
when the OS prompt appears on the system
console.
NOTE:
All events listed below are logged in both FPL and SEL unless otherwise noted. An event of alert
level 2 or greater will be logged in both places, alert level 1 or 0 is logged in FPL only.
6a"POWER_FAIL_NO_SBL" - Power
command failed because no Blade Link is
installed.
6b"SBL_REMOVED" - Blade Link was
removed
6c"ILO_RST_REASON_SBL" - The iLO 3
was reset because the Blade Link was
installed (FPL only).
1. Be sure that the Blade Link is properly
installed and fully seated.
2. Look for bent connector pins on the Blade Link
1. If the Blade Link was not removed, be sure
that is properly installed and fully seated.
2. Look for bent connector pins on the Blade
Link.
1. Be sure that the Blade Link is properly
installed and fully seated.
2. Look for bent connector pins on the Blade
Link.
Table Continued
Troubleshooting89
Page 90
StepConditionAction
6d"SYSVARS_MISMATCH" - The system
variable stored on the Blade Link does not
match the system variable stored on the
blade.
6e"SBL_DOMAIN_IMPROPER_SBL" - The
Blade Link currently installed is not
appropriate for the blade slot (even/odd)
and enclosure (c7000/c3000) being used.
6f"POWER_FAIL_IMPROPER_SBL" - Power
command failed because the Blade Link
currently installed is not appropriate for
the blade slot (even/odd) and enclosure
(c7000/c3000) being used.
Use the sysset command.
1. Be sure that the Blade Link installed in the
correct slot and enclosure, see Installing the
Blade Link for BL860c i4, BL870c i4 or
BL890c i4 configurations on page 25.
2. Be sure that the Blade Link is properly
installed and fully seated.
3. Look for bent connector pins on the Blade
Link.
1. Be sure that the Blade Link installed in the
correct slot and enclosure, see Installing the
Blade Link for BL860c i4, BL870c i4 or
BL890c i4 configurations on page 25.
2. Be sure that the Blade Link is properly
installed and fully seated.
3. Look for bent connector pins on the Blade
Link.
6gIncorrect number of processors displayed
for the conjoined blades
6h
"QPI_CPU_LINK_DEGRADED"
1. Be sure that the Blade Link is properly
installed and fully seated.
2. Look for bent connector pins on the Blade
Link.
3. Check the MP logs for CPUs that were deconfigured by system firmware. The UEFI
cpuconfig command can be used to show
and re-configure hardware or software deconfigured CPUs.
Be sure that the Blade Link is properly installed
and fully seated.
90Troubleshooting
Page 91
Table 11: Advanced Low End Troubleshooting
StepSymptom/ConditionAction
6Cannot read SEL from the
system console.
7OS is non-responsive (hung) Front panel LEDs indicate that the server blade power is
SEL logging has stopped (health is steady green and power
is steady green).
Examine console messages for any UEFI errors or warnings
about operation or communications.
This issue is fixed when the SEL resumes logging.
turned on, and it is either booting or running the OS (for
example, health is steady green and power is steady green).
Nothing can be logged for this condition.
1. Use the iLO 3 MP Command Menu to initial a ToC, using
the tc command.
2. Reboot the OS and escalate.
3. Use the following commands to obtain the system
hardware status by capturing the logs.
•sl -e -nc
•l -f -nc
•df -d
•ps
•ss
•sr
•sysset
•npar
•UEFI info all
•VFP
4. Examine the iLO 3 MP logs for entries related to
processors, processor power modules, shared memory,
and core I/O devices (see
103 for more details).
Errors and error logs on page
This issue is fixed when the root cause is determined.
Table Continued
Troubleshooting91
Page 92
StepSymptom/ConditionAction
8aMCA occurs during server
blade operation. The server
blade reboots the OS.
NOTE:
The server blade
reboots OS if enabled.
8bMCA occurs during server
blade operation and it is
repeated; server blade
reboot of OS is prevented.
NOTE:
The troubleshooting
actions for this step
are identical to those
in Step 8a, except that
the server blade in this
step must be hard
reset to begin the
booting process.
Front panel LEDs indicate that the server blade detected a
fatal error that it cannot recover from through OS recovery
routines (for example, health is flashing red and power is
steady green).
1. Capture the MCA dump with the UEFI command,
errdumpmca. If the server blade can boot the OS, you
can capture binary MCA dump files online.
2. Examine the iLO 3 MP logs for entries related to
processors, processor power modules, shared memory,
and core I/O devices (See Errors and error logs on page
103 for more details).
The issue is fixed when the MCA does not repeat or the
source of the MCA has been eliminated.
Front panel LEDs indicate that the server blade detected a
fatal, front side bus error.
System firmware is running to gather and log all error data for
this MCA event.
1. Examine the iLO 3 MP logs for entries related to
processors, processor power modules, shared memory,
and core I/O devices (see Errors and error logs on page
103 for more details).
If the system repeats the MCA, it is unable to capture
MCA dumps since the system is unable to boot to UEFI
shell. Reduce the system configuration and try the boot
with the minimum configuration. For the i4 server blades,
you may need to manage the nPar configuration for the
minimum configuration.
2. Try dc power off/on.
3. Try reseating the blades.
Troubleshooting tools
Cause
The HPE Integrity BL860c i4 Server Blade uses LEDs and other tools to help troubleshoot problems that
occur in the server blade.
Controls and ports
92 Troubleshooting tools
This issue is fixed when the MCA does not repeat.
Page 93
Front panel view
1Monarch blade indicator7HDD bay 2
2UID LED8Blade power LED
3Blade health LED9Partition Identifier
4NICs 1, 2, 3, 410 Physical Presence Button
5Monarch power button11 SUV connector
6HDD bay 1
Rear panel view
Front panel view93
Page 94
1Power connectors
2GBX signal connectors
Server blade LEDs
Front panel LEDs
ItemDescriptionStatus
1Monarch blade indicatorGreen = Blade is acting as
Monarch blade
Off = Blade is not Monarch or is
not conjoined
2UID LEDBlue = Identified
Blue flashing = Active remote
management
Off = No active remote
management
3Blade health LEDGreen = Normal operation
Amber flashing = Degraded
condition
Red flashing = Critical condition
Table Continued
94 Server blade LEDs
Page 95
ItemDescriptionStatus
4NICs 1, 2, 3, 4Green = Network linked
Green flashing = Network activity
Off = No link or activity
5Monarch power button / Monarch
power LED
Green = Blade is acting as
Monarch and is powered on
Amber = Blade is acting as
Monarch and is in standby
Off = Blade is not Monarch
6SUV connectorN/A
7Physical Presence ButtonN/A
8Partition IdentifierGreen = Blade is in a partition
Off = Blade is not in a partition
9Blade power LEDGreen = Server blade is powered
on
Amber = standby (auxiliary
power available)
1
Off = Off
10iLO 3 Heartbeat (behind grill)Green flashing = iLO 3 Active
Amber flashing = iLO 3 failure
1
If the Onboard Administrator denies power to the server blade, the server blade returns to Standby mode.
SAS disk drive LEDs
The BL860c i4 Sever Blade may have up to 2 hard disk drives installed. Each drive has two LEDs which
display the drive status.
ItemDescription
1Fault/UID LED (amber/blue)
2Online LED (green)
Off = no standby voltage
SAS disk drive LEDs95
Page 96
SAS hard drive LED combinations
NOTE:
Predictive failure alerts only occur when the hard drive is connected to a Smart Array controller.
Table 12: SAS disk drive LEDs: RAID mode
Online/activity LED (green)Fault/UID LED (amber/blue)Interpretation
On, off, or flashingAlternating amber and blueThe drive has failed, or a
predictive failure alert has been
received for this drive; it also has
been selected by a management
application.
On, off, or flashingSteadily blueThe drive is operating normally,
and it has been selected by a
management application.
OnAmber, flashing regularly (1 Hz)A predictive failure alert has been
received for this drive. Replace
the drive as soon as possible.
Do not remove the drive.
Removing a drive might
terminate the current operation
and cause data loss.
The drive is part of an array that
is undergoing capacity expansion
or stripe migration, but a
predictive failure alert has been
received for this drive. To
minimize the risk of data loss, do
not replace the drive until the
expansion or migration is
complete.
Do not remove the drive.
Removing a drive might
terminate the current operation
and cause data loss.
The drive is rebuilding, erasing,
or it is part of an array that is
undergoing capacity expansion
or stripe migration.
96Troubleshooting
Table Continued
Page 97
Online/activity LED (green)Fault/UID LED (amber/blue)Interpretation
Flashing irregularlyAmber, flashing regularly (1 Hz)The drive is active, but a
predictive failure alert has been
received for this drive. Replace
the drive as soon as possible.
Flashing irregularlyOffThe drive is active, and it is
operating normally.
OffSteadily amberA critical fault condition has been
identified for this drive, and the
controller has placed it offline.
Replace the drive as soon as
possible.
OffAmber, flashing regularly (1 Hz)A predictive failure alert has been
received for this drive. Replace
the drive as soon as possible.
OffOffThe drive is offline, a spare, or
not configured as part of an
array.
Table 13: SAS disk drive LEDs : HBA mode
Drive activity LED status Drive status LED
status
Off, or flashingAlternating amber
and blue
Off, or flashingBlueThe drive is operating normally, and it has been
Flashing irregularlyRegularly flashing
amber(1 Hz)
Flashing irregularlyOffThe drive is active, and it is operating normally.
OffAmberA critical fault condition has been identified for this
OffRegularly flashing
amber(1 Hz)
Indication
The drive has failed, or a predictive failure alert has
been received for this drive; it also has been
selected by a management application.
selected by a management application.
The drive is active, but a predictive failure alert has
been received for this drive. Replace the drive as
soon as possible.
drive. Replace the drive as soon as possible.
A predictive failure alert has been received for this
drive. Replace the drive as soon as possible.
OffOffThe drive is online.
OffOffNormal operation; no activity
Troubleshooting97
Page 98
Blade Link LEDs
DescriptionStatus
Busy LEDGreen = Blade Link is currently active. Do not
Virtual Front Panel LEDs in the iLO 3 TUI
remove.
Off = Blade Link is not currently active. Safe to
remove.
iLO 3 has no LED that equates to the Blade Health LED located on the front panel of each individual
BL860c i4 Server Blade. The Blade Health LED represents the health of the individual server blade. The
virtual LEDs in the iLO 3 GUI and TUI reflect system/partition health. The closest equivalency to the
Blade Health LED is a field called Bay [x] Health on the System Health page of the iLO 3 GUI.
The VFP, located in the iLO 3 TUI directly off the MP MAIN MENU, is a textual representation of the
system/partition state as well as the system’s status with respect to where it is in the boot process
(running non-OS code, etc).
The following virtual LEDs are located in the VFP:
•HEALTH
•SYSTEM
•LOCATOR
•POWER
98 Blade Link LEDs
Page 99
Table 14: Indications for the iLO 3 TUI LEDs
VFP LEDPurposeEquivalentIndicationsValue
HEALTHRepresents the
health of the entire
system/partition.
Shows the worst
health case of any
blade that is part of
the partition.
System Health LED in the
iLO 3 GUI's Virtual Front
Panel
Flashing
amber
Flashing redHealth of the partition is
Health of the partition and the
health of one or more blades
within the partition is
DEGRADED, meaning that:
•A FRU has failed
•A fatal event has been
logged (Level seven alert
events cause this
indication if there are no
FRU failures.)
NOTE:
If the SYSTEM LED
lights red, this
counts as a FRU
failure for the
HEALTH LED.
FATAL and the health of one
or more blades within the
partition are FATAL, meaning
that:
SYSTEMThere is only one
SYSTEM LED per
partition. The state
will be the same for
System Event Log Health
as displayed on the
A FRU has failed and A fatal
event has been logged (Level
seven events cause this
indication if there are FRU
failures on one of the blades
in the partition.)
A reset will clear this LED of a
fatal event, but not a FRU
failure. A power cycle will
clear both.
Steady green All internal FRUs in the
partition are healthy, there are
no fatal events, and the
system power is on.
OffPower to the system/partition
is off and there were no
AMBER or RED HEALTH
indications on the last boot.
Flashing redThere has been a FATAL
event since the last power on
or system reset.
A reset or power on will clear
this LED.
Table Continued
Troubleshooting99
Page 100
every blade in the
partition.
System Health page in
the iLO 3 GUI
There is no physical LED
on the blade's front panel
that represents this state.
Steady green System is beginning to boot
the OS.
OffOS is not booting and there
are no fatal events.
LOCATOR Identifies the system/
partition and will blink
while a firmware
update is in progress.
NOTE:
All blades in a
partition show
the same
LOCATOR
status at all
times.
NOTE:
You can toggle
the UID in
software from
either the iLO 3
TUI or web
GUI. There is
no UID
hardware
button on the
server blade
itself to turn it
on or off.
•UID LED on the front
panel of the individual
blade
•Locator UID in iLO 3
•Virtual Indicator in the
Onboard
Administrator’s
interface
OnThe system/partition is
identified.
OffSystem/partition is not
identified and no firmware
update is actively writing to
any of the system ROMs.
FlashingPower should not be
interrupted; firmware is
actively burning a ROM.
NOTE:
On ProLiant servers this
LED blinks to indicate
remote management;
however this is not
currently a supported
indication on Integrity
servers.
ON/OFF
toggle
Toggles the current LED state
to indicate transitional mode
states activated by a push and
hold of the Physical Presence
button.
If the Locator is off it will
toggle on and then off, but if
the Locator is on, it will toggle
off and then on. If the Locator
is flashing, no action will be
taken.
POWERIdentifies the system/
partition power state.
SUV Cable and Ports
The SUV port on the front of the server blade is used in conjunction with an SUV cable to connect the
server to external devices such as a terminal emulator or monitor. In a conjoined server, only the SUV
port on the Monarch Server is active.
100 SUV Cable and Ports
•Monarch Power button
LED on the front panel
of the Monarch blade
•System Power LED in
the iLO 3 GUI
Steady green System/partition is fully
powered on.
Steady amber System/partition is in standby.
OffThe system/partition is
powered off.
Loading...
+ hidden pages
You need points to download manuals.
1 point = 1 manual.
You can buy points or you can get point for every manual you upload.