This document is for the person who installs, administers, and troubleshoots servers and
storage systems. Hewlett Packard Enterprise assumes you are qualified in the servicing of
computer equipment and trained in recognizing hazards in products with hazardous energy
levels.
Part Number: 880928-001
Published: November 2017
Edition: 1
Copyright 2017, Hewlett Packard Enterprise Development LP
Notices
The information contained herein is subject to change without notice. The only warranties for Hewlett
Packard Enterprise products and services are set forth in the express warranty statements accompanying
such products and services. Nothing herein should be construed as constituting an additional warranty.
Hewlett Packard Enterprise shall not be liable for technical or editorial errors or omissions contained
herein.
Confidential computer software. Valid license from Hewlett Packard Enterprise required for possession,
use, or copying. Consistent with FAR 12.211 and 12.212, Commercial Computer Software, Computer
Software Documentation, and Technical Data for Commercial Items are licensed to the U.S. Government
under vendor's standard commercial license.
Links to third-party websites take you outside the Hewlett Packard Enterprise website. Hewlett Packard
Enterprise has no control over and is not responsible for information outside the Hewlett Packard
Enterprise website.
4Serial label pull tab or optional Systems Insight Display
5iLO service port
6USB 3.0 port
Universal media bay components
ItemDescription
1USB 2.0 port
2Video display port
3Optical disk drive (optional)
4Drives (optional)
8 Component identification
12LFF front panel components
ItemDescription
1Drive bays
8LFF front panel components
ItemDescription
1Drives (optional)
2LFF power switch module
3Drive bays
LFF power switch module components
Component identification9
ItemDescription
1Optical disk drive
2Serial label pull tab
3USB 3.0 port
4iLO service port
5Video display port
Front panel LEDs and buttons
SFF front panel LEDs and button
ItemDescriptionStatus
1Power On/Standby button and
system power LED
1
•Solid green = System on
•Flashing green (1 Hz/cycle per sec) = Performing
power on sequence
•Solid amber = System in standby
•Off = No power present
2Health LED
1
•Solid green = Normal
•Flashing green (1 Hz/cycle per sec) = iLO is
rebooting
•Flashing amber = System degraded
•Flashing red (1 Hz/cycle per sec) = System critical
2
3
Table Continued
10 Front panel LEDs and buttons
ItemDescriptionStatus
3NIC status LED
1
•Solid green = Link to network
•Flashing green (1 Hz/cycle per sec) = Network
active
•Off = No network activity
4UID button/LED
1
•Solid blue = Activated
•Flashing blue:
◦1 Hz/cycle per sec = Remote management or
firmware upgrade in progress
◦4 Hz/cycle per sec = iLO manual reboot
sequence initiated
◦8 Hz/cycle per sec = iLO manual reboot
sequence in progress
◦Off = Deactivated
1
When all four LEDs described in this table flash simultaneously, a power fault has occurred.
2
Facility power is not present, the power cord is not attached, no power supplies are installed, power supply failure has
occurred, or the power button cable is disconnected.
3
If the health LED indicates a degraded or critical state, review the system IML or use iLO to review the system health
status.
12LFF front panel LEDs and button
Component identification11
ItemDescriptionStatus
1Health LED
1
•Solid green = Normal
•Flashing green (1 Hz/cycle per sec) = iLO is
rebooting
•Flashing amber = System degraded
•Flashing red (1 Hz/cycle per sec) = System critical
2
2Power On/Standby button and
system power LED
3NIC status LED
4UID button/LED
1
1
1
•Solid green = System on
•Flashing green (1 Hz/cycle per sec) = Performing
power on sequence
•Solid amber = System in standby
•Off = No power present
3
•Solid green = Link to network
•Flashing green (1 Hz/cycle per sec) = Network
active
•Off = No network activity
•Solid blue = Activated
•Flashing blue:
◦1 Hz/cycle per sec = Remote management or
firmware upgrade in progress
◦4 Hz/cycle per sec = iLO manual reboot
sequence initiated
1
When all four LEDs described in this table flash simultaneously, a power fault has occurred.
2
If the health LED indicates a degraded or critical state, review the system IML or use iLO to review the system health
status.
3
Facility power is not present, the power cord is not attached, no power supplies are installed, power supply failure has
occurred, or the power button cable is disconnected.
12Component identification
◦8 Hz/cycle per sec = iLO manual reboot
sequence in progress
•Off = Deactivated
LFF power switch module LEDs and button
ItemDescriptionStatus
1UID button/LED
1
•Solid blue = Activated
2Health LED
•Flashing blue:
◦1 Hz/cycle per sec = Remote management or
firmware upgrade in progress
◦4 Hz/cycle per sec = iLO manual reboot
sequence initiated
◦
◦8 Hz/cycle per sec = iLO manual reboot
sequence in progress
•Off = Deactivated
1
•Solid green = Normal
•Flashing green (1 Hz/cycle per sec) = iLO is
rebooting
•Flashing amber = System degraded
•Flashing red (1 Hz/cycle per sec) = System critical
2
Table Continued
Component identification13
ItemDescriptionStatus
3NIC status LED
1
•Solid green = Link to network
•Flashing green (1 Hz/cycle per sec) = Network
active
•Off = No network activity
4Power On/Standby button and
system power LED
1
When all four LEDs described in this table flash simultaneously, a power fault has occurred.
2
If the health LED indicates a degraded or critical state, review the system IML or use iLO to review the system health
status.
3
Facility power is not present, the power cord is not attached, no power supplies are installed, power supply failure has
occurred, or the power button cable is disconnected.
UID button functionality
The UID button can be used to display the HPE ProLiant Pre-boot Health Summary when the server will
not power on. For more information, see the latest HPE iLO User Guide on the Hewlett Packard
Enterprise website.
Power fault LEDs
The following table provides a list of power fault LEDs, and the subsystems that are affected. Not all
power faults are used by all servers.
1
•Solid green = System on
•Flashing green (1 Hz/cycle per sec) = Performing
power on sequence
•Solid amber = System in standby
•Off = No power present
3
SubsystemLED behavior
System board1 flash
Processor2 flashes
Memory3 flashes
Riser board PCIe slots4 flashes
FlexibleLOM5 flashes
Removable HPE Flexible Smart Array
controller/Smart SAS HBA controller
System board PCIe slots7 flashes
Power backplane or storage backplane8 flashes
Power supply9 flashes
14 UID button functionality
6 flashes
Systems Insight Display LEDs
The Systems Insight Display LEDs represent the system board layout. The display enables diagnosis with
the access panel installed.
DescriptionStatus
Processor LEDs
DIMM LEDs
Fan LEDs
NIC LEDs
Power supply LEDs
PCI riser LED
Off = Normal
Amber = Failed processor
Off = Normal
Amber = Failed DIMM or configuration issue
Off = Normal
Amber = Failed fan or missing fan
Off = No link to network
Solid green = Network link
Flashing green = Network link with activity
If power is off, the front panel LED is not active. For
status, see Rear panel LEDs.
Off = Normal
Solid amber = Power subsystem degraded, power
supply failure, or input power lost.
Off = Normal
Over temp LED
Amber = Incorrectly installed PCI riser cage
Off = Normal
Amber = High system temperature detected
Table Continued
Systems Insight Display LEDs15
DescriptionStatus
Amp Status LED
Power cap LED
When the health LED on the front panel illuminates either amber or red, the server is experiencing a
health event. For more information on the combination of these LEDs, see sSystem Insight Displaycombined LED description).
Off = AMP modes disabled
Solid green = AMP mode enabled
Solid amber = Failover
Flashing amber = Invalid configuration
Off = System is in standby, or no cap is set.
Solid green = Power cap applied
Systems Insight Display combined LED descriptions
The combined illumination of the following LEDs indicates a system condition:
•Systems Insight Display LEDs
•System power LED
•Health LED
Systems Insight Display
LED and color
Processor (amber)RedAmber
Processor (amber)AmberGreenProcessor in socket X is in a pre-
DIMM (amber)RedGreenOne or more DIMMs have failed.
DIMM (amber)AmberGreenDIMM in slot X is in a pre-failure
Over temp (amber)AmberGreenThe Health Driver has detected a
Health
LED
System
power LED
Status
One or more of the following
conditions may exist:
•Processor in socket X has failed.
•Processor X is not installed in the
socket.
•Processor X is unsupported.
•ROM detects a failed processor
during POST.
failure condition.
condition.
cautionary temperature level.
Over temp (amber)RedAmberThe server has detected a hardware
16 Systems Insight Display combined LED descriptions
critical temperature level.
Table Continued
Systems Insight Display
LED and color
PCI riser (amber)RedGreenThe PCI riser cage is not seated
Fan (amber)AmberGreenOne fan has failed or has been
Fan (amber)RedGreenTwo or more fans have failed or been
Health
LED
System
power LED
Status
properly.
removed.
removed.
Power supply (amber)RedAmber
Power supply (amber)AmberGreen
One or more of the following
conditions may exist:
•Only one power supply is installed
and that power supply is in
standby.
•Power supply fault
•System board fault
One or more of the following
conditions may exist:
•Redundant power supply is
installed and only one power
supply is functional.
•AC power cord is not plugged into
redundant power supply.
•Redundant power supply fault
•Power supply mismatch at POST
or power supply mismatch through
hot-plug addition
Power cap (off)—AmberStandby
Power cap (green)—Flashing
green
Power cap (green)—GreenPower is available.
Power cap (flashing amber)—AmberPower is not available.
IMPORTANT:
If more than one DIMM slot LED is illuminated, further troubleshooting is required. Test each bank of
DIMMs by removing all other DIMMs. Isolate the failed DIMM by replacing each DIMM in a bank
with a known working DIMM.
Waiting for power
Component identification17
Rear panel components
ItemDescription
1Primary riser slots 1 to 3 (Optional drive cage)
2Optional riser slots 4 to 6 (Optional drive cage)
3Optional riser slots 7 to 8 (Optional drive cage)
4Power supply 1
5Power supply 2
6Video port
7Serial port (optional)
1
81Gb RJ-45 ports 1 to 4
9iLO management port
10USB 3.0 ports
11FlexibleLOM slot
1
When a tertiary riser cage is installed as shown, the serial port can be installed in riser slot 6.
Rear panel LEDs
18 Rear panel components
ItemDescriptionStatus
1UID LED
•Off = Deactivated
•Solid blue = Activated
•Flashing blue = System being
managed remotely
2Link LED
•Off = No network link
•Green = Network link
3Activity LED
•Off = No network activity
•Solid green = Link to network
•Flashing green = Network activity
4Power supply
LEDs
•Off = System is off or power supply
has failed.
•Solid green = Normal
Component identification19
System board components
ItemDescription
1FlexibleLOM connector
2System maintenance switch
3Primary (processor 1) PCIe riser connector
4Front display port/USB 2.0 connector
20 System board components
Table Continued
ItemDescription
5M.2 connectors (x2)
6Dual USB port
7x1 SATA port 4
8Optical/SATA port 5
9Front power/USB 3.0 connector
10Drive backplane power connectors
11Smart Storage Battery connector
12Chassis Intrusion Detection connector
13Drive backplane power connector
14microSD card slot
15NVMe connector (port 7b)
16NVMe connector (port 8b)
17Flexible Smart Array connector
18Secondary (processor 2) PCIe riser connector
19System battery
20Tertiary (processor 2) PCI riser connector
1
1
21TPM connector
22Serial port connector
1
Requires a second processor.
System maintenance switch descriptions
PositionDefaultFunction
1
S1
S2Off
S3OffReserved
S4OffReserved
S5Off
Off
•Off = iLO security is enabled.
•On = iLO security is disabled.
•Off = System configuration can be changed.
•On = System configuration is locked.
•Off = Power-on password is enabled.
•On = Power-on password is disabled.
System maintenance switch descriptions21
Table Continued
PositionDefaultFunction
3
S6 2,
Off
•Off = No function
•On = Restore default manufacturing settings
S7Off
•Off = Set default boot mode to UEFI.
•On = Set default boot mode to legacy.
S8—Reserved
S9—Reserved
S10—Reserved
S11—Reserved
S12—Reserved
1
To access the redundant ROM, set S1, S5, and S6 to On.
2
When the system maintenance switch position 6 is set to the On position, the system is prepared to restore all
configuration settings to their manufacturing defaults.
3
When the system maintenance switch position 6 is set to the On position and Secure Boot is enabled, some
configurations cannot be restored. For more information, see Secure Boot on page 158.
DIMM slot locations
DIMM slots are numbered sequentially (1 through 16) for each processor. The supported AMP modes use
the letter assignments for population guidelines.
22 DIMM slot locations
Processor and socket components
ItemDescription
1Pin field
2Rail frame
3Carrier frame
4Processor
5Force frame
6Captive screws (Torx T-20)
Drives
SAS/SATA drive components and LEDs
Processor and socket components23
ItemDescriptionStatus
1Locate
•Solid blue = The drive is being identified by a host
application.
•Flashing blue = The drive carrier firmware is being
updated or requires an update.
2Activity ring LED
•Rotating green = Drive activity.
•Off = No drive activity.
3Do not remove LED
•Solid white = Do not remove the drive. Removing
the drive causes one or more of the logical drives to
fail.
•Off = Removing the drive does not cause a logical
drive to fail.
4Drive status LED
•Solid green = The drive is a member of one or more
logical drives.
•Flashing green = The drive is rebuilding or
•Flashing amber/green = The drive is a member of
•Flashing amber = The drive is not configured and
•Solid amber = The drive has failed.
•Off = The drive is not configured by a RAID
NVMe drive components and LEDs
performing a RAID migration, strip size migration,
capacity expansion, or logical drive extension, or is
erasing.
one or more logical drives and predicts the drive will
fail.
predicts the drive will fail.
controller.
24 NVMe drive components and LEDs
ItemDescription
1Release lever
2Activity ring
3Do Not Remove LED
1
4Request to Remove NVMe Drive button
1
Do not remove an NVMe SSD from the drive bay while the Do Not Remove button LED is flashing. The Do Not
Remove button LED flashes to indicate that the device is still in use. Removing the NVMe SSD before the device has
completed and ceased signal/traffic flow can cause loss of data.
uFF drive components and LEDs
ItemDescriptionStatus
1Locate
•Off—Normal
•Solid blue—The drive is being identified by a host
application
•Flashing blue—The drive firmware is being updated
or requires an update
2uFF drive ejection latchRemoves the uFF drive when released
3Do not remove LED
•Off—OK to remove the drive. Removing the drive
does not cause a logical drive to fail.
•Solid white—Do not remove the drive. Removing
the drive causes one or more of the logical drives to
fail.
Table Continued
uFF drive components and LEDs25
ItemDescriptionStatus
4Drive status LED
•Off—The drive is not configured by a RAID
controller
•Solid green—The drive is a member of one or more
logical drives
•Flashing green (4 Hz)—The drive is operating
normally and has activity
•Flashing green (1 Hz)—The drive is rebuilding or
performing a RAID migration, stripe size migration,
capacity expansion, logical drive extension, or is
erasing
•Flashing amber/green (1 Hz)—The drive is a
member of one or more logical drives that predicts
the drive will fail
•Solid amber—The drive has failed
•Flashing amber (1 Hz)—The drive is not configured
and predicts the drive will fail
5Adapter ejection release latch
and handle
Fan bay numbering
Removes the SFF flash adapter when released
26 Fan bay numbering
Drive box identification
Front boxes
ItemDescription
1Box 1
2Box 2
3Box 3
ItemDescription
1Box 1
2Box 2
3Box 3
Rear boxes
ItemDescription
1Box 4
2Box 5
3Box 6
Drive box identification27
ItemDescription
1Box 4
2Box 6
Midplane box (LFF only)
ItemDescription
1Box 7
Drive bay numbering
Drive bay numbering depends on how the drive backplanes are connected:
•To a controller:
◦Embedded controllers use the onboard SATA ports.
◦Type-a controllers install to the type-a smart array connector.
◦Type-p controllers install to a PCIe riser.
•To a SAS expander:
Installs in the primary or secondary PCIe riser
28 Drive bay numbering
Drive bay numbering: Smart Array controller
When the drive backplane is connected directly to a storage controller, then each drive box starts at 1.
The following images are examples of common configurations.
24SFF
12LFF
6SFF rear panel
2SFF + 3LFF rear panel
Drive bay numbering: Smart Array controller29
Universal media bay
8LFF + Universal media bay, optional 2SFF, and optical drive
Midplane box
Drive bay numbering: SAS expander
Drive numbering through a SAS Expander is continuous:
•SAS expander port 1 always connects to port 1 of the controller.
•SAS expander port 2 always connects to port 2 of the controller.
•SAS expander port 3 = drive numbers 1 to 4.
30 Drive bay numbering: SAS expander
•SAS expander port 4 = drive numbers 5 to 8.
•SAS expander port 5 = drive numbers 9 to 12.
•SAS expander port 6 = drive numbers 13 to 16.
•SAS expander port 7 = drive numbers 17 to 20.
•SAS expander port 8 = drive numbers 21 to 24.
•SAS expander port 9 = drive numbers 25 to 28.
Common configuration examples:
NVMe drives
When any stacked 2SFF drive cage is connected to the SAS expander, the drive numbering skips the
second number to allow uFF drive bay numbering on page 33. For example, when a rear 2SFF drive
cage is connected to SAS expander port 9, then the drive numbers are 25 and 27.
When the front 24SFF bays are populated, any installed rear 2SFF drives are always 25 and 27.
If a 2SFF drive cage is connected to SAS expander port 3, then the drive numbers are 1 and 3.
Universal media bay
Component identification31
8LFF front panel
12LFF front panel
Midplane box (LFF only)
32Component identification
Front 12LFF + Midplane 4LFF + All rear 2SFF
Drive bay numbering: NVMe drives
Server populated with NVMe drives and NVMe risers
uFF drive bay numbering
There are two uFF drives in each drive carrier.
Drives connected to a controller
•The left bay = The default bay number of the server
•The right bay = The default bay number of the server + 100
Drives connected to a SAS expander
Drive bay numbering: NVMe drives33
For example:
•If the drives are connected to port 3 of the SAS expander, then the uFF drives are 1 to 4.
•If the drives are connected to port 9 of the SAS expander, then the uFF drives are 25 to 28.
34Component identification
Operations
Power up the server
Procedure
To power up the server, press the Power On/Standby button.
Power down the server
Before powering down the server for any upgrade or maintenance procedures, perform a backup of
critical server data and programs.
IMPORTANT:
When the server is in standby mode, auxiliary power is still being provided to the system.
To power down the server, use one of the following methods:
•Press and release the Power On/Standby button.
This method initiates a controlled shutdown of applications and the OS before the server enters
standby mode.
•Press and hold the Power On/Standby button for more than 4 seconds to force the server to enter
standby mode.
This method forces the server to enter standby mode without properly exiting applications and the OS.
If an application stops responding, you can use this method to force a shutdown.
•Use a virtual power button selection through iLO.
This method initiates a controlled remote shutdown of applications and the OS before the server
enters standby mode.
Before proceeding, verify that the server is in standby mode by observing that the system power LED is
amber.
Extend the server from the rack
WARNING:
To reduce the risk of personal injury or equipment damage, be sure that the rack is adequately
stabilized before extending a component from the rack.
1. Pull down the quick release levers on each side of the server.
2. Extend the server from the rack.
Operations35
3. After performing the installation or maintenance procedure, slide the server back into the rack, and
then press the server firmly into the rack to secure it in place.
WARNING:
To reduce the risk of personal injury, be careful when pressing the server rail-release latches and
sliding the server into the rack. The sliding rails could pinch your fingers.
Extending the server from the rack
WARNING:
To reduce the risk of personal injury or equipment damage, be sure that the rack is adequately
stabilized before extending anything from the rack.
Procedure
Pull down the quick release levers on each side of the server, and then extend the server from the rack.
36 Extending the server from the rack
Removing the server from the rack
To remove the server from a Hewlett Packard Enterprise, Compaq-branded, Telco, or third-party rack:
Procedure
1. Power down the server.
2. Extend the server from the rack.
3. Disconnect the cabling and remove the server from the rack.
For more information, see the documentation that ships with the rack mounting option.
4. Place the server on a sturdy, level surface.
Installing the server into the rack
Procedure
1. Observe the following alert:
CAUTION:
Always plan the rack installation so that the heaviest item is on the bottom of the rack. Install the
heaviest item first, and continue to populate the rack from the bottom to the top.
2. Install the server and cable management arm into the rack. For more information, see the installation
instructions that ship with the rack rail system.
3. Connect peripheral devices to the server. For information on identifying connectors, see "Rear panel
components."
WARNING:
To reduce the risk of electric shock, fire, or damage to the equipment, do not plug telephone or
telecommunications connectors into RJ-45 connectors.
Removing the server from the rack37
4. Connect the power cord to the rear of the server.
5. Install the power cord anchors.
6. Secure the cables to the cable management arm.
IMPORTANT:
When using cable management arm components, be sure to leave enough slack in each of the
cables to prevent damage to the cables when the server is extended from the rack.
7. Connect the power cord to the AC power source.
38Operations
WARNING:
To reduce the risk of electric shock or damage to the equipment:
•Do not disable the power cord grounding plug. The grounding plug is an important safety
feature.
•Plug the power cord into a grounded (earthed) electrical outlet that is easily accessible at all
times.
•Unplug the power cord from the power supply to disconnect power to the equipment.
•Do not route the power cord where it can be walked on or pinched by items placed against it.
Pay particular attention to the plug, electrical outlet, and the point where the cord extends
from the server.
Remove the access panel
WARNING:
To reduce the risk of personal injury from hot surfaces, allow the drives and the internal system
components to cool before touching them.
CAUTION:
Do not operate the server for long periods with the access panel open or removed. Operating the
server in this manner results in improper airflow and improper cooling that can lead to thermal
damage.
To remove the component:
Procedure
1. Power down the server.
2. Extend the server from the rack.
3. Open or unlock the locking latch, slide the access panel to the rear of the chassis, and remove the
access panel.
Installing the access panel
Procedure
1. Place the access panel on top of the server with the latch open.
Allow the panel to extend past the rear of the server approximately 1.25 cm (0.5 in).
2. Push down on the latch.
The access panel slides to a closed position.
3. Tighten the security screw on the latch, if needed.
Remove the access panel39
Removing the fan cage
CAUTION:
Do not operate the server for long periods with the access panel open or removed. Operating the
server in this manner results in improper airflow and improper cooling that can lead to thermal
damage.
IMPORTANT:
For optimum cooling, install fans in all primary fan locations.
Procedure
1. Power down the server.
2. Do one of the following:
•Disconnect each power cord from the power source.
•Disconnect each power cord from the server.
3. Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack.
4. Remove the access panel.
5. Remove the air baffle.
6. Remove the fan cage.
40 Removing the fan cage
Installing the fan cage
CAUTION:
Do not operate the server for long periods with the access panel open or removed. Operating the
server in this manner results in improper airflow and improper cooling that can lead to thermal
damage.
IMPORTANT:
For optimum cooling, install fans in all primary fan locations.
Removing the air baffle or midplane drive cage
CAUTION:
Do not detach the cable that connects the battery pack to the cache module. Detaching the cable
causes any unsaved data in the cache module to be lost.
CAUTION:
For proper cooling, do not operate the server without the access panel, baffles, expansion slot
covers, or blanks installed. If the server supports hot-plug components, minimize the amount of time
the access panel is open.
Procedure
1. Power down the server.
2. Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3. Do one of the following:
Installing the fan cage41
•Extend the server from the rack.
•Remove the server from the rack.
4. Remove the access panel.
5. Do one of the following:
•Remove the air baffle.
•Remove the 4LFF midplane drive cage:
a. Disconnect all cables.
b. Remove all drives.
Be sure to note the location of each drive.
c. Remove the drive cage.
CAUTION:
Do not drop the drive cage on the system board. Dropping the drive cage on the system
board might damage the system or components. Remove all drives and use two hands
when installing or removing the drive cage.
42Operations
Installing the air baffle
Procedure
1. Observe the following alerts.
CAUTION:
For proper cooling, do not operate the server without the access panel, baffles, expansion slot
covers, or blanks installed. If the server supports hot-plug components, minimize the amount of
time the access panel is open.
CAUTION:
Do not detach the cable that connects the battery pack to the cache module. Detaching the cable
causes any unsaved data in the cache module to be lost.
2. Install the air baffle.
Installing the air baffle43
Removing a riser cage
CAUTION:
To prevent damage to the server or expansion boards, power down the server and remove all AC
power cords before removing or installing the PCI riser cage.
Procedure
1. Power down the server.
2. Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3. Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack.
4. Remove the access panel.
5. Remove the riser cage.
Primary and secondary riser cages
44 Removing a riser cage
Tertiary riser cage
Removing a riser slot blank
CAUTION:
To prevent improper cooling and thermal damage, do not operate the server unless all PCI slots
have either an expansion slot cover or an expansion board installed.
Procedure
1. Power down the server.
2. Remove all power:
Removing a riser slot blank45
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3. Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack .
4. Remove the access panel.
5. Remove the riser cage.
6. Remove the blank.
Removing the hard drive blank
Remove the component as indicated.
Releasing the cable management arm
Release the cable management arm and then swing the arm away from the rack.
46 Removing the hard drive blank
Accessing the Systems Insight Display
The Systems Insight Display is only supported on SFF models. To access a Systems Insight Display, use
the following procedure.
Procedure
1. Press and release the panel.
2. After the display fully ejects, rotate the display to view the LEDs.
Accessing the Systems Insight Display47
Setup
HPE support services
Delivered by experienced, certified engineers, HPE support services help you keep your servers up and
running with support packages tailored specifically for HPE ProLiant systems. HPE support services let
you integrate both hardware and software support into a single package. A number of service level
options are available to meet your business and IT needs.
HPE support services offer upgraded service levels to expand the standard product warranty with easyto-buy, easy-to-use support packages that will help you make the most of your server investments. Some
of the HPE support services for hardware, software or both are:
•Foundation Care – Keep systems running.
◦6-Hour Call-to-Repair
◦4-Hour 24x7
◦Next Business Day
•Proactive Care – Help prevent service incidents and get you to technical experts when there is one.
◦6-Hour Call-to-Repair
◦4-Hour 24x7
◦Next Business Day
•Startup and implementation services for both hardware and software
•HPE Education Services – Help train your IT staff.
For more information on HPE support services, see the
Setup overview
Procedure
1. Review the operational requirements for the server (Operational requirements on page 49).
2. Read the following safety notices, warnings, and cautions:
•Server warnings and cautions (Server warnings and cautions on page 51)
•Rack warnings (Rack warnings on page 52)
•Electrostatic discharge (Electrostatic discharge on page 52)
Hewlett Packard Enterprise website.
3. Verify the contents in the server box (Server box contents on page 53).
4. Install hardware options (Hardware options installation on page 55).
5. Install the server into a rack (Installing the server into the rack on page 37).
48 Setup
6. Configure the server (Configuring the server on page 53).
7. Install or deploy an operating system (Installing or deploying an operating system on page 53).
8. Register your server (Registering the server on page 54).
Operational requirements
Space and airflow requirements
To allow for servicing and adequate airflow, observe the following space and airflow requirements when
deciding where to install a rack:
•Leave a minimum clearance of 63.5 cm (25 in) in front of the rack.
•Leave a minimum clearance of 76.2 cm (30 in) behind the rack.
•Leave a minimum clearance of 121.9 cm (48 in) from the back of the rack to the back of another rack
or row of racks.
Hewlett Packard Enterprise servers draw in cool air through the front door and expel warm air through the
rear door. Therefore, the front and rear rack doors must be adequately ventilated to allow ambient room
air to enter the cabinet, and the rear door must be adequately ventilated to allow the warm air to escape
from the cabinet.
CAUTION:
To prevent improper cooling and damage to the equipment, do not block the ventilation openings.
When vertical space in the rack is not filled by a server or rack component, the gaps between the
components cause changes in airflow through the rack and across the servers. Cover all gaps with
blanking panels to maintain proper airflow.
CAUTION:
Always use blanking panels to fill empty vertical spaces in the rack. This arrangement ensures
proper airflow. Using a rack without blanking panels results in improper cooling that can lead to
thermal damage.
The 9000 and 10000 Series Racks provide proper server cooling from flow-through perforations in the
front and rear doors that provide 64 percent open area for ventilation.
CAUTION:
When using a Compaq branded 7000 series rack, install the high airflow rack door insert (PN
327281-B21 for 42U rack, PN 157847-B21 for 22U rack) to provide proper front-to-back airflow and
cooling.
Operational requirements49
CAUTION:
If a third-party rack is used, observe the following additional requirements to ensure adequate
airflow and to prevent damage to the equipment:
•Front and rear doors—If the 42U rack includes closing front and rear doors, you must allow
5,350 sq cm (830 sq in) of holes evenly distributed from top to bottom to permit adequate airflow
(equivalent to the required 64 percent open area for ventilation).
•Side—The clearance between the installed rack component and the side panels of the rack must
be a minimum of 7 cm (2.75 in).
Temperature requirements
To ensure continued safe and reliable equipment operation, install or position the system in a wellventilated, climate-controlled environment.
The maximum recommended ambient operating temperature (TMRA) for most server products is 35°C
(95°F). The temperature in the room where the rack is located must not exceed 35°C (95°F).
CAUTION:
To reduce the risk of damage to the equipment when installing third-party options:
•Do not permit optional equipment to impede airflow around the server or to increase the internal
rack temperature beyond the maximum allowable limits.
•Do not exceed the manufacturer’s TMRA.
Power requirements
Installation of this equipment must comply with local and regional electrical regulations governing the
installation of information technology equipment by licensed electricians. This equipment is designed to
operate in installations covered by NFPA 70, 1999 Edition (National Electric Code) and NFPA-75, 1992
(code for Protection of Electronic Computer/Data Processing Equipment). For electrical power ratings on
options, refer to the product rating label or the user documentation supplied with that option.
WARNING:
To reduce the risk of personal injury, fire, or damage to the equipment, do not overload the AC
supply branch circuit that provides power to the rack. Consult the electrical authority having
jurisdiction over wiring and installation requirements of your facility.
CAUTION:
Protect the server from power fluctuations and temporary interruptions with a regulating
uninterruptible power supply. This device protects the hardware from damage caused by power
surges and voltage spikes and keeps the system in operation during a power failure.
Electrical grounding requirements
The server must be grounded properly for proper operation and safety. In the United States, you must
install the equipment in accordance with NFPA 70, 1999 Edition (National Electric Code), Article 250, as
well as any local and regional building codes. In Canada, you must install the equipment in accordance
with Canadian Standards Association, CSA C22.1, Canadian Electrical Code. In all other countries, you
must install the equipment in accordance with any regional or national electrical wiring codes, such as the
International Electrotechnical Commission (IEC) Code 364, parts 1 through 7. Furthermore, you must be
50 Temperature requirements
sure that all power distribution devices used in the installation, such as branch wiring and receptacles, are
listed or certified grounding-type devices.
Because of the high ground-leakage currents associated with multiple servers connected to the same
power source, Hewlett Packard Enterprise recommends the use of a PDU that is either permanently wired
to the building’s branch circuit or includes a nondetachable cord that is wired to an industrial-style plug.
NEMA locking-style plugs or those complying with IEC 60309 are considered suitable for this purpose.
Using common power outlet strips for the server is not recommended.
Server warnings and cautions
WARNING:
This server is heavy. To reduce the risk of personal injury or damage to the equipment:
•Observe local occupational health and safety requirements and guidelines for manual material
handling.
•Get help to lift and stabilize the product during installation or removal, especially when the
product is not fastened to the rails. Hewlett Packard Enterprise recommends that a minimum of
two people are required for all rack server installations. If the server is installed higher than chest
level, a third person may be required to help align the server.
•Use caution when installing the server in or removing the server from the rack; it is unstable
when not fastened to the rails.
WARNING:
To reduce the risk of personal injury from hot surfaces, allow the drives and the internal system
components to cool before touching them.
WARNING:
To reduce the risk of personal injury, electric shock, or damage to the equipment, remove the power
cord to remove power from the server. The front panel Power On/Standby button does not
completely shut off system power. Portions of the power supply and some internal circuitry remain
active until AC/DC power is removed.
CAUTION:
Protect the server from power fluctuations and temporary interruptions with a regulating
uninterruptible power supply. This device protects the hardware from damage caused by power
surges and voltage spikes and keeps the system in operation during a power failure.
CAUTION:
Do not operate the server for long periods with the access panel open or removed. Operating the
server in this manner results in improper airflow and improper cooling that can lead to thermal
damage.
Server warnings and cautions51
Rack warnings
WARNING:
To reduce the risk of personal injury or damage to the equipment, be sure that:
•The leveling jacks are extended to the floor.
•The full weight of the rack rests on the leveling jacks.
•The stabilizing feet are attached to the rack if it is a single-rack installation.
•The racks are coupled together in multiple-rack installations.
•Only one component is extended at a time. A rack may become unstable if more than one
component is extended for any reason.
WARNING:
To reduce the risk of personal injury or equipment damage when unloading a rack:
•At least two people are needed to safely unload the rack from the pallet. An empty 42U rack can
weigh as much as 115 kg (253 lb), can stand more than 2.1 m (7 ft) tall, and might become
unstable when being moved on its casters.
•Never stand in front of the rack when it is rolling down the ramp from the pallet. Always handle
the rack from both sides.
WARNING:
To reduce the risk of personal injury or damage to the equipment, adequately stabilize the rack
before extending a component outside the rack. Extend only one component at a time. A rack may
become unstable if more than one component is extended.
WARNING:
When installing a server in a telco rack, be sure that the rack frame is adequately secured at the top
and bottom to the building structure.
Electrostatic discharge
Be aware of the precautions you must follow when setting up the system or handling components. A
discharge of static electricity from a finger or other conductor may damage system boards or other staticsensitive devices. This type of damage may reduce the life expectancy of the system or component.
To prevent electrostatic damage:
•Avoid hand contact by transporting and storing products in static-safe containers.
•Keep electrostatic-sensitive parts in their containers until they arrive at static-free workstations.
•Place parts on a grounded surface before removing them from their containers.
•Avoid touching pins, leads, or circuitry.
•Always be properly grounded when touching a static-sensitive component or assembly. Use one or
more of the following methods when handling or installing electrostatic-sensitive parts:
52 Rack warnings
◦Use a wrist strap connected by a ground cord to a grounded workstation or computer chassis. Wrist
straps are flexible straps with a minimum of 1 megohm ±10 percent resistance in the ground cords.
To provide proper ground, wear the strap snug against the skin.
◦Use heel straps, toe straps, or boot straps at standing workstations. Wear the straps on both feet
when standing on conductive floors or dissipating floor mats.
◦Use conductive field service tools.
◦Use a portable field service kit with a folding static-dissipating work mat.
If you do not have any of the suggested equipment for proper grounding, have an authorized reseller
install the part.
For more information on static electricity or assistance with product installation, contact an authorized
reseller.
Server box contents
The server shipping box contains the following contents:
•A server
•A power cord
•Rack-mounting hardware
•Documentation
Installing hardware options
Install any hardware options before initializing the server. For options installation information, refer to the
option documentation. For server-specific information, refer to "Hardware options installation."
Configuring the server
When the server is powered on, the POST screen is displayed. Use the following options to configure the
server:
•System utilities (F9)
Use this option to configure UEFI, RBSU, or other boot settings.
•Intelligent Provisioning (F10)
Use this option to configure drives, access Smart Storage Administrator, or begin installing or
deploying an operating system.
•Boot order (F11)
Use this option to select a boot device.
•Network boot (F12)
Use this option to PXE boot the server from the network.
Installing or deploying an operating system
Before installing an operating system, observe the following:
Server box contents53
•Be sure to read the HPE UEFI requirements for ProLiant servers on the Hewlett Packard Enterprisewebsite. If UEFI requirements are not met, you might experience boot failures or other errors when
installing the operating system.
•Update firmware before using the server for the first time, unless software or components require an
older version. For more information, see "Keeping the system current on page 160."
•For the latest information on supported operating systems, see the Hewlett Packard Enterprisewebsite.
•The server does not ship with OS media. All system software and firmware is preloaded on the server.
Registering the server
To experience quicker service and more efficient support, register the product at the Hewlett Packard
Enterprise Product Registration website.
54 Registering the server
Hardware options installation
Product QuickSpecs
For more information about product features, specifications, options, configurations, and compatibility, see
the product QuickSpecs on the Hewlett Packard Enterprise website (
Introduction
If more than one option is being installed, read the installation instructions for all the hardware options
and identify similar steps to streamline the installation process.
WARNING:
To reduce the risk of personal injury from hot surfaces, allow the drives and the internal system
components to cool before touching them.
CAUTION:
To prevent damage to electrical components, properly ground the server before beginning any
installation procedure. Improper grounding can cause electrostatic discharge.
Installing the bezel and bezel lock
http://www.hpe.com/info/qs).
Hardware options installation55
Power supply options
Hot-plug power supply calculations
For hot-plug power supply specifications and calculators to determine electrical and heat loading for the
server, see the Hewlett Packard Enterprise Power Advisor website (http://www.hpe.com/info/
poweradvisor/online).
Installing a redundant hot-plug power supply
CAUTION:
All power supplies installed in the server must have the same output power capacity. Verify that all
power supplies have the same part number and label color. The system becomes unstable and
might shut down if it detects different power supplies.
CAUTION:
To prevent improper cooling and thermal damage, do not operate the server unless all bays are
populated with either a component or a blank.
Procedure
1. Release the cable management arm to access the rear panel.
2. Remove the blank.
WARNING:
To reduce the risk of personal injury from hot surfaces, allow the power supply or power supply
blank to cool before touching it.
56 Power supply options
3. Insert the power supply into the power supply bay until it clicks into place.
4. Connect the power cord to the power supply.
5. Route the power cord.
Use the cable management arm and best practices when routing cords and cables.
6. Connect the power cord to the power source.
7. Observe the power supply LED.
Drive options
Drive guidelines
Depending on the configuration, the server supports SAS, SATA, and NVMe drives.
Observe the following general guidelines:
•The system automatically sets all drive numbers.
•If only one hard drive is used, install it in the bay with the lowest drive number.
For drive numbering, see "Drive bay numbering on page 28".
•The NVMe SSD is a PCIe bus device. Devices attached to a PCIe bus cannot be removed without
allowing the device and bus to complete and cease the signal/traffic flow.
Drive options57
Do not remove an NVMe SSD from the drive bay while the Do Not Remove button LED is flashing.
The Do Not Remove button LED flashes to indicate that the device is still in use. Removal of the
NVMe SSD before the device has completed and ceased signal/traffic flow can cause loss of data.
•Drives with the same capacity provide the greatest storage space efficiency when grouped into the
same drive array.
Supported drive carriers
Depending on the drive cage, the server supports the following drive carriers:
•SFF Smart Carrier (SC)
•SFF Smart Carrier NVMe (SCN)
•SFF Smart Carrier M.2 (SCM)
•LFF Smart Carrier (SC)
•LFF to SFF Smart Carrier Converter
Installing a hot-plug SAS or SATA drive
Procedure
1. Remove the drive blank.
2. Prepare the drive.
3. Install the drive.
58 Supported drive carriers
4. Observe the LED status of the drive.
Installing an NVMe drive
CAUTION:
To prevent improper cooling and thermal damage, do not operate the server unless all drive and
device bays are populated with either a component or a blank.
Procedure
1. Remove the drive blank.
2. Prepare the drive.
3. Install the drive.
Installing an NVMe drive59
4. Observe the LED status of the drive.
Installing a uFF drive and SCM drive carrier
IMPORTANT:
Not all drive bays support the drive carrier. To find supported bays, see the server QuickSpecs.
Procedure
1. If needed, install the uFF drive into the drive carrier.
2. Remove the drive blank.
3. Install the drives.
Push firmly near the ejection handle until the latching spring engages with the drive bay.
60 Installing a uFF drive and SCM drive carrier
4. Power on the server.
To configure the drive, use HPE Smart Storage Administrator.
Installing an M.2 drive
This procedure is for replacing M.2 drives located on an expansion card, riser, or the system board only.
Do not use this procedure to replace uFF drives.
Prerequisites
Before installing this option, be sure that you have the following:
•The components included with the hardware option kit
•T-10 Torx screwdriver
Procedure
1. Power down the server .
2. Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3. Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack .
4. Remove the access panel.
5. Remove the riser cage.
6. Install the drive.
Installing an M.2 drive61
The installation is complete.
Fan options
CAUTION:
To avoid damage to server components, fan blanks must be installed in fan bays 1 and 2 in a singleprocessor configuration.
CAUTION:
To avoid damage to the equipment, do not operate the server for extended periods of time if the
server does not have the optimal number of fans installed. Although the server might boot, Hewlett
Packard Enterprise does not recommend operating the server without the required fans installed
and operating.
Valid fan configurations are listed in the following table.
ConfigurationFan bay 1Fan bay 2Fan bay 3Fan bay 4Fan bay 5Fan bay 6
1 processorFan blankFan blankFanFanFanFan
1 processor 24-SFF
or 12-LFF
configuration with
high-performance
fans
2 processorsFanFanFanFanFanFan
For a single-processor configuration, excluding 24-SFF and 12-LFF configurations, four fans and two
blanks are required in specific fan bays for redundancy. A fan failure or missing fan causes a loss of
redundancy. A second fan failure or missing fan causes an orderly shutdown of the server.
For a dual-processor configuration or single-processor 24-SFF or 12-LFF configurations, six fans are
required for redundancy. A fan failure or missing fan causes a loss of redundancy. A second fan failure or
missing fan causes an orderly shutdown of the server.
High-performance fans might be necessary in 24-SFF and 12-LFF configurations for the following
installations:
62 Fan options
FanFanFanFanFanFan
•Optional GPU riser installations
•ASHRAE compliant configurations
For more information, see the Hewlett Packard Enterprise website.
The server supports variable fan speeds. The fans operate at minimum speed until a temperature change
requires a fan speed increase to cool the server. The server shuts down during the following temperaturerelated scenarios:
•At POST and in the OS, iLO performs an orderly shutdown if a cautionary temperature level is
detected. If the server hardware detects a critical temperature level before an orderly shutdown
occurs, the server performs an immediate shutdown.
•When the Thermal Shutdown feature is disabled in the BIOS/Platform Configuration (RBSU), iLO does
not perform an orderly shutdown when a cautionary temperature level is detected. Disabling this
feature does not disable the server hardware from performing an immediate shutdown when a critical
temperature level is detected.
CAUTION:
A thermal event can damage server components when the Thermal Shutdown feature is
disabled in the BIOS/Platform Configuration (RBSU).
Installing high-performance fans
CAUTION:
Caution: To prevent damage server, ensure that all DIMM latches are closed and locked before
installing the fans.
CAUTION:
Do not operate the server for long periods with the access panel open or removed. Operating the
server in this manner results in improper airflow and improper cooling that can lead to thermal
damage.
Procedure
1. Extend the server from the rack.
2. Remove the access panel.
3. If installed, remove all fan blanks.
Installing high-performance fans63
4. Remove the air baffle.
5. Remove all standard fans.
IMPORTANT:
Do not mix standard fans and high-performance fans in the same server.
6. Install high-performance fans in all fan bays.
64Hardware options installation
7. Install the air baffle.
8. Install the access panel.
9. Install the server into the rack.
Memory options
IMPORTANT:
This server does not support mixing LRDIMMs and RDIMMs. Attempting to mix any combination of
these DIMMs can cause the server to halt during BIOS initialization. All memory installed in the
server must be of the same type.
Memory-processor compatibility information
The server processor determines the type of DIMM that will be supported in the server.
For the latest memory configuration information, see the product QuickSpecs on the Hewlett Packard
Enterprise website.
DIMM specifications
DIMM specifications for AMD (X) processor
TypeRankCapacity (GB)Native speed
(MT/s)
RDIMMSingle8, 162666STD
RDIMMDual16, 322666STD
LRDIMMQuad642666STD
LRDIMMOctal1282666STD
Maximum memory capacity
The maximum memory capacity is a function of the DIMM capacity, number of installed DIMMs, memory
type, and number of installed processors.
Voltage
Memory options65
Maximum memory capacity for AMD processors
DIMM typeDIMM rankCapacity (GB)One processor
RDIMMSingle8128256
RDIMMSingle16256512
RDIMMDual16256512
RDIMMDual325121024
LRDIMMQuad6410242048
LRDIMMOctal12820484096
SmartMemory
HPE SmartMemory is included in each Gen10 ProLiant Server. Hewlett Packard Enterprise works closely
with leading memory device manufacturers in the development, qualification, and production testing of
DIMMs to ensure that they meet the tighter tolerances specified by Hewlett Packard Enterprise.
Memory subsystem architecture
The memory subsystem in this server is divided into channels. Each processor supports 8 channels, and
each channel supports 2 DIMM slots, as shown in the following table.
ChannelPopulation orderSlot number
(GB)
Two processors
(GB)
AG O10 9
BC K12 11
CE M14 13
DA I16 15
EH P7 8
FD L5 6
GF N3 4
HB J1 2
For the location of the slot numbers, see "DIMM slot locations on page 22."
This multi-channel architecture provides enhanced performance in Advanced ECC mode.
DIMM slots in this server are identified by number and by letter. Letters identify the population order. Slot
numbers indicate the DIMM slot ID for spare replacement.
DIMM ranks
To understand and configure memory protection modes properly, an understanding of DIMM rank is
helpful. Some DIMM configuration requirements are based on these classifications.
A single-rank DIMM has one set of memory chips that is accessed while writing to or reading from the
memory. A dual-rank DIMM is similar to having two single-rank DIMMs on the same module, with only
one rank accessible at a time. A quad-rank DIMM is, effectively, two dual-rank DIMMs on the same
module. Only one rank is accessible at a time. The server memory control subsystem selects the proper
rank within the DIMM when writing to or reading from the DIMM.
66 SmartMemory
Dual- and quad-rank DIMMs provide the greatest capacity with the existing memory technology. For
example, if current DRAM technology supports 8 GB single-rank DIMMs, a dual-rank DIMM would be 16
GB, and a quad-rank DIMM would be 32 GB, and an octal-rank LRDIMM would be 64 GB.
LRDIMMs are labeled as quad- and octal-rank DIMMs. There are four and eight ranks of DRAM on the
DIMM, but the LRDIMM buffer creates an abstraction that allows the DIMM to appear as a logical dualrank DIMM to the system. This is called Rank Multiplication. The LRDIMM buffer also isolates the
electrical loading of the DRAM from the system to allow for faster operation. These two changes allow the
system to support up to two LRDIMMs per memory channel, providing for greater memory capacity and
higher memory operating speed compared to quad-rank RDIMMs.
DIMM label identification
To determine DIMM characteristics, see the label attached to the DIMM. The information in this section
helps you to use the label to locate specific information about the DIMM.
ItemDescriptionDefinition
1Capacity
2Rank
3Data width on DRAM
4Memory generation
8 GB
16 GB
32 GB
64 GB
128 GB
1R = Single rank
2R = Dual rank
4R = Quad rank
8R = Octal rank
x4 = 4-bit
x8 = 8-bit
PC4 = DDR4
Table Continued
DIMM label identification67
ItemDescriptionDefinition
5Maximum memory speed
6CAS latency
7DIMM type
For more information about product features, specifications, options, configurations, and compatibility, see
the product QuickSpecs on the Hewlett Packard Enterprise website (http://www.hpe.com/info/qs).
Memory configurations
To optimize server availability, the server supports the following Advanced Memory Protection (AMP)
mode:
2133 MT/s
2400 MT/s
2666 MT/s
P = CAS 15-15-15
T = CAS 17-17-17
U = CAS 20-18-18
V = CAS 19-19-19 (for RDIMM, LRDIMM)
V = CAS 22-19-19 (for 3DS TSV LRDIMM)
R = RDIMM (registered)
L = LRDIMM (load reduced)
E = Unbuffered ECC (UDIMM)
Advanced ECC—Provides up to 4-bit error correction. This mode is the default option for this server.
AMP options are configured in the BIOS/Platform Configuration (RBSU). For more information, see the
UEFI System Utilities User Guide for HPE ProLiant Gen10 Servers and HPE Synergy on the Hewlett
Packard Enterprise website.
Advanced ECC memory configuration
Advanced ECC memory is the default memory protection mode for this server. Standard ECC can correct
single-bit memory errors and detect multi-bit memory errors. When multi-bit errors are detected using
Standard ECC, the error is signaled to the server and causes the server to halt.
Advanced ECC protects the server against some multi-bit memory errors. Advanced ECC can correct
both single-bit memory errors and 4-bit memory errors if all failed bits are on the same DRAM device on
the DIMM.
Advanced ECC provides additional protection over Standard ECC because it is possible to correct certain
memory errors that would otherwise be uncorrected and result in a server failure. Using HPE Advanced
Memory Error Detection technology, the server provides notification when a DIMM is degrading and has a
higher probability of uncorrectable memory error.
DIMM population information
For specific DIMM population information, see the DIMM population guidelines on the Hewlett Packard
Enterprise website (http://www.hpe.com/docs/memory-population-rules).
Identifying the processor type
The processor type installed in the server is briefly displayed during POST. To view this information and
additional processor specifications, do the following:
68 Memory configurations
Procedure
1. Reboot the server.
The server restarts and the POST screen appears.
2. Press F9.
The System Utilities screen appears.
3. Select System Information | Processor Information.
The Processor Information screen shows detailed information about the processors installed in the
server.
4. Press Esc until the main menu is displayed.
5. Select Reboot the System to exit the utility and resume the boot process.
Installing a DIMM
The server supports up to 32 DIMMs.
Prerequisites
Before installing this option, be sure you have the following:
The components included with the hardware option kit
For more information on specific options, see the server QuickSpecs on the Hewlett Packard Enterprise
website.
Procedure
1.Power down the server.
2.Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3.Do one of the following:
a. Extend the server from the rack.
b. Remove the server from the rack.
4.Remove the access panel.
5.Open the DIMM slot latches.
6.Install the DIMM.
Installing a DIMM69
7.Install the access panel.
8.Install the server in the rack.
9.Connect each power cord to the server.
10. Connect each power cord to the power source.
11. Power up the server.
Use the BIOS/Platform Configuration (RBSU) in the UEFI System Utilities to configure the memory mode.
For more information about LEDs and troubleshooting failed DIMMs, see "Systems Insight Display
combined LED descriptions."
Controller options
The server supports the following storage controllers:
•Embedded controllers
Enabled through System Utilities and configured through HPE Smart Storage Administrator (Intelligent
Provisioning)
•Type-a controllers
Type-a controllers install in the type-a smart array connector.
•Type-p controllers
Type-p controllers install in a PCIe expansion slot
Installing a storage controller
Prerequisites
Before installing this option, be sure that you have the following:
The components included with the hardware option kit
70 Controller options
Procedure
1. Power down the server.
2. Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3. Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack.
4. Remove the access panel.
5. Do one of the following:
•Remove the air baffle.
•If installed, remove the 4LFF midplane drive cage.
6. Do one of the following:
•For Type-a Smart Array controllers, install the controller into the Smart Array connector.
•For Type-p Smart Array controllers, install the controller into an expansion slot.
7. Cable the controller.
The installation is complete.
Hardware options installation71
Installing a Universal Media Bay
Prerequisites
Before installing this option, be sure that you have the following:
•The components included with the hardware option kit
•T-10 Torx screwdriver
Procedure
1.Power down the server.
2.Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3.Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack .
4.Remove the access panel.
5.Remove the air baffle.
6.Remove the fan cage.
7.Remove the bay blank.
8.Route the USB and video cables through the opening.
9.If installing a two-bay SFF front drive cage, install the drive cage.
10. Install the universal media bay.
72 Installing a Universal Media Bay
11. (Optional) Install the optical disk drive.
12. Connect the cables.
13. Install the fan cage.
14. Install the air baffle.
15. Install the access panel.
16. Slide the server into the rack.
17. Connect each power cord to the server.
18. Connect each power cord to the power source.
19. Power up the server.
The installation is complete.
Hardware options installation73
Drive cage options
Installing a front 8NVMe SSD Express Bay drive cage
Observe the following:
•The drive cage can be installed in any box. This procedure covers installing the drive cage in box 1.
•When installing in box 1, the NVMe riser must be installed in the tertiary PCIe slot.
•When installing in box 2, the NVMe riser must be installed in the secondary PCIe slot.
•When installing in box 3, the NVMe riser must be installed in the primary PCIe slot.
Prerequisites
An associated NVMe riser and high-performance fans are required when installing this option.
Procedure
1.Observe the following alerts.
WARNING:
To reduce the risk of personal injury from hot surfaces, allow the drives and the internal system
components to cool before touching them.
CAUTION:
To prevent damage to electrical components, properly ground the server before beginning any
installation procedure. Improper grounding can cause ESD.
2.Power down the server .
3.Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack .
4.Remove the access panel.
5.Remove the air baffle.
6.Remove the fan cage.
7.Remove the blank.
74 Drive cage options
8.Install the drive cage:
a. Remove all drives and drive blanks.
b. Install the drive cage.
9.Install the associated NVMe riser.
10. Connect the power cable to the drive backplane power connector.
11. Connect the data cables from the drive backplane to the NVMe riser.
12. Install drives or drive blanks.
The installation is complete.
Installing a front 6SFF SAS/SATA + 2NVMe Premium drive cage
The drive cage can be installed in any box. This procedure covers installing the drive cage in box 1.
Installing a front 6SFF SAS/SATA + 2NVMe Premium drive cage75
Prerequisites
A storage controller and high-performance fans are required when installing this drive cage.
Procedure
1.Observe the following alerts.
WARNING:
To reduce the risk of personal injury from hot surfaces, allow the drives and the internal system
components to cool before touching them.
CAUTION:
To prevent damage to electrical components, properly ground the server before beginning any
installation procedure. Improper grounding can cause ESD.
2.Power down the server .
3.Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack .
4.Remove the access panel.
5.Remove the air baffle.
6.Remove the fan cage.
7.Remove the blank.
8.Install the drive cage:
a. If drive blanks are installed in the drive cage assembly, remove the drive blanks. Retain the
drive blanks for use in empty drive bays.
b. Install the drive cage.
76Hardware options installation
9.Connect the power cable.
10. Install a storage controller.
11. Connect the data cables from the drive backplane to the controller.
12. Install drives or drive blanks.
The installation is complete.
Installing airflow labels
When an Express Bay drive cage is installed, airflow labels might be required:
Prerequisites
Before installing this option, be sure that you have the following:
The components included with the hardware option kit
Procedure
•If an eight-bay SFF drive cage is installed in box 1, then airflow labels are not required.
•If a blank is installed in box 1, replace it with the blank that comes with the kit.
Installing airflow labels77
•If a Universal Media Bay is installed in box 1, do one of the following:
◦If the 2SFF drive cage is not installed, then install airflow labels as shown.
◦If a 2SFF drive cage is installed, then install the airflow labels as shown.
78Hardware options installation
Installing a front 8SFF SAS/SATA drive cage in box 1
Prerequisites
Before installing this option, be sure that you have the following:
•T-10 Torx screwdriver
•The components included with the hardware option kit
Procedure
1.Power down the server.
2.Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3.Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack .
4.Remove the access panel.
5.Remove the air baffle.
6.Remove the fan cage.
7.Remove the bay blank.
Installing a front 8SFF SAS/SATA drive cage in box 179
8.Install the 8SFF front drive cage option.
9.Connect the power and data cables.
10. Install the fan cage.
11. Install the air baffle.
12. Install the access panel.
13. Slide the server into the rack.
14. Connect each power cord to the server.
15. Connect each power cord to the power source.
16. Power up the server.
The installation is complete.
80Hardware options installation
Installing a front 8SFF SAS/SATA drive cage in box 2
Procedure
1.Power down the server.
2.Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3.Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack .
4.Remove the access panel.
5.Remove the air baffle.
6.
Remove the fan cage.
7.Remove the bay blank.
8.Install the 8SFF front drive cage option.
Installing a front 8SFF SAS/SATA drive cage in box 281
9.Connect the power and data cables.
10. Install the fan cage.
11. Install the access panel.
12. Slide the server into the rack.
13. Connect each power cord to the server.
14. Connect each power cord to the power source.
15. Power up the server.
The installation is complete.
Installing a front 2SFF NVMe/SAS/SATA Premium drive cage
Prerequisites
Before installing this option, be sure that you have the following:
•T-10 Torx screwdriver
•The components included with the hardware option kit
•This installation requires a universal media bay.
Procedure
1.Power down the server.
2.Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3.Do one of the following:
82 Installing a front 2SFF NVMe/SAS/SATA Premium drive cage
•Extend the server from the rack.
•Remove the server from the rack .
4.Remove the access panel.
5.Remove the front bay blank.
6.Remove the optical disk drive tray from the universal media bay.
7.Remove the SFF drive blank from the universal media bay.
Hardware options installation83
8.Install the drive cage into the universal media bay.
9.Install the optical disk drive tray.
84Hardware options installation
10. Install the universal media bay.
11. Connect the power and data cables.
12. Install the access panel.
13. Slide the server into the rack.
14. Connect each power cord to the server.
15. Connect each power cord to the power source.
16. Power up the server.
The installation is complete.
Installing a midplane 4LFF SAS/SATA drive cage
Observe the following:
Installing a midplane 4LFF SAS/SATA drive cage85
•A 1U heatsink is required for each processor when installing this option.
•If you have a TPM, install it prior to this option.
•If you have a type-a controller, install it prior to this option.
Prerequisites
Before installing this option, be sure that you have the following:
The components included with the hardware option kit
Procedure
1.Power down the server.
2.Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3.Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack.
4.Remove the access panel.
5.Remove the air baffle.
The air baffle is no longer needed. The drive cage acts as an air baffle for the server.
6.Remove all riser cages.
86Hardware options installation
7.Connect the power cable to the drive backplane power connector on the system board.
8.If connecting the data cable to the system board or a controller, connect the data cable.
9.Prepare the drive cage for installation by lifting the latches on the drive cage.
10. Install the drive cage:
CAUTION:
Do not drop the drive cage on the system board. Dropping the drive cage on the system board
might damage the system or components. Remove all drives and use two hands when
installing or removing the drive cage.
a. Locate the alignment pins on the rear of the drive cage.
b. Align the pin on the rear left of the drive cage to the server and then insert the pin.
c. Gently lower the opposite side of the drive cage.
d. Pull the plunger pin on the rear right of the drive cage and then lower the drive cage until the
plunger pin engages.
Hardware options installation87
11. Install drives or drive blanks.
12. Push down on the latches to lower the drive cage into place.
13. Connect the power and data cables to the drive backplane.
The installation is complete.
Installing a rear 2SFF SAS/SATA drive cage in the primary or secondary
riser
Prerequisites
Before installing this option, be sure that you have the following:
•T-10 Torx screwdriver
•The components included with the hardware option kit
88 Installing a rear 2SFF SAS/SATA drive cage in the primary or secondary riser
•The front drive bays are fully populated with 12 LFF or 24 SFF drives.
•High performance fans are installed in all fan bays.
Procedure
1.Power down the server.
2.Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3.Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack .
4.Remove the access panel.
5.Do one of the following:
For primary bays, remove the riser cage.
For secondary bays, remove the rear wall blank.
Hardware options installation89
6.Install a SAS expander or other expansion card, if needed.
7.Install the drive cage.
8.Cable the drive backplane.
9.Install drives or drive blanks.
10. Install the access panel.
11. Slide the server into the rack.
12. Connect each power cord to the server.
13. Connect each power cord to the power source.
14. Power up the server.
The installation is complete.
90Hardware options installation
Installing a rear 2SFF SAS/SATA drive cage over the power supplies
Prerequisites
Before installing this option, be sure that you have the following:
•T-10 Torx screwdriver
•The components included with the hardware option kit
•The front bays are fully populated with 12 LFF or 24 SFF drives
•High performance fans are installed in all fan bays
Procedure
1.Power down the server.
2.Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3.Do one of the following:
•
Extend the server from the rack.
•Remove the server from the rack .
4.Remove the access panel.
5.Do one of the following:
If installed, remove the secondary riser cage.
Remove the secondary wall blank.
Installing a rear 2SFF SAS/SATA drive cage over the power supplies91
6.Remove the tertiary wall blank.
7.Install the drive cage compatible rear wall.
92Hardware options installation
8.Install the drive cage.
9.Install drives or drive blanks.
10. Install the secondary rear wall or a riser cage.
11. Cable the drive backplane.
12. Install the access panel.
13. Slide the server into the rack.
14. Connect each power cord to the server.
15. Connect each power cord to the power source.
16. Power up the server.
The installation is complete.
Hardware options installation93
Installing a rear 3LFF SAS/SATA drive cage
Before installing this option, the front bays must be fully populated with 12 LFF drives. High-performance
fans must be installed in all fan bays.
Prerequisites
Before installing this option, be sure that you have the following:
•T-10 Torx screwdriver
•The components included with the hardware option kit
Procedure
1.Power down the server.
2.Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
3.Do one of the following:
•
Extend the server from the rack.
•Remove the server from the rack .
4.Remove the access panel.
5.If installed, remove the secondary riser cage. The secondary riser cage is not supported with a three-
bay LFF rear configuration.
6.Remove the rear wall blank.
94 Installing a rear 3LFF SAS/SATA drive cage
7.Install the three-bay LFF rear drive cage option.
8.Install drives or drive blanks.
9.Connect the power and data cables.
10. Install the access panel.
11. Slide the server into the rack.
12. Connect each power cord to the server.
13. Connect each power cord to the power source.
14. Power up the server.
The installation is complete.
Riser and riser cage options
Riser and riser cage options95
Installing primary and secondary risers
Prerequisites
Before installing this option, be sure that you have the following:
•The components included with the hardware option kit
•T-10 Torx screwdriver
Procedure
1.Power down the server .
2.Do one of the following:
•Disconnect each power cord from the power source.
•Disconnect each power cord from the server.
3.Do one of the following:
•
Extend the server from the rack.
•Remove the server from the rack.
4.Remove the access panel.
5.Remove the riser cage.
6.Remove the riser board.
7.Install the riser.
96 Installing primary and secondary risers
8.Install any expansion boards, if needed
9.Connect data cables to the riser or expansion board, if needed.
10. Install the riser cage.
11. If needed, connect data cables to drive backlanes.
The installation is complete.
Installing tertiary risers
Prerequisites
Before installing this option, be sure that you have the following:
•The components included with the hardware option kit
•T-10 Torx screwdriver
•A tertiary riser cage is required to install this option.
Procedure
1.Power down the server .
2.Do one of the following:
•Disconnect each power cord from the power source.
•Disconnect each power cord from the server.
3.Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack.
4.Remove the access panel.
5.Remove the riser cage.
6.Install the riser.
Installing tertiary risers97
7.Install any expansion cards, if needed.
8.Connect any data cables to riser or expansion boards.
9.Install the tertiary riser cage.
10. Connect cables to drive backplanes, if needed.
The installation is complete.
Installing a secondary riser cage
Prerequisites
Before installing this option, be sure that you have the following:
•The components included with the hardware option kit
•T-10 Torx screwdriver
Procedure
1. Observe the following alert:
CAUTION:
To prevent damage to the server or expansion boards, power down the server and remove all
AC power cords before removing or installing the PCI riser cage.
2. Power down the server.
3. Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
4. Do one of the following:
98 Installing a secondary riser cage
•Extend the server from the rack.
•Remove the server from the rack.
5. Remove the access panel.
6. Remove the rear wall blank.
7. Install any expansion boards, if needed.
8. Install the riser cage:
The installation is complete.
Installing a tertiary riser cage
Prerequisites
Before installing this option, be sure that you have the following:
Installing a tertiary riser cage99
•The components included with the hardware option kit
•T-10 Torx screwdriver
Procedure
1. Observe the following alert.
CAUTION:
To prevent damage to the server or expansion boards, power down the server and remove all
AC power cords before removing or installing the PCI riser cage.
2. Power down the server.
3. Remove all power:
a. Disconnect each power cord from the power source.
b. Disconnect each power cord from the server.
4. Do one of the following:
•Extend the server from the rack.
•Remove the server from the rack.
5. Remove the access panel.
6. Remove the rear wall blanks.
100Hardware options installation
Loading...
+ hidden pages
You need points to download manuals.
1 point = 1 manual.
You can buy points or you can get point for every manual you upload.