viIBM Power Systems E870 and E880 Technical Overview and Introduction
Page 9
Draft Document for Review October 14, 2014 10:19 am5137spec.fm
Notices
This information was developed for products and services offered in the U.S.A.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area. Any
reference to an IBM product, program, or service is not intended to state or imply that only that IBM product,
program, or service may be used. Any functionally equivalent product, program, or service that does not
infringe any IBM intellectual property right may be used instead. However, it is the user's responsibility to
evaluate and verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not grant you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, Armonk, NY 10504-1785 U.S.A.
The following paragraph does not apply to the United Kingdom or any other country where such
provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION
PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR
IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT,
MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of
express or implied warranties in certain transactions, therefore, this statement may not apply to you.
This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.
Any references in this information to non-IBM websites are provided for convenience only and do not in any
manner serve as an endorsement of those websites. The materials at those websites are not part of the
materials for this IBM product and use of those websites is at your own risk.
IBM may use or distribute any of the information you supply in any way it believes appropriate without incurring
any obligation to you.
Any performance data contained herein was determined in a controlled environment. Therefore, the results
obtained in other operating environments may vary significantly. Some measurements may have been made
on development-level systems and there is no guarantee that these measurements will be the same on
generally available systems. Furthermore, some measurements may have been estimated through
extrapolation. Actual results may vary. Users of this document should verify the applicable data for their
specific environment.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.
This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to the names and addresses used by an actual business
enterprise is entirely coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the sample
programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore,
cannot guarantee or imply reliability, serviceability, or function of these programs.
5137spec.fmDraft Document for Review October 14, 2014 10:19 am
Trademarks
IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of International Business Machines
Corporation in the United States, other countries, or both. These and other IBM trademarked terms are
marked on their first occurrence in this information with the appropriate symbol (® or ™), indicating US
registered or common law trademarks owned by IBM at the time this information was published. Such
trademarks may also be registered or common law trademarks in other countries. A current list of IBM
trademarks is available on the Web at http://www.ibm.com/legal/copytrade.shtml
The following terms are trademarks of the International Business Machines Corporation in the United States,
other countries, or both:
Active Memory™
AIX®
Electronic Service Agent™
EnergyScale™
eServer™
Focal Point™
IBM®
Micro-Partitioning®
Power Architecture®
POWER Hypervisor™
Power Systems™
Power Systems Software™
POWER6+™
POWER6®
POWER7+™
POWER7®
PowerHA®
PowerPC®
PowerVM®
POWER®
pSeries®
Rational Team Concert™
Rational®
Redbooks®
Redpaper™
Redbooks (logo)®
RS/6000®
System p®
System x®
System z®
Tivoli®
The following terms are trademarks of other companies:
Intel Xeon, Intel, Intel logo, Intel Inside logo, and Intel Centrino logo are trademarks or registered trademarks
of Intel Corporation or its subsidiaries in the United States and other countries.
Linux is a trademark of Linus Torvalds in the United States, other countries, or both.
Microsoft, and the Windows logo are trademarks of Microsoft Corporation in the United States, other
countries, or both.
Java, and all Java-based trademarks and logos are trademarks or registered trademarks of Oracle and/or its
affiliates.
UNIX is a registered trademark of The Open Group in the United States and other countries.
Other company, product, or service names may be trademarks or service marks of others.
viiiIBM Power Systems E870 and E880 Technical Overview and Introduction
Page 11
Draft Document for Review October 14, 2014 10:19 am5137pref.fm
Preface
This IBM® Redpaper™ publication is a comprehensive guide covering the IBM Power System
E870 (9119-MME) and IBM Power System E880 (9119-MHE) servers that support IBM AIX®,
IBM i, and Linux operating systems. The objective of this paper is to introduce the major
innovative Power E870 and Power E880 offerings and their relevant functions:
The new IBM POWER8 processor, available at frequencies of 4.024 GHz, 4.190 GHz, and
4.350 GHz
Significantly strengthened cores and larger caches
Two integrated memory controllers with improved latency and bandwidth
Integrated I/O subsystem and hot-pluggable PCIe Gen3 I/O slots
Improved reliability, serviceability, and availability (RAS) functions
IBM EnergyScale™ technology that provides features such as power trending,
power-saving, capping of power, and thermal measurement
This publication is for professionals who want to acquire a better understanding of IBM Power
Systems™ products. The intended audience includes the following roles:
Authors
Clients
Sales and marketing professionals
Technical support professionals
IBM Business Partners
Independent software vendors
This paper expands the current set of IBM Power Systems documentation by providing a
desktop reference that offers a detailed technical description of the Power E870 and
Power E880 systems.
This paper does not replace the latest marketing materials and configuration tools. It is
intended as an additional source of information that, together with existing sources, can be
used to enhance your knowledge of IBM server solutions.
This paper was produced by a team of specialists from around the world working at the
International Technical Support Organization, Austin Center.
Alexandre Bicas Caldeira is a Certified IT Specialist and is a member of the Power Systems
Advanced Technical Sales Support team for IBM Brazil. He holds a degree in Computer
Science from the Universidade Estadual Paulista (UNESP) and an MBA in Marketing. His
major areas of focus are competition, sales, and technical sales support. Alexandre has more
than 14 years of experience working on IBM Systems & Technology Group Solutions and has
worked also as an IBM Business Partner on Power Systems hardware, AIX, and IBM
PowerVM® virtualization products.
YoungHoon Cho is a Power Systems Top Gun with the post-sales Technical Support Team
for IBM in Korea. He has over 10 years of experience working on RS/6000®, System p®, and
Power Systems products. He provides second-line technical support to Field Engineers
working on Power Systems and system management.
5137pref.fmDraft Document for Review October 14, 2014 10:19 am
James Cruickshank works in the Power System Client Technical Specialist team for IBM in
the UK. He holds an honors degree in Mathematics from the University of Leeds. James has
over 13 years experience working with RS/6000, pSeries®, System p and Power Systems
products. James supports customers in the financial services sector in the UK.
Bartłomiej Grabowski is a Principal System Support Specialist in DHL IT Services in the
Czech Republic. He holds a Bachelor’s degree in Computer Science from the Academy of
Computer Science and Management in Bielsko-Biala. His areas of expertise include IBM i,
PowerHA® solutions that are based on hardware and software replication, Power Systems
hardware, and PowerVM. He is an IBM Certified Systems Expert and a coauthor of several
IBM Redbooks® publications.
The project that produced this publication was managed by:
Scott Vetter
Executive Project Manager, PMP
Thanks to the following people for their contributions to this project:
Tamikia Barrow, Ella Buslovich, Mark Clark, Volker Haug, Daniel Henderson,
Stephanie Jensen, Bob Kovacks, Jai Lei Ma, Chris Mann, Dwayne Moore, Mark Olson,
Monica Sanchez, Bill Starke, Jeff Stuecheli, Doug Szerdi, Jacobo Vargas, Steve Will
IBM
Now you can become a published author, too!
Here’s an opportunity to spotlight your skills, grow your career, and become a published
author—all at the same time! Join an ITSO residency project and help write a book in your
area of expertise, while honing your experience using leading-edge technologies. Your efforts
will help to increase product acceptance and customer satisfaction, as you expand your
network of technical contacts and relationships. Residencies run from two to six weeks in
length, and you can participate either in person or as a remote resident working from your
home base.
Find out more about the residency program, browse the residency index, and apply online at:
ibm.com/redbooks/residencies.html
Comments welcome
Your comments are important to us!
We want our papers to be as helpful as possible. Send us your comments about this paper or
other IBM Redbooks publications in one of the following ways:
Use the online Contact us review Redbooks form found at:
ibm.com/redbooks
Send your comments in an email to:
redbooks@us.ibm.com
Mail your comments to:
IBM Corporation, International Technical Support Organization
Dept. HYTD Mail Station P099
xIBM Power Systems E870 and E880 Technical Overview and Introduction
Page 13
Draft Document for Review October 14, 2014 10:19 am5137pref.fm
2455 South Road
Poughkeepsie, NY 12601-5400
Stay connected to IBM Redbooks
Find us on Facebook:
http://www.facebook.com/IBMRedbooks
Follow us on Twitter:
http://twitter.com/ibmredbooks
Look for us on LinkedIn:
http://www.linkedin.com/groups?home=&gid=2130806
Explore new Redbooks publications, residencies, and workshops with the IBM Redbooks
Stay current on recent Redbooks publications with RSS Feeds:
http://www.redbooks.ibm.com/rss.html
Preface xi
Page 14
5137pref.fmDraft Document for Review October 14, 2014 10:19 am
xiiIBM Power Systems E870 and E880 Technical Overview and Introduction
Page 15
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
1
Chapter 1.General description
The IBM Power System E870 (9119-MME) and IBM Power System E880 (9119-MHE)
servers use the latest POWER8 processor technology that is designed to deliver
unprecedented performance, scalability, reliability, and manageability for demanding
commercial workloads.
The Power E870 is a highly-scalable rack-mount system optimized for running AIX, IBM i, and
Linux workloads. The Power E870 is a modular-built system and uses one or two system
nodes together with a system control unit. Each system node is five EIA-units tall (5U) and the
system control unit is two EIA-units (2U) tall. The Power E870 is housed in a 19-inch rack.
The Power E880 is a highly-scalable rack-mount system optimized for running AIX, IBM i, and
Linux workloads. The Power E880 is a modular-built system. In 2014, the system is built of
one or two system nodes together with a system control unit. Each system node is five
EIA-units tall (5U) and the system control unit is two EIA-units (2U) tall. The Power E880 is
housed in a 19-inch rack.
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
1.1 Systems overview
The following sections provide detailed information about the Power E870 and Power E880
systems.
1.1.1 Power E870 server
The Power E870 (9119-MME) server is a powerful POWER8-based system that scales up to
eight sockets. Each socket contains a single 8-core or 10-core POWER8 processor. Thus, a
fully-configured Power E870 can scale up to 64 or 80 cores.
The Power E870 is a modular system that is built from a combination of the following four
building blocks:
System control unit
System node
PCIe Gen3 I/O expansion drawer
EXP24S SFF Gen2-bay drawer
The system control unit provides redundant system master clock and redundant system
master Flexible Service Processor (FSP) and support for the Operator Panel, the system
VPD, and the base DVD. The system control unit provides clock signals to the system nodes
with semi-flexible connectors.
Each system node provides four processor sockets for POWER8 processors and 32 CDIMM
slots supporting a maximum of 8 memory features. Using the 256 GB memory features, the
system node can support a maximum of 2 TB of RAM. Thus a fully-configured Power E870
can support up to 4 TB of RAM. Each system node provides eight PCIe Gen3 x16 slots.
Each optional 19-inch PCIe Gen3 4U I/O Expansion Drawer provides 12 PCIe Gen 3 slots.
The I/O expansion drawer connects to the system unit with a PCIe X16 to optical CXP
converter adapter housed in the system unit.
Each EXP24S SFF Gen2-bay Drawer provides 24 x 2.5-inch form-factor (SFF) SAS bays.
The EXP24S is connected to the Power E870 server using a SAS adapter in a PCIe slot in a
system node or in a I/O expansion drawer.
2IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 17
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
Figure 1-1 shows a single system node Power E870 in a T42 rack with two PCIe I/O drawers
and an EXP24S disk drawer.
Figure 1-1 A Power E870 in a T42 rack
1.1.2 Power E880 server
The Power E880 (9119-MHE) server is a powerful POWER8-based system that scales up to
16 sockets. Each socket contains a single 8-core POWER8 processor. Thus, a
fully-configured Power E880 can scale up to 128 cores. In 2014 up to two system nodes per
server are supported or a maximum of 64-cores.
The Power E880 is a modular system that is built from a combination of the following four
building blocks:
System control unit
System node
PCIe Gen3 I/O expansion drawer
EXP24S SFF Gen2-bay drawer
The system control unit provides redundant system master clock and redundant system
master Flexible Service Processor (FSP) and support for the Operator Panel, the system
VPD, and the base DVD.
Each system node provides four processor sockets for POWER8 processors and 32 CDIMM
slots supporting a maximum of 8 memory features. Using the 512 GB memory features, the
system node can support a maximum of 4 TB of RAM. Thus a two-node Power E880 can
support up to 8 TB of memory. A four-node Power E880 can support up to 16 TB of RAM.
Each system node provides eight PCIe Gen3 x16 slots.
Chapter 1. General description 3
Page 18
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
Each optional 19-inch PCIe Gen3 4U I/O Expansion Drawer provides twelve PCIe Gen 3
slots. The I/O expansion drawer connects to the system unit with a PCIe X16 to Optical CXP
converter adapter housed in the system unit.
Each EXP24S SFF Gen2-bay Drawer provides 24 x 2.5-inch form-factor (SFF) SAS bays.
The EXP24S is connected to the Power E880 server using a SAS adapter in a PCIe slot in a
system node or in a I/O expansion drawer.
Figure 1-2 shows a four system node Power E880 with two PCIe I/O drawers and an EXP24S
disk drawer.
Figure 1-2 A Power E880 in a T42 rack
1.1.3 System Control Unit
The system control unit (SCU) in a Power E870 and E880 is a new innovation to increase the
reliability, availability, and serviceability of the servers. The 2 EIA rack unit (U) unit provides
redundant clock and service processor capability to Power E870 and E880 systems even if
they have only one system unit installed. The SCU also provides a DVD option connected to a
PCIe adapter with a USB cable.
The SCU is powered from the system nodes using cables plugged into the first and second
system nodes in a two-, three-, or four-system node server. The SCU is powered from the
single system node in servers with only one.
The SCU provides redundant clock signalling to the system nodes using semi-flexible
connectors. These connectors are located at the rear of the system and route up and down
the middle of the system nodes. In this way they do not cause any restrictions on the allowed
rack specification.
4IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 19
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
Power connections
Backup
battery
Clock
connections
DVD USB
connection
FSP
connections
HMC
connections
FSP
connections
HMC
connections
DVD Drive
Operator panel
The SCU provides redundant service processor function to the system nodes using flexible
service processor (FSP) cables. Each system node has two FSP connections to the system
control unit.
The SCU provides redundant connections to one or two HMCs using RJ45 Ethernet
connections.
Figure 1-3 shows the front and rear view of a system control unit. The locations of the
connectors and features are indicated.
Figure 1-3 Front and rear view of the system control unit
1.1.4 System nodes
The system nodes in the Power E870 and E880 servers host the processors, memory, PCIe
slots, and power supplies for the system. Each system node is 5U high and connects to the
system control unit with FSP, clock, and power connectors. The system node connects to
other system nodes using SMP connectors.
Each system node in a Power E870 or E880 server provides four POWER8 processors, 32
CDIMM slots, and eight PCIe Gen3 x16 slots.
Chapter 1. General description 5
Page 20
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
PCIe Gen3 x16 slots
Clock connectors
Local clock and control units
Power sockets
FSP connector
SCU power
FSP connector
SCU power
SMP connectors x6
Power socketsPower sockets
Figure 1-4 shows the front view of the system node.
Figure 1-4 Front view of the system node
Figure 1-5 shows the rear view of the system node with notable features highlighted.
Figure 1-5 Rear view of the system node
6IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 21
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
Unlike the inter-node flex connectors on the Power 770 and Power 780 servers, the SMP
cables on the Power E870 and E880 servers are fully flexible. These cables do not impose
restrictions on allowed rack specification. Figure 1-6 shows a diagram of how SMP cables can
be routed within a rack on a Power E880 with all four system nodes installed.
Figure 1-6 An example of SMP cable routing in a four-drawer system
Figure 1-7 shows SMP cable routing in a two-drawer Power E870 and E880.
Figure 1-7 SMP cable routing in a two-drawer Power E870 and E880.
Chapter 1. General description 7
Page 22
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
1.1.5 I/O drawers
I/O drawers to provide additional PCIe adapter slots and SAS disk slots are available for the
Power E870 and E880. More details can be found in section 1.6, “I/O drawers” on page 17.
1.2 Operating environment
Table 1-1details the operating environment for the Power E870 and E880 servers.
Table 1-1 Operating environment for Power E870 and Power E880
Power E870 and Power E880 operating environment
DescriptionOperatingNon-operating
Power E870Power E880Power E870Power E880
Temperature5 to 35 degrees C (41 to 95 F)5 to 45 degrees C (41 to 114F)
Relative humidity20 - 80%8 - 80%
Maximum dew
point
Operating voltage200 to 240 V ACN/A
Operating
frequency
Maximum power
consumption
Maximum power
source loading
Maximum
thermal output
Maximum altitude3,048 m (10,000 ft.)N/A
Noise levelTBDN/A
1.3 Physical package
Table 1-2 on page 9 lists the physical dimensions of the system control unit and of individual
system nodes. Both servers are available only in a rack-mounted form factor.
29 degrees C (84 F)28 degrees C (82 F)
47 - 63 HzN/A
4150 Watts per enclosureN/A
4.2 kVA per enclosureN/A
14,164 BTU/hour per enclosureN/A
The Power E870 is a modular system that can be constructed from a single system control
unit and one or two building-block enclosures.
The Power E880 is a modular system that can be constructed from a single system control
unit and one, two, three, or four building-block enclosures. In 2014, up to two system nodes
per server are supported.
The system control unit requires 2U and each system node requires 5U. Thus, a
single-enclosure system requires 7U, a two-enclosure system requires 12U, a
three-enclosure system requires 17U, and a four-enclosure system requires 22U.
8IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 23
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
Table 1-2 Physical dimensions of the Power E870 and Power E870
DimensionSystem control unitPower E870 or
Power E880
enclosure
Width437 mm (17.2 in)445 mm (17.5 in)434 mm (17.1 in)
Depth784 mm (30.9 in)902 mm (35.5 in)813 mm (32.0 in)
PCIe I/O Expansion
drawer
Height88.9 mm (3.5 in) 2 EIA
units
Weight23.68 kg (52 lb)75.7 kg (167 lb)23.6 kg (52 lb)
219 mm (8.6 in) 5 EIA
units
86 mm (3.4 in) 2 EIA
units
Figure 1-8 shows a picture of a Power E870 system control unit and system node from the
front.
Figure 1-8 Power 870 system control unit and system node
1.4 System features
Both Power E870 and Power E880 system node contain four processor modules with 512 KB
L2 cache and 8 MB L3 cache per core.
1.4.1 Power E870 system features
The following features are available on the Power E870:
One or two 5U 19-inch rack mount system node drawers
One 2U 19-inch rack-mount system control unit drawer
Chapter 1. General description 9
Page 24
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
12U for a system with two system node drawers
One processor feature per system node. All system nodes must have the same feature:
– 4.024 GHz, (4 X 0/8-core) 32-core POWER8 processor (#EPBA)
– 4.190 GHz, (4 X 0/10-core) 40-core POWER8 processor (#EPBC)
Static or mobile processor activation features available on a per core basis
POWER8 DDR3 memory CDIMMs (32 CDIMM slots per system node, 16 sites populated
per system node minimum):
– 0/64 GB (4 X 16 GB), 1600 MHz (#EM8J)
– 0/128 GB (4 X 32 GB), 1600 MHz (#EM8K)
– 0/256 GB (4 X 64 GB), 1600 MHz (#EM8L)
Active Memory™ Expansion - optimized onto the processor chip (#EM82)
Elastic CoD Processor Day - no-charge, temporary usage of inactive, Elastic CoD
resources on initial orders (#EPJ3 or #EPJ5)
90 Days Elastic CoD Temporary Processor Enablement (#EP9T)
Eight PCIe Gen3 x16 I/O expansion slots per system node drawer (maximum 16 with
2-drawer system)
One slim-line, SATA media bay per system control unit enclosure (DVD drive selected as
default with the option to de-select)
Redundant hot-swap AC power supplies in each system node drawer
Two HMC ports per Flexible Service Processor (FSP) in the system control unit
Dynamic logical partition (LPAR) support
Processor and memory Capacity Upgrade on Demand (CUoD)
PowerVM virtualization:
– Micro-Partitioning®
– Dynamic logical partitioning
– Shared processor pools
– Shared storage pools
– Live Partition Mobility
– Active Memory Sharing
– Active Memory Deduplication
– NPIV support
– PowerVP Performance Monitor
Optional PowerHA for AIX, IBM i, and Linux
Optional PCIe Gen3 I/O Expansion Drawer with PCIe Gen3 slots:
– Zero or two PCIe Gen3 Drawers per system node drawer (#EMX0)
– Each Gen3 I/O drawer holds two 6-slot PCIe3 Fan-out Modules (#EMXF)
– Each Gen3 I/O drawer attaches to the system node via two PCIe3 Optical Cable
Adapters (#EJ07)
10IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 25
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
1.4.2 Power E880 system features
The following features are available on the Power E880:
One or two 5U 19-inch rack-mount system node drawers
One 2U 19-inch rack-mount system control unit drawer
12U for a system with two system node drawers
One processor feature per system node:
– 4.35 GHz, (4 X 0/8-core) 32-core POWER8 processor (#EPBB)
Static or mobile processor activation features available on a per core basis
POWER8 DDR3 memory CDIMMs (32 CDIMM slots per system node, 16 sites populated
per system node minimum):
– 0/64 GB (4 X 16 GB), 1600 MHz (#EM8J)
– 0/128 GB (4 X 32 GB), 1600 MHz (#EM8K)
– 0/256 GB (4 X 64 GB), 1600 MHz (#EM8L)
– 0/512 GB (4 X 128 GB), 1600 MHz (#EM8M)
Active Memory Expansion - optimized onto the processor chip (#EM82)
90 Days Elastic CoD Temporary Processor Enablement (#EP9T)
Eight PCIe Gen3 x16 I/O expansion slots per system node drawer (maximum 16 with
2-drawer system)
One slim-line, SATA media bay per system control unit enclosure (DVD drive defaulted on
order, option to de-select)
Redundant hot-swap ac power supplies in each system node drawer
Two HMC ports per Flexible Service Processor (FSP) in the system control unit
Dynamic logical partition (LPAR) support
Processor and memory CUoD
PowerVM virtualization:
– Micro-Partitioning
– Dynamic logical partitioning
– Shared processor pools
– Shared storage pools
– Live Partition Mobility
– Active Memory Sharing
– Active Memory Deduplication
– NPIV support
– PowerVP Performance Monitor
Optional PowerHA for AIX, IBM i, and Linux
Optional PCIe Gen3 I/O Expansion Drawer with PCIe Gen3 slots:
– Zero or two PCIe Gen3 Drawers per system node drawer (#EMX0)
– Each Gen3 I/O drawer holds two 6-slot PCIe3 Fan-out Modules (#EMXF)
– Each Gen3 I/O drawer attaches to the system node via two PCIe3 Optical Cable
Adapters (#EJ07)
Chapter 1. General description 11
Page 26
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
1.4.3 Minimum features
Each Power E870 or Power E880 initial order must include a minimum of the following items:
1 x processor module
– 4.024 GHz, 32-core POWER8 processor module (#EPBA) for Power E870
– 4.190 GHz, 40-core POWER8 processor module (#EPBC) for Power E870
– 4.35 GHz, 32-core POWER8 processor module (#EPBB) for Power E880
8 x 1 core processor activation
– 1 core Processor Activation for #EPBA (#EPBJ) for Power E870
– 1 core Processor Activation for #EPBC (#EPBL) for Power E870
– 1 core Processor Activation for #EPBB (#EPBK) for Power E880
4 x 64 GB (4 x 16 GB) CDIMMs, 1600 MHz, 4 Gb DDR3 DRAM (#EM8J)
50% of installed memory must be activated using features:
1 x 5U system node drawer (#EBA0)
2 x Service Processor (#EU0A)
1 x Load Source Specify
– EXP24S SFF Gen2 (#5887 or #EL1S) Load Source Specify (#0728)
– SAN Load Source Specify (#0837)
1 x Rack-mount Drawer Bezel and Hardware
– IBM Rack-mount Drawer Bezel and Hardware (#EBA2) for Power E870
– IBM Rack-mount Drawer Bezel and Hardware (#EBA3) for Power E880
– OEM Rack-mount Drawer Bezel and Hardware (#EBA4)
1 x System Node to System Control Unit Cable Set for Drawer 1 (#ECCA)
1 x AC Power Chunnels for routing power cables from back of machine to front (#EBAA)
1 x Language Group Specify (#9300/#97xx)
1 x Primary Operating System Indicator
– Primary Operating System Indicator - AIX (#2146)
– Primary Operating System Indicator - Linux (#2147)
Optional 1 x Media
– SATA Slimline DVD-RAM with write cache (#EU13)
– PCIe2 LP 4-Port USB 3.0 Adapter (#EC45)
– PCIe2 4-Port USB 3.0 Adapter (#EC46)
When either AIX or Linux are the primary operating systems, the order must include a
minimum of the following items also:
1 x Primary Operating System Indicator
– Primary Operating System Indicator - AIX (#2146)
– Primary Operating System Indicator - Linux (#2147)
12IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 27
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
When IBM i is the primary operating system, the order must include a minimum of the
following items:
1 x Specify Code
– Mirrored System Disk Level, Specify Code (#0040)
– Device Parity Protection-All, Specify Code (#0041)
– Mirrored System Bus Level, Specify Code (#0043)
– Device Parity RAID 6 All, Specify Code (#0047)
– Mirrored Level System Specify Code (#0308)
1 x System Console
– Sys Console On HMC (#5550)
– System Console-Ethernet No IOP (#5557)
1 x Primary Operating System Indicator - IBM i (#2145)
Note:
Additional optional features can be added, as desired.
IBM i systems require a DVD to be available to the system when required. This DVD
can be located in the system control unit (DVD feature #EU13) or it can be located
externally in an enclosure like the 7226-1U3. A USB PCIe adapter such as #EC45 is
required for #EU13. A SAS PCIe adapter such as #EJ11 is required to attach a SATA
DVD in the 7226-1U3. A virtual media repository may be used to substitute for a DVD
device.
Feature-coded racks are allowed for I/O expansion only.
A machine type/model rack, if desired, should be ordered as the primary rack.
A minimum number of eight processor activations must be ordered per system.
A minimum of four memory features per drawer are required.
At least 50% of available memory must be activated through a combination of feature
#EMA5 and #EMA6, and #EMA9.
Memory sizes can differ across CPU modules, but the eight CDIMM slots connected to
the same processor module must be filled with identical CDIMMs.
If SAN Load Source Specify (#0837) is ordered #0040, #0041, #0043, #0047, #0308
are not supported.
The language group is auto-selected based on geographic rules.
No feature codes are assigned for the following:
– Four AC power supplies are delivered as part of the system node. No features are
assigned to power supplies. Four line cords are auto-selected according to
geographic rules.
– Two default AC PDU to wall cables are included. No features are assigned. Cables
are auto-selected according to geographic rules.
– There must be one system control unit on each system. The system control unit is
considered the system with the serial number.
One HMC is required for every Power E870 or Power E880; however, a communal HMC
is acceptable. HMCs supported on POWER8 hardware are 7042-C08 and 7042-CR5
through 7042-CR8.
Chapter 1. General description 13
Page 28
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
1.4.4 Power supply features
This section describes how the system nodes and system control units are powered.
System node power
Four AC power supplies are required for each system node enclosure. This arrengement
provides 2+2 redundant power with dual power sources for enhanced system availability. A
failed power supply can be hot swapped but must remain in the system until the replacement
power supply is available for exchange.
Four AC power cords are used for each system node (one per power supply) and are ordered
using the AC Power Chunnel feature (#EBAA). The chunnel carries power from the rear of the
system node to the hot swap power supplies located in the front of the system node where
they are more accessible for service.
System control unit power
The system control unit is powered from the system nodes. UPIC cables provide redundant
power to the system control unit. Two UPIC cables attach to system node drawer 1 and two
UPIC cables attach to system node drawer 2. They are ordered with #ECCA and #ECCB. The
UPIC cords provides N+1 redundant power to the system control unit.
1.4.5 Processor card features
Each system must have a minimum of four active processors. Each processor card feature
will deliver a set of four identical processors. All processor cards in the system must be
identical. Cable features are required to connect system nodes to the system control unit and
to other system nodes.
For a single system node configuration, #ECCA is required. #ECCA provides cables to
connect the system node with the system control unit.
For a dual system node configuration, #ECCB is required. #ECCB provides cables to
connect the system nodes with the system control unit and cables to connect the two
system nodes.
The Power E870 has two types of processor cards, offering the following features:
4.024 GHz, (4 X 0/8-core) 32-core POWER8 processor (#EPBA)
4.190 GHz, (4 X 0/10-core) 40-core POWER8 processor (#EPBC)
The Power E880 has one type of processor card, offering the following features:
4.35 GHz, (4 X 0/8-core) 32-core POWER8 processor (#EPBB)
Several types of capacity on demand (CoD) processor options are available on the
Power E870 and Power 880 servers to help meet changing resource requirements in an
on-demand environment by using resources installed on the system but not activated. CoD
allows you to purchase additional permanent processor or memory capacity and dynamically
activate it when needed. The #EPJ3 provide no-charge elastic processor days for
Power E880.
1.4.6 Summary of processor features
Table 1-3 on page 15 summarizes the processor feature codes for the Power E870.
14IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 29
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
Table 1-3 Summary of processor features for the Power E870
Feature CodeDescription
EPBA4.02 GHz, 32-core POWER8 processor
EPBC4.19 GHz, 40-core POWER8 processor
EB2RSingle 5250 Enterprise Enablement, IBM i
EB30Full 5250 Enterprise Enablement, IBM i
EP2S1-Core Mobile Activation
EP2U1-Core Mobile Activation from POWER 7
EP9T90 Days Elastic CoD Processor Core Enablement
EPBJ1 core Processor Activation for #EPBA
EPBL1 core Processor Activation for #EPBC
EPBN1 core Processor Activation for #EPBA, Mobile Enabled
EPBQ1 core Processor Activation for #EPBC, Mobile Enabled
EPJ61 Proc-Day Elastic CoD Billing for #EPBA, AIX/Linux
EPJ71 Proc-Day Elastic CoD Billing for #EPBA, IBM i
EPJ8100 Elastic CoD Proc-Days of Billing for Processor #EPBA. AIX/Linux
EPJ9100 Elastic CoD Proc-Days of Billing for Processor #EPBA. IBM i
EPJAProc CoD Utility Billing, 100 Proc-mins. for #EPBA, AIX/Linux
EPJBProc CoD Utility Billing, 100 Proc-mins. for #EPBA, IBM i
EPJJ1 Proc-Day Elastic CoD Billing for #EPBC, AIX/Linux
EPJK1 Proc-Day Elastic CoD Billing for #EPBC, IBM i
EPJL100 Elastic CoD Proc-Days of Billing for Processor #EPBC. AIX/Linux
EPJM100 Elastic CoD Proc-Days of Billing for Processor #EPBC. IBM i
EPJNProc CoD Utility Billing, 100 Proc-mins. for #EPBC, AIX/Linux
EPJPProc CoD Utility Billing, 100 Proc-mins. for #EPBC, IBM i
Table 1-4 summarizes the processor feature codes for the Power E880.
Table 1-4 Summary of processor features for the Power E880
Feature CodeDescription
EPBB4.35 GHz, 32-core POWER8 processor
EB2RSingle 5250 Enterprise Enablement
EB30Full 5250 Enterprise Enablement
EP2T1-Core Mobile Activation
EP2V1-Core Mobile Activation from POWER 7
EP9T90 Days Elastic CoD Processor Core Enablement
EPBK1 core Processor Activation for #EPBB
Chapter 1. General description 15
Page 30
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
Feature CodeDescription
EPBP1 core Processor Activation for #EPBB, Mobile Enabled
EPJ348 Proc-Days of Elastic CoD Processor Resources
EPJC1 Proc-Day Elastic CoD Billing for #EPBB, AIX/Linux
EPJD1 Proc-Day Elastic CoD Billing for #EPBB, IBM i
EPJE100 Elastic CoD Proc-Days of Billing for Processor #EPBB. AIX/Linux
EPJF100 Elastic CoD Proc-Days of Billing for Processor #EPBB. IBM i
EPJGProc CoD Utility Billing, 100 Proc-mins. for #EPBB, AIX/Linux
EPJHProc CoD Utility Billing, 100 Proc-mins. for #EPBB, IBM i
1.4.7 Memory features
1600 MHz memory CDIMMs are available as 64 GB (#EM8J), 128 GB (#EM8K), 256 GB
(#EM8L), and 512 GB (#EM8M) memory features. Each memory feature provides four
CDIMMs. CDIMMs are custom DIMMs which enhance memory performance and memory
reliability. Each system node has 32 CDIMM slots which support a maximum of eight memory
features. Memory activations of 50% of the installed capacity are required. The feature
#EMJ8 provides no-charge elastic memory days for the Power E880.
Table 1-5 lists memory features that are available for the Power E870 and E880.
EMB6Bundle of eight #EM8M, 512 GB 1600 MHz Memory features
EM9T90 Days Elastic CoD Memory Enablement
EMA51 GB Memory Activation
EMA6Quantity of 100 1 GB Memory Activations (#EMA5)
EMA7100 days of 1 GB Mobile Memory Activations
EMA9100 days of 1 GB Mobile Enabled Memory Activations
EMB7Bundle of forty, #EMA6 1600 MHz memory features
EMB8Bundle of 512 #EMA5 1600 MHz memory features
EMJ41 GB-Day billing for Elastic CoD memory
EMJ5100 GB-Day billing for Elastic CoD memory
EMJ6999 GB-Day billing for Elastic CoD memory
EMJ8384 GB-Days of Elastic CoD Memory Resources (E880 only)
16IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 31
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
1.4.8 System node PCIe slots
Each Power E870 or Power E880 system node enclosure provides eight half-length, half-high
x16 PCIe Gen3 slots. These PCIe slots can be used for either low profile PCIe adapters or for
attaching a PCIe I/O drawer.
A new form factor blind swap cassette (BSC) is used to house the low profile adapters which
go into these slots.
PCIe Gen1, Gen2, and Gen3 adapter cards are supported in these Gen3 slots.
Table 1-6 provides details of the PCI slots in the Power E870 and Power E880 system.
Table 1-6 PCIe slot locations and descriptions for the Power E870 and Power E880
All slots support enhanced error handling (EEH).
All PCIe slots are hot swappable and support concurrent maintenance.
1.5 Disk and media features
A system node does not support internal disks. Any required disk must reside within a SAN
disk subsystem or an external disk drawer. At the time of writing, the EXP24S SFF Gen2-bay
Drawer (#5887) is the only supported disk drawer for Power E870 or Power E880.
Each system control unit enclosure has one slim-line bay which can support one DVD drive
(#EU13). The #EU13 DVD is cabled to a USB PCIe adapter located in either a system node
or in a PCIe Gen3 I/O drawer. A USB to SATA converter is included in the configuration
without a separate feature code.
For IBM i, a DVD drive must be available on the server when required. The DVD can be in the
system control unit or it can be located in an external enclosure such as a 7226-1U3
Multimedia drawer.
1.6 I/O drawers
If additional Gen3 PCIe slots beyond the system node slots are required, a system node x16
slot is used to attach a 6-slot expansion module in the I/O Drawer. A PCIe Gen3 I/O
Chapter 1. General description 17
Page 32
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
expansion drawer (#EMX0) holds two expansion modules which are attached to any two x16
PCIe slots in the same system node or in different system nodes.
Disk-only I/O drawers (#5887) are also supported, providing storage capacity.
1.6.1 PCIe Gen3 I/O expansion drawer
The 19-inch 4 EIA (4U) PCIe Gen3 I/O expansion drawer (#EMX0) and two PCIe FanOut
Modules (#EMXF) provide twelve PCIe I/O full-length, full-height slots. One FanOut Module
provides six PCIe slots labeled C1 through C6. C1 and C4 are x16 slots and C2, C3, C5, and
C6 are x8 slots. PCIe Gen1, Gen2, and Gen3 full-high adapter cards are supported.
A blind swap cassette (BSC) is used to house the full-high adapters that go into these slots.
The BSC is the same BSC as used with the previous generation server's 12X attached I/O
drawers (#5802, #5803, #5877, #5873). The drawer is shipped with a full set of BSC, even if
the BSC is empty.
Concurrent repair and add/removal of PCIe adapter cards is done by HMC guided menus or
by operating system support utilities.
A PCIe X16 to Optical CXP converter adapter (#EJ07) and 2.0 M (#ECC6) or 10.0 M
(#ECC8) CXP 16X Active Optical cables (AOC) connect the system node to a PCIe FanOut
module in the I/O expansion drawer. One feature #ECC6 or one #ECC8 ships two AOC
cables. Each PCIe Gen3 I/O expansion drawer has two power supplies.
In 2014, a system node supports either zero or two PCIe Gen3 I/O expansion drawers.
Connecting a single drawer to a system node is not supported.
Figure 1-9 shows a PCIe Gen3 I/O expansion drawer.
Figure 1-9 PCIe Gen3 I/O expansion drawer
18IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 33
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
P1-C3
P1-C2
P1-C1
P1-C4
P1-C5
P1-C6
P2-C1
P2-C2
P2-C3
P2-C6
P2-C5
P2-C4
1.6.2 I/O drawers and usable PCI slot
Figure 1-10 shows the rear view of the PCIe Gen3 I/O expansion drawer with the location
codes for the PCIe adapter slots in the PCIe3 6-slot fanout module.
Figure 1-10 Rear view of a PCIe Gen3 I/O expansion drawer with PCIe slots location codes
Table 1-7 provides details of the PCI slots in the PCIe Gen3 I/O expansion drawer.
Table 1-7 PCIe slot locations and descriptions for the PCIe Gen3 I/O expansion drawer
SlotLocation codeDescription
Slot 1P1-C1PCIe3, x16
Slot 2P1-C2PCIe3, x8
Slot 3P1-C3PCIe3, x8
Slot 4P1-C4PCIe3, x16
Slot 5P1-C5PCIe3, x8
Slot 6P1-C6PCIe3, x8
Slot 7P2-C1PCIe3, x16
Slot 8P2-C2PCIe3, x8
Slot 9P2-C3PCIe3, x8
Slot 10P2-C4PCIe3, x16
Slot 11P2-C5PCIe3, x8
Slot 12P2-C6PCIe3, x8
All slots support full-length, regular-height adapter or short (low-profile) with a
regular-height tailstock in single-wide, generation 3, blind-swap cassettes.
Chapter 1. General description 19
Page 34
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
1 2 3 4 5 6 7 8 910111213141516 1718192021222324
FC 58 87
C1
C2
LNK LNK LN K LN K
T2T 3T1
LNK LNK
10/100 LNK
LNK LNK LNK LNK
T2T3T1
LNK LNK
10/100 LNK
ESMESM
OK
IN
DC
OK
IN
DC
Front
Rear
Slots C1 and C4 in each PCIe3 6-slot fanout module are x16 PCIe3 buses and slots C2,
C3, C5, and C6 are x8 PCIe buses.
All slots support enhanced error handling (EEH).
All PCIe slots are hot swappable and support concurrent maintenance.
Table 1-8 summarizes the maximum number of I/O drawers supported and the total number
of PCI slots that are available when expansion consists of a single drawer type.
Table 1-8 Maximum number of I/O drawers supported and total number of PCI slots
System nodesMaximum #EMX0
drawers
Total number of slots
PCIe3, x16PCIe3, x8
1 system node2816
2 system nodes41632
1.6.3 EXP24S SFF Gen2-bay Drawer
The EXP24S SFF Gen2-bay Drawer (#5887) is an expansion drawer with 24 2.5-inch
form-factor SAS bays. The EXP24S supports up to 24 hot-swap SFF-2 SAS hard disk drives
(HDDs) or solid state drives (SSDs). It uses only 2 EIA of space in a 19-inch rack. The
EXP24S includes redundant AC power supplies and uses two power cords.
With AIX, Linux, and VIOS, you can order the EXP24S with four sets of six bays, two sets of
12 bays, or one set of 24 bays (mode 4, 2, or 1). With IBM i, you can order the EXP24S as
one set of 24 bays (mode 1). Mode setting is done by IBM Manufacturing and there is no
option provided to change the mode after it is shipped.
The EXP24S SAS ports are attached to a SAS PCIe adapter or pair of adapters using SAS
YO or X cables.
Figure 1-11 shows the EXP24S SFF Gen2-bay drawer.
20IBM Power Systems E870 and E880 Technical Overview and Introduction
Figure 1-11 EXP24S SFF Gen2-bay drawer
Page 35
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
1.7 Comparison between models
The Power E870 offers configuration options where the POWER8 processor can have one of
two processor speeds. Each system node is populated with four single chip modules (SCMs).
Each system node will contain one of the following processor configurations:
Four 4.02 GHz 8-core SCMs
Four 4.19 GHz 10-core SCMs
A Power E870 with either of the processor configurations can have as few as eight cores
activated or up to 100% of the cores can be activated. Incrementing one core at a time is
available through built-in capacity on demand (CoD) functions to the full capacity of the
system. The Power E870 can be installed with one or two system nodes connected to the
system control unit. Each system node can have up to 2 TB of memory installed.
The Power E880 also offers system nodes populated with four SCMs. The Power E880
system node has one processor configuration:
Four 4.35 GHz 8-core SCMs
A Power E880 with either of the processor configurations can have as few as eight cores
activated or up to 100% of the cores can be activated. Incrementing one core at a time is
available through built-in capacity on demand (CoD) functions to the full capacity of the
system. The Power E880 can be installed with one, two, three, or four system nodes
connected to the system control unit. In 2014 up to two system nodes are supported. Each
system node can have up to 4 TB of memory installed.
Statement of direction:
IBM plans to enhance the Power Systems enterprise system portfolio with greater
scalability, availability and flexibility. IBM intends to deliver the following offerings:
A more scalable Power E880 enterprise-class system with up to 192 POWER8
processor cores and up to 16 TB of total memory
Support for concurrent maintenance on the Power E880 I/O Expansion Drawer by
enabling hot add and repair capabilities
IBM's statements regarding its plans, directions, and intent are subject to change or
withdrawal without notice at IBM's sole discretion. Information regarding potential future
products is intended to outline our general product direction and it should not be relied on
in making a purchasing decision. The information mentioned regarding potential future
products is not a commitment, promise, or legal obligation to deliver any material, code, or
functionality. Information about potential future products may not be incorporated into any
contract. The development, release, and timing of any future features or functionality
described for our products remains at our sole discretion.
Table 1-9 shows a summary of processor and memory maximums for the Power E870 and
Power E880.
Table 1-9 Summary of processor and memory maximums for the Power E870 and Power E880
SystemCores
per
SCM
Core
speed
System
node core
maximum
System
core
maximum
System
node
memory
maximum
System
memory
maximum
Power E87084.02 GHz32642 TB4TB
Chapter 1. General description 21
Page 36
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
SystemCores
Power E870104.19 GHz40802 TB4 TB
Power E880 in
2014
Power E88084.35 GHz321284 TB16TB
1.8 Build to order
You can do a build to order (also called a la carte) configuration by using the IBM
Configurator for e-business (e-config). With it, you specify each configuration feature that you
want on the system.
This method is the only configuration method for the Power E870 and Power E880 servers.
1.9 IBM editions
Core
per
SCM
84.35 GHz32644 TB8TB
speed
System
node core
maximum
System
core
maximum
System
node
memory
maximum
System
memory
maximum
IBM edition offerings are not available for the Power E870 and Power E880 servers.
1.10 Model upgrades
The following sections describe the various upgrades that are available.
1.10.1 Power E870
A model conversion from a Power 770 (9117-MMD) to a Power E870 (9119-MME) is
available. One-step upgrades from previous models of the Power 770 (9117-MMB and
9117-MMC) are not available. To upgrade from a 9117-MMB or 9117-MMC, an upgrade to a
9117-MMD is required first.
The existing components being replaced during a model or feature conversion become the
property of IBM and must be returned.
1.10.2 Power E880
A model conversion from a Power 780 (9179-MHD) to a Power E880 (9119-MHE) is available.
One-step upgrades from previous models of the Power 780 (9179-MHB and 9179-MHC) are
not available. To upgrade from a 9179-MHB or 9179-MHC, an upgrade to a 9179-MHD is
required first.
Upgrades to a Power E880 from a Power 795 (9119-FHB) are not available.
The existing components being replaced during a model or feature conversion become the
property of IBM and must be returned.
22IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 37
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
1.10.3 Upgrade considerations
Feature conversions are set up for the following items:
PCIe Crypto blind swap cassettes
Power IFL processor activations
Power IFL PowerVM for Linux
Active Memory Expansion Enablement
DDR3 memory DIMMS to CDIMMS
Static and mobile memory activations
5250 enterprise enablement
POWER7+™ processor cards to POWER8 processors
Static and mobile processor activations
System CEC enclosure and bezel to 5U system node drawer
PowerVM standard and enterprise
The following features that are present on the current system can be moved to the new
system if they are supported in the Power E870 and E880:
Disks
SSDs
PCIe adapters with cables, line cords, keyboards, and displays
Racks
Doors
EXP24S I/O drawers
For POWER7+ processor-based systems that have the Elastic CoD function enabled, you
must reorder the elastic CoD enablement features when placing the upgrade MES order for
the new Power E870 or E880 system to keep the elastic CoD function active. To initiate the
model upgrade, the on/off enablement features should be removed from the configuration file
before the MES order is started. Any temporary use of processors or memory owed to IBM on
the existing system must be paid before installing the Power E870 or E880.
1.11 Management consoles
This section discusses the supported management interfaces for the servers.
The Hardware Management Console (HMC) is required for managing the IBM
Power E870 and Power E880. It has a set of functions that are necessary to manage
the system:
Creating and maintaining a multiple partition environment
Displaying a virtual operating system session terminal for each partition
Displaying a virtual operator panel of contents for each partition
Detecting, reporting, and storing changes in hardware conditions
Powering managed systems on and off
Acting as a service focal point for service representatives to determine an appropriate
service strategy
Chapter 1. General description 23
Page 38
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
In 2014, a new HMC model was announced, machine type 7042-CR8. Hardware features on
the CR8 model include a new x86 processor, dual hard drives (RAID 1), and redundant power
supply. At the time of writing, the latest version of HMC code was V8.2.0. This code level is
required for new the new systems.
The IBM Power E870 and Power E880 are not supported by the Integrated Virtualization
Manager (IVM).
Several HMC models are supported to manage POWER8 based systems. Two models
(7042-CR7 and 7042-CR8) are available for ordering at the time of writing, but you can also
use one of the withdrawn models listed in Table 1-10.
7042-C08WithdrawnIBM 7042 Model C08 Deskside Hardware Management Console
7042-CR5WithdrawnIBM 7042 Model CR5 Rack-Mounted Hardware Management Console
7042-CR6WithdrawnIBM 7042 Model CR6 Rack mounted Hardware Management Console
7042-CR7AvailableIBM 7042 Model CR7 Rack mounted Hardware Management Console
7042-CR8AvailableIBM 7042 Model CR8 Rack mounted Hardware Management Console
At the time of writing, base Licensed Machine Code Version 8 Revision 8.2.0 or later is
required to support the Power E870 (9119-MME) and Power E880 (9119-MHE).
Fix Central: You can download or order the latest HMC code from the Fix Central website:
http://www.ibm.com/support/fixcentral
Existing HMC models 7042 can be upgraded to Licensed Machine Code Version 8 to support
environments that might include POWER6®, POWER6+®,POWER7®, POWER7+ and
POWER8 processor-based servers.
If you want to support more than 254 partitions in total, the HMC will require a memory
upgrade to a minimum of 4 GB.
1.12 System racks
The Power E870 and E880 and its I/O drawers are designed to be mounted in the following
existing IBM racks: 7014-T00, 7014-T42, 7965-94Y, #0553, #0551, and #ER05.
However, for initial system orders, the racks must be ordered as machine type 7014-T42,
#0553. The 36U (1.8-meter) rack (#0551) and the 42U (2.0-meter) rack (#ER05) are available
for order only on Miscellaneous Equipment Specification (MES) upgrade orders only.
Shipping without a rack: If you require the system to be shipped without an IBM rack,
feature code #ER21 must be used to remove the IBM rack from the order. The server will
then ship as separate packages for installation into an existing rack.
The Power E870 and Power E880 use a new type of connector between system drawers.
Therefore, the systems do not require wider racks, and an OEM rack or cabinet which meet
the requirements can be used.
24IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 39
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
Installing in non-IBM racks: The client is responsible for ensuring that the installation of
the drawer in the preferred rack or cabinet results in a configuration that is stable,
serviceable, safe, and compatible with the drawer requirements for power, cooling, cable
management, weight, and rail security.
1.12.1 IBM 7014 model T00 rack
The 1.8-meter (71-inch) model T00 is compatible with past and present IBM Power systems.
The features of the T00 rack are as follows:
36U (EIA units) of usable space.
Optional removable side panels.
Optional highly perforated front door.
Optional side-to-side mounting hardware for joining multiple racks.
Standard business black or optional white color in OEM format.
Increased power distribution and weight capacity.
Supports both AC and DC configurations.
The rack height is increased to 1926 mm (75.8 in.) if a power distribution panel is fixed to
the top of the rack.
The #6068 feature provides a cost effective plain front door
Weights are as follows:
– T00 base empty rack: 244 kg (535 lb.)
– T00 full rack: 816 kg (1795 lb.)
– Maximum Weight of Drawers is 572 kg (1260 lb.)
– Maximum Weight of Drawers in a zone 4 earthquake environment is 490 kg (1080 lb.).
This equates to 13.6 kg (30 lb.)/EIA.
1.12.2 IBM 7014 model T42 rack
The 2.0-meter (79.3-inch) Model T42 addresses the requirement for a tall enclosure to house
the maximum amount of equipment in the smallest possible floor space. The following
features differ in the model T42 rack from the model T00:
The T42 rack has 42U (EIA units) of usable space (6U of additional space).
The model T42 supports AC power only.
Weights are as follows:
– T42 base empty rack: 261 kg (575 lb.)
– T42 full rack: 930 kg (2045 lb.)
The feature #ERG7 provides an attractive black full-height rack door. The door is steel,
with a perforated flat front surface. The perforation pattern extends from the bottom to the
top of the door to enhance ventilation and provide some visibility into the rack.
The feature #6069 provides a cost-effective plain front door.
The feature #6249 provides a special acoustic door
Special door: The Power 780 logo rack door (#6250) is not supported.
Chapter 1. General description 25
Page 40
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
1.12.3 IBM 7953 model 94Y rack
The 2.0-meter (79.3 inch) model 94Y rack has the following features:
42U (EIA units)
Weights:
– Base empty rack: 187 kg (221 lb.)
– Maximum load limit: 664 kg (1460 lb.)
The IBM 42U Slim Rack (7953-94Y) differs from the IBM 42U enterprise rack (7014-T42) in
several aspects. Both provide 42U of vertical space, are 1100 mm deep, and have an interior
rail-to-rail depth of 715 mm. However, the IBM 42U Slim Rack is 600 mm wide; the T42 is
645 mm wide with side covers. For clients with 2-foot floor tiles, the extra 45 mm (1.77-inch)
width of the enterprise rack can sometimes cause challenges when cutting holes in the floor
tiles for cabling.
The 42U Slim Rack has a lockable perforated front steel door, providing ventilation, physical
security, and visibility of indicator lights in the installed equipment within. In the rear, either a
lockable perforated rear steel door (#EC02) or a lockable rear door heat exchanger (#EC15)
is used. Lockable optional side panels (#EC03) increase the rack’s aesthetics, help control
airflow through the rack, and provide physical security. Multiple 42U Slim Racks can be bolted
together to create a rack suite (indicated with feature #EC04).
Weight calculation: Maximum weight limits must include everything that will be installed
in the rack. This must include servers, I/O drawers, PDUs, switches, and anything else
installed. In zone 4 earthquake environments the rack should be configured starting with
the heavier drawers at the bottom of the rack.
1.12.4 Feature code #0551 rack
The 1.8-meter rack (#0551) is a 36U (EIA units) rack. The rack that is delivered as #0551 is
the same rack delivered when you order the 7014-T00 rack. The included features might
differ. Several features that are delivered as part of the 7014-T00 must be ordered separately
with the #0551. The #0551 is not available for initial orders of Power E870 and E880 servers.
1.12.5 Feature code #0553 rack
The 2.0-meter rack (#0553) is a 42U (EIA units) rack. The rack that is delivered as #0553 is
the same rack delivered when you order the 7014-T42. The included features might differ.
Several features that are delivered as part of the 7014-T42 or must be ordered separately
with the #0553.
1.12.6 The AC power distribution unit and rack content
For rack models T00, T42 and the slim 94Y, 12-outlet PDUs are available. The PDUs
available include:
PDU mounting: Only horizontal PDUs are allowed in racks hosting Power E870 and
Power E880 system. Vertically mounted PDUs limit access to the cable routing space on
the side of the rack and cannot be used.
26IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 41
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
41
42
39
40
37
38
35
36
33
34
31
32
29
30
27
28
25
26
23
24
21
22
19
20
17
18
15
16
13
14
11
12
09
10
07
08
05
06
03
04
01
02
41
42
39
40
37
38
35
36
33
34
31
32
29
30
27
28
25
26
23
24
21
22
19
20
17
18
15
16
13
14
11
12
09
10
07
08
05
06
03
04
01
02
System Node
#1
Control Unit
I/O Expansion Drawer
P4P3P1P2
30A PDU
30A PDU
P1P2
I/O Expansion Drawer
P1P2
System Node
#2
Control Unit
I/O Expansion Drawer
I/O Expansion Drawer
P4P3P1 P2
P4P3P1 P2
30A PDU
30A PDU
30A PDU
30A PDU
P2P1
P2P1
When mounting the horizontal PDUs it is a good practice to place them almost at the top or
almost at the bottom of the rack leaving 2U or more of space at the very top or very bottom of
the open for cable management. Mounting a horizontal PDU in the middle of the rack is
generally not optimal for cable management.
For the Power E870 or E880 installed in IBM 7014 or FC 055x racks, the following PDU rules
apply:
For PDU #7109 and #7188 when using 24 Amp power cord #6654, #6655, #6656, #6657,
or #6658, each pair of PDUs can power one Power E870 or Power E880 system nodes
and two I/O expansion drawers or eight I/O expansion drawers. 24A PDU cables are used
to supply 30A PDUs. In the Figure 1-12 on page 27 you can see the rack configuration
with two pairs of 30A PDUs suppling two system nodes configuration and two I/O
expansion drawers.
Figure 1-12 Two system node configuration and two I/O expansion drawers supplied by 30A PDUs
For PDU #7109 and #7188 when using three phase power cords or 48 Amp power cords
#6491 or #6492, each pair of PDUs can power up to two Power E870 or Power E880
system nodes and two I/O expansion drawer or eight I/O expansion drawers. 48A PDU
cables are used to supply 60A PDU. In the Figure 1-13 on page 28 you can see the rack
Chapter 1. General description 27
Page 42
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
41
42
39
40
37
38
35
36
33
34
31
32
29
30
27
28
25
26
23
24
21
22
19
20
17
18
15
16
13
14
11
12
09
10
07
08
05
06
03
04
01
02
41
42
39
40
37
38
35
36
33
34
31
32
29
30
27
28
25
26
23
24
21
22
19
20
17
18
15
16
13
14
11
12
09
10
07
08
05
06
03
04
01
02
System Node
#2
System Node
#1
Control Unit
I/O Expansion Drawer
I/O Expansion Drawer
I/O Expansion Drawer
I/O Expansion Drawer
EXP24S
EXP24S
EXP24S
EXP24S
P4P3P1 P2
P4P3P1 P2
60A PDU
60A PDU
60A PDU
60A PDU
P2P1
P2P1
P2P1
P2P1
P2P1
P2P1
P2P1
P2P1
configuration with two pairs of 60A PDUs suppling a two system node configuration, four
I/O expansion drawers, and four EXP24S disk drawers.
28IBM Power Systems E870 and E880 Technical Overview and Introduction
Figure 1-13 Two system nodes configuration with I/O expansions and disk drawers supplied by 60A
PDUs
For detailed power cord requirements and power cord feature codes, see the IBM Power
Systems Hardware Information Center website:
<< add an infocenter link once is available>>
Power cord: Ensure that the appropriate power cord feature is configured to support the
power being supplied.
For rack integrated systems a minimum quantity of two PDUs (#7109 or #7188) are required.
The PDUs (#7109, #7188) support a wide range of country requirements and electrical power
specifications. The PDU receives power through a UTG0247 power line connector. Each PDU
requires one PDU-to-wall power cord. Various power cord features are available for countries
and applications by selecting a PDU-to-wall power cord, which must be ordered separately.
Each power cord provides the unique design characteristics for the specific power
requirements. To match new power requirements and save previous investments, these
Page 43
Draft Document for Review October 14, 2014 10:19 am5137ch01.fm
power cords can be requested with an initial order of the rack or with a later upgrade of the
rack features.
The PDUs have 12 client-usable IEC 320-C13 outlets. There are six groups of two outlets fed
by six circuit breakers. Each outlet is rated up to 10 Amps, but each group of two outlets is fed
from one 20 A circuit breaker.
The Universal PDUs are compatible with previous models.
Power cord and PDU: Based on the power cord that is used, the PDU can supply a range
of 4.8 - 21 kVA. The total kilovolt ampere (kVA) of all the drawers that are plugged into the
PDU must not exceed the power cord limitation.
Each system node mounted in the rack requires four power cords. For maximum
availability, be sure to connect power cords from the same system to two separate PDUs in
the rack, and to connect each PDU to independent power sources.
Chapter 1. General description 29
Page 44
5137ch01.fmDraft Document for Review October 14, 2014 10:19 am
30IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 45
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
2
Chapter 2.Architecture and technical
overview
This chapter describes the overall system architecture for the IBM Power System E870
(9119-MME) and IBM Power System E880 (9119-MHE) servers. The bandwidths that are
provided throughout the section are theoretical maximums that are used for reference.
The speeds that are shown are at an individual component level. Multiple components and
application implementation are key to achieving the best performance.
Always do the performance sizing at the application workload environment level and evaluate
performance by using real-world performance measurements and production workloads.
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
SCU Interface
P8 (1)
Memory
Controller 0
Memory
Controller 1
PHB 0
PHB 1
P8 (2)
PHB 0
PHB 1
P8 (0)
PHB 0
PHB 1
P8 (3)
PHB 0
PHB 1
SMP
X Bus
SMP
X Bus
SMP
X Bus
SMP
X Bus
SMP
A Bus
SMP
A Bus
SMP
A Bus
SMP
A Bus
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
Memory
Controller 0
Memory
Controller 1
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
Memory
Controller 0
Memory
Controller 1
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
Memory
Controller 0
Memory
Controller 1
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
CDIMM
To other
enclosures
To other
enclosures
To other
enclosures
To other
enclosures
SCU Interface
card
12v power
to SCU
Service
Processor
Connection
Clock
Interface
To SCU
Clock
SCU Interface
SCU Interface
card
12v power
to SCU
Service
Processor
Connection
Clock
Interface
To SCU
Clock
25.6 GB/s
76.8 GB/s
15.75 GB/s
28.8 GB/s
2.1 Logical diagrams
This section contains logical diagrams for the Power E870 and E880.
Figure 2-1 shows the logical system diagram for a single system node of a Power E870 or
Power E880.
Figure 2-1 Logical system diagram for a system node of a Power E870 or a Power E880
32IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 47
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
Power in
Operator Panel
From
CEC
Power in
From
CEC
Power in
From
CEC
Power in
From
CEC
USB
cable
to
PCIe card
DVD
FSP 0FSP 1Clock 0Clock 1
VPD
Figure 2-2 on page 33 shows the logical system diagram for the system control unit of a
Power E870 or a Power E880.
Figure 2-2 Logical system diagram for the system control unit
Chapter 2. Architecture and technical overview 33
Page 48
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
Drawer 1
Drawer 0
X3
X1
X0
X3
X1
X0
X3
X1
X0
X3
X1
X0
X0
X1
X3
X0
X1
X3
X0
X1
X3
X0
X1
X3
P0
P2
P3
P1
T1T2T5T6T3T4
T6
T5
T4
T3
T2
T1
A2
A1
A0
A2
A1
A0
A2
A1
A0
A2
A1
A0
A0
A1
A2
A0
A1
A2
A0
A1
A2
A0
A1
A2
P1
P3
P2
P0
Flexible symmetric multiprocessing (SMP) cables are used to connect system nodes when a
Power E870 or Power E880 is configured with more than one system node. Figure 2-3 on
page 34 shows the SMP connection topology for a two drawer Power E870 or Power E880
system.
Figure 2-3 SMP connection topology for a two-drawer Power E870 or Power E880
34IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 49
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
T1T2T5T6T3T4
P3
X3
X1
X0
A2
A1
A0
Drawer 3
P2
X3
X1
X0
A2
A1
A0
P0
X3
X1
X0
A2
A1
A0
P1
X3
X1
X0
A2
A1
A0
Drawer 2
T1T2T5T6T3T4
Drawer 1
X3
X1
X0
X3
X1
X0
X3
X1
X0
X3
X1
X0
P0
P2
P3
P1
A2
A1
A0
A2
A1
A0
A2
A1
A0
A2
A1
A0
T6
T5
T4
T3
T2
T1
A0
A1
A2
A0
A1
A2
A0
A1
A2
A0
A1
A2
X0
X1
X3
X0
X1
X3
X0
X1
X3
X0
X1
X3
P1
P3
P2
P0
Figure 2-4 shows the SMP connection topology for a three-drawer Power E880.
Figure 2-4 SMP connection topology for a three-drawer Power E880
Chapter 2. Architecture and technical overview 35
Page 50
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
T1T2T5T3T4
P3
X3
X1
X0
A2
A1
A0
Drawer 3
P2
X3
X1
X0
A2
A1
A0
P0
X3
X1
X0
A2
A1
A0
P1
X3
X1
X0
A2
A1
A0
Drawer 2
Drawer 4
T1T2T5T6T3T4
Drawer 1
X3
X1
X0
P0
P2
P3
P1
X3
X1
X0
X3
X1
X0
X3
X1
X0
A2
A1
A0
A2
A1
A0
A2
A1
A0
A2
A1
A0
T6
T6
T5
T2
T4T3
T1
T6
T5
T4
T3
T2
T1
A2
A1
A0
A2
A1
A0
A2
A1
A0
A2
A1
A0
X3
X1
X0
X3
X1
X0
X3
X1
X0
X3
X1
X0
P0
P2
P3
P1
A0
A1
A2
A0
A1
A2
A0
A1
A2
A0
A1
A2
X0
X1
X3
X0
X1
X3
X0
X1
X3
X0
X1
X3
P1
P3
P2
P0
Figure 2-5 shows the SMP connection topology for a four-drawer Power E880.
Figure 2-5 SMP connection topology for a four-drawer Power E880
2.2 The IBM POWER8 processor
This section introduces the latest processor in the IBM Power Systems product family, and
describes its main characteristics and features.
36IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 51
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
2.2.1 POWER8 processor overview
The POWER8 processor is manufactured using the IBM 22 nm Silicon-On-Insulator (SOI)
technology. Each chip is 649 mm
2
and contains 4.2 billion transistors. As shown in Figure 2-6,
the chip contains 12 cores, two memory controllers, PCIe Gen3 I/O controllers, and an
interconnection system that connects all components within the chip. On some systems only
6, 8, 10, or 12 cores per processor may be available to the server. Each core has 512 KB of
L2 cache, and all cores share 96 MB of L3 embedded DRAM (eDRAM). The interconnect
also extends through module and board technology to other POWER8 processors in addition
to DDR3 memory and various I/O devices.
POWER8 systems use memory buffer chips to interface between the POWER8 processor
and DDR3 or DDR4 memory
1
. Each buffer chip also includes an L4 cache to reduce the
latency of local memory accesses.
Figure 2-6 The POWER8 processor chip
The POWER8 processor is for system offerings from single-socket servers to multi-socket
Enterprise servers. It incorporates a triple-scope broadcast coherence protocol over local and
global SMP links to provide superior scaling attributes. Multiple-scope coherence protocols
reduce the amount of SMP link bandwidth that is required by attempting operations on a
limited scope (single chip or multi-chip group) when possible. If the operation cannot
complete coherently, the operation is reissued using a larger scope to complete the operation.
1
At the time of the publication, the available POWER8 processor-based systems use DDR3 memory.
Chapter 2. Architecture and technical overview 37
Page 52
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
The following are additional features that can augment the performance of the POWER8
processor:
Support for DDR3 and DDR4 memory through memory buffer chips that offload the
memory support from the POWER8 memory controller.
Each memory CDIMM has 16 MB of L4 cache within the memory buffer chip that reduces
the memory latency for local access to memory behind the buffer chip; the operation of the
L4 cache is not apparent to applications running on the POWER8 processor. Up to 128
MB of L4 cache can be available for each POWER8 processor.
Hardware transactional memory.
On-chip accelerators, including on-chip encryption, compression, and random number
generation accelerators.
Coherent Accelerator Processor Interface, which allows accelerators plugged into a PCIe
slot to access the processor bus using a low latency, high-speed protocol interface.
Adaptive power management.
There are two versions of the POWER8 processor chip. Both chips use the same building
blocks. The scale-out systems use a 6-core version of POWER8. The 6-core chip is installed
in pairs in a Dual Chip Module (DCM) that plugs into a socket in the system board of the
systems. Functionally, it works as a single chip module (SCM).
Figure 2-7 shows a graphic representation of the 6-core processor. A 6-core processor is only
available on the scale-out systems. It is shown here for informational purposes.
Figure 2-7 6-core POWER8 processor chip
38IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 53
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
Table 2-1 summarizes the technology characteristics of the POWER8 processor.
Table 2-1 Summary of POWER8 processor technology
TechnologyPOWER8 processor
Die size649 mm
Fabrication technology 22 nm lithography
Copper interconnect
SOI
eDRAM
Maximum processor cores6 or 12
Maximum execution threads core/chip8/96
Maximum L2 cache core/chip512 KB/6 MB
Maximum On-chip L3 cache core/chip8 MB/96 MB
Maximum L4 cache per chip128 MB
Maximum memory controllers2
SMP design-point16 sockets with IBM POWER8 processors
2
CompatibilityWith prior generations of POWER® processor
2.2.2 POWER8 processor core
The POWER8 processor core is a 64-bit implementation of the IBM Power Instruction Set
Architecture (ISA) Version 2.07 and has the following features:
Multi-threaded design, which is capable of up to eight-way simultaneous multithreading
(SMT)
32 KB, eight-way set-associative L1 instruction cache
64 KB, eight-way set-associative L1 data cache
Enhanced prefetch, with instruction speculation awareness and data prefetch depth
awareness
Enhanced branch prediction, using both local and global prediction tables with a selector
table to choose the best predictor
Improved out-of-order execution
Two symmetric fixed-point execution units
Two symmetric load/store units and two load units, all four of which can also run simple
fixed-point instructions
An integrated, multi-pipeline vector-scalar floating point unit for running both scalar and
SIMD-type instructions, including the Vector Multimedia eXtension (VMX) instruction set
and the improved Vector Scalar eXtension (VSX) instruction set, and capable of up to
eight floating point operations per cycle (four double precision or eight single precision)
In-core Advanced Encryption Standard (AES) encryption capability
Hardware data prefetching with 16 independent data streams and software control
Hardware decimal floating point (DFP) capability
Chapter 2. Architecture and technical overview 39
Page 54
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
More information about Power ISA Version 2.07 can be found at the following website:
Figure 2-8 shows a picture of the POWER8 core, with some of the functional units
highlighted.
Figure 2-8 POWER8 processor core
2.2.3 Simultaneous multithreading
POWER8 processor advancements in multi-core and multi-thread scaling are remarkable. A
significant performance opportunity comes from parallelizing workloads to enable the full
potential of the microprocessor, and the large memory bandwidth. Application scaling is
influenced by both multi-core and multi-thread technology.
Simultaneous Multithreading (SMT) allows a single physical processor core to simultaneously
dispatch instructions from more than one hardware thread context. With SMT, each POWER8
core can present eight hardware threads. Because there are multiple hardware threads per
physical processor core, additional instructions can run at the same time. SMT is primarily
beneficial in commercial environments where the speed of an individual transaction is not as
critical as the total number of transactions that are performed. SMT typically increases the
throughput of workloads with large or frequently changing working sets, such as database
servers and web servers.
Table 2-2 on page 41 shows a comparison between the different POWER processors in terms
of SMT capabilities that are supported by each processor architecture.
40IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 55
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
Table 2-2 SMT levels that are supported by POWER processors
The architecture of the POWER8 processor, with its larger caches, larger cache bandwidth,
and faster memory, allows threads to have faster access to memory resources, which
translates in to a more efficient usage of threads. Because of that, POWER8 allows more
threads per core to run concurrently, increasing the total throughput of the processor and of
the system.
2.2.4 Memory access
On the Power E870 and Power E880, each POWER8 processor has two memory controllers,
each connected to four memory channels. Each memory channel operates at 1600 MHz and
connects to a DIMM. Each DIMM on a POWER8 system has a memory buffer that is
responsible for many functions that were previously on the memory controller, such as
scheduling logic and energy management. The memory buffer also has 16 MB of level 4 (L4)
cache.
On the Power E870 each memory channel can address up to 64 GB. Therefore a single
system node can address 2 TB of memory. A two node system can address up to 4 TB of
memory.
On the Power E880 each memory channel can address up to 128 GB. Therefore a single
system node can address 4 TB of memory. A two node system can address up to 8 TB of
memory and a four node system can address up to 16 TB of memory.
Chapter 2. Architecture and technical overview 41
Page 56
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
POWER8
Processor
Memory
Buffers
DRAM
Chips
Memory
Buffers
DRAM
Chips
Figure 2-9 gives a simple overview of the POWER8 processor memory access structure in
the Power E870 and Power E880.
Figure 2-9 Overview of POWER8 memory access structure
2.2.5 On-chip L3 cache innovation and Intelligent Cache
Similar to POWER7 and POWER7+, the POWER8 processor uses a breakthrough in material
engineering and microprocessor fabrication to implement the L3 cache in eDRAM and place it
on the processor die. L3 cache is critical to a balanced design, as is the ability to provide good
signaling between the L3 cache and other elements of the hierarchy, such as the L2 cache or
SMP interconnect.
The on-chip L3 cache is organized into separate areas with differing latency characteristics.
Each processor core is associated with a fast 8 MB local region of L3 cache (FLR-L3) but also
has access to other L3 cache regions as shared L3 cache. Additionally, each core can
negotiate to use the FLR-L3 cache that is associated with another core, depending on the
reference patterns. Data can also be cloned and stored in more than one core’s FLR-L3
cache, again depending on the reference patterns. This
enables the POWER8 processor to optimize the access to L3 cache lines and minimize
overall cache latencies.
Figure 2-6 on page 37 and Figure 2-7 on page 38 show the on-chip L3 cache, and highlight
one fast 8 MB L3 region closest to a processor core.
The benefits of using eDRAM on the POWER8 processor die is significant for several
reasons:
Latency improvement
Bandwidth improvement
A six-to-one latency improvement occurs by moving the L3 cache on-chip compared to L3
accesses on an external (on-ceramic) ASIC.
Intelligent Cache management
A 2x bandwidth improvement occurs with on-chip interconnect. Frequency and bus sizes
are increased to and from each core.
42IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 57
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
No off-chip driver or receivers
Removing drivers or receivers from the L3 access path lowers interface requirements,
conserves energy, and lowers latency.
Small physical footprint
The performance of eDRAM when implemented on-chip is similar to conventional SRAM
but requires far less physical space. IBM on-chip eDRAM uses only a third of the
components that conventional SRAM uses, which has a minimum of six transistors to
implement a 1-bit memory cell.
Low energy consumption
The on-chip eDRAM uses only 20% of the standby power of SRAM.
2.2.6 Level 4 cache and memory buffer
POWER8 processor-based systems introduce an additional level of memory hierarchy. The
Level 4 (L4) cache is implemented together with the memory buffer in the Custom DIMM
(CDIMM). Each memory buffer contains 16 MB of L4 cache. Figure 2-10 shows a picture of
the memory buffer, where you can see the 16 MB L4 cache, and processor links and memory
interfaces.
Figure 2-10 Memory buffer chip
Table 2-3 shows a comparison of the different levels of cache in the POWER7, POWER7+,
and POWER8 processors.
Table 2-3 POWER8 cache hierarchy
CachePOWER7POWER7+POWER8
L1 instruction cache:
Capacity/associativity
L1 data cache:
Capacity/associativity
bandwidth
32 KB, 4-way32 KB, 4-way32 KB, 8-way
32 KB, 8-way
Two 16 B reads or
one 16 B writes per cycle
32 KB, 8-way
Two 16 B reads or
one 16 B writes per cycle
Chapter 2. Architecture and technical overview 43
64 KB, 8-way
Two 16 B reads or
one 16 B writes per cycle
Page 58
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
CachePOWER7POWER7+POWER8
L2 cache:
Capacity/associativity
bandwidth
L3 cache:
Capacity/associativity
bandwidth
L4 cache:
Capacity/associativity
bandwidth
256 KB, 8-way
Private
32 B reads and 16 B writes
per cycle
On-Chip
4 MB/core, 8-way
16 B reads and 16 B writes
per cycle
N/AN/AOn-Chip
For more information about the POWER8 memory subsystem, see 2.3, “Memory subsystem”
on page 47.
2.2.7 Hardware transactional memory
Transactional memory is an alternative to lock-based synchronization. It attempts to simplify
parallel programming by grouping read and write operations and running them like a single
operation. Transactional memory is like database transactions where all shared memory
accesses and their effects are either committed all together or discarded as a group. All
threads can enter the critical region simultaneously. If there are conflicts in accessing the
shared memory data, threads try accessing the shared memory data again or are stopped
without updating the shared memory data. Therefore, transactional memory is also called a
lock-free synchronization. Transactional memory can be a competitive alternative to
lock-based synchronization.
256 KB, 8-way
Private
32 B reads and 16 B writes
per cycle
On-Chip
10 MB/core, 8-way
16 B reads and 16 B writes
per cycle
512 KB, 8-way
Private
32 B reads and 16 B writes
per cycle
On-Chip
8 MB/core, 8-way
32 B reads and 32 B writes
per cycle
16 MB/buffer chip, 16-way
Up to 8 buffer chips per
socket
Transactional Memory provides a programming model that makes parallel programming
easier. A programmer delimits regions of code that access shared data and the hardware
runs these regions atomically and in isolation, buffering the results of individual instructions,
and retrying execution if isolation is violated. Generally, transactional memory allows
programs to use a programming style that is close to coarse-grained locking to achieve
performance that is close to fine-grained locking.
Most implementations of transactional memory are based on software. The POWER8
processor-based systems provide a hardware-based implementation of transactional memory
that is more efficient than the software implementations and requires no interaction with the
processor core, therefore allowing the system to operate at maximum performance.
2.2.8 Coherent Accelerator Processor Interface
The Coherent Accelerator Interface Architecture (CAIA) defines a coherent accelerator
interface structure for attaching peripherals to Power Systems.
The Coherent Accelerator Processor Interface (CAPI) can attach accelerators that have
coherent shared memory access to the processors in the server and share full virtual address
translation with these processors, using a standard PCIe Gen3 bus.
Applications can have customized functions in Field Programmable Gate Arrays (FPGA) and
be able to enqueue work requests directly in shared memory queues to the FPGA, and using
the same effective addresses (pointers) it uses for any of its threads running on a host
44IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 59
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
Custom
Hardware
Application
CAPP
Coh erenc e Bus
PSL
FPGA or ASIC
POW ER8
PCIe G en3
Tra nsport f o r en ca psul ated messag es
processor. From the practical perspective, CAPI allows a specialized hardware accelerator to
be seen as an additional processor in the system, with access to the main system memory,
and coherent communication with other processors in the system.
The benefits of using CAPI include the ability to access shared memory blocks directly from
the accelerator, perform memory transfers directly between the accelerator and processor
cache, and reduction in the code path length between the adapter and the processors. This is
because the adapter is not operating as a traditional I/O device, and there is no device driver
layer to perform processing. It also presents a simpler programming model.
Figure 2-11 shows a high-level view of how an accelerator communicates with the POWER8
processor through CAPI. The POWER8 processor provides a Coherent Attached Processor
Proxy (CAPP), which is responsible for extending the coherence in the processor
communications to an external device. The coherency protocol is tunneled over standard
PCIe Gen3, effectively making the accelerator part of the coherency domain.
The accelerator adapter implements the Power Service Layer (PSL), which provides address
translation and system memory cache for the accelerator functions. The custom processors
on the board, consisting of an FPGA or an Application Specific Integrated Circuit (ASIC) use
this layer to access shared memory regions, and cache areas as through they were a
processor in the system. This ability greatly enhances the performance of the data access for
the device and simplifies the programming effort to use the device. Instead of treating the
hardware accelerator as an I/O device, it is treated as a processor. This eliminates the
requirement of a device driver to perform communication, and the need for Direct Memory
Access that requires system calls to the operating system kernel. By removing these layers,
the data transfer operation requires fewer clock cycles in the processor, greatly improving the
I/O performance.
Figure 2-11 CAPI accelerator that is attached to the POWER8 processor
The implementation of CAPI on the POWER8 processor allows hardware companies to
develop solutions for specific application demands and use the performance of the POWER8
processor for general applications and custom acceleration of specific functions using a
hardware accelerator, with a simplified programming model and efficient communication with
the processor and memory resources.
Chapter 2. Architecture and technical overview 45
Page 60
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
2.2.9 Power management and system performance
The POWER8 processor has power saving and performance enhancing features that can be
used to lower overall energy usage, while yielding higher performance when needed. The
following modes can be enabled and modified to use these features.
Dynamic Power Saver: Favor Performance
This mode is intended to provide the best performance. If the processor is being used even
moderately, the frequency is raised to the maximum frequency possible to provide the best
performance. If the processors are lightly used, the frequency is lowered to the minimum
frequency, which is potentially far below the nominal shipped frequency, to save energy. The
top frequency that is achieved is based on system type and is affected by environmental
conditions. Also, when running at the maximum frequency, more energy is being consumed,
which means this mode can potentially cause an increase in overall energy consumption.
Dynamic Power Saver: Favor Power
This mode is intended to provide the best performance per watt consumed. The processor
frequency is adjusted based on the processor usage to maintain the workload throughput
without using more energy than required to do so. At high processor usage levels, the
frequency is raised above nominal, as in the Favor Performance mode. Likewise, at low
processor usage levels, the frequency is lowered to the minimum frequency. The frequency
ranges are the same for the two Dynamic Power Saver modes, but the algorithm that
determines which frequency to set is different.
Dynamic Power Saver: Tunable Parameters
The Dynamic Power Saver: Favor Performance and Dynamic Power Saver: Favor Power
modes are tuned to provide both energy savings and performance increases. However, there
might be situations where only top performance is of concern, or, conversely, where peak
power consumption is an issue. The tunable parameters can be used to modify the setting of
the processor frequency in these modes to meet these various objectives. Modifying these
parameters should be done only by advanced users. If you must address any issues
concerning the Tunable Parameters, IBM support personal should be directly involved in the
parameter value selection.
Idle Power Saver
This mode is intended to save the maximum amount of energy when the system is nearly idle.
When the processors are found to be nearly idle, the frequency of all processors is lowered to
the minimum. Additionally, workloads are dispatched onto a smaller number of processor
cores so that the other processor cores can be put into a low energy usage state. When
processor usage increases, the process is reversed: The processor frequency is raised back
up to nominal, and the workloads are spread out once again over all of the processor cores.
There is no performance boosting aspect in this mode, but entering or exiting this mode might
affect overall performance. The delay times and usage levels for entering and exiting this
mode can be adjusted to allow for more or less aggressive energy savings.
The controls for all of these modes are available on the Advanced System Management
Interface (ASMI) and are described in more detail in a white paper that is found at the
following link:
For more information, see 2.13, “Energy management” on page 105.
46IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 61
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
2.2.10 Comparison of the POWER7, POWER7+, and POWER8 processors
Table 2-4 shows comparable characteristics between the generations of POWER7,
POWER7+, and POWER8 processors.
Table 2-4 Comparison of technologies for the POWER8 processor and the prior generations
CharacteristicsPOWER7POWER7+POWER8
Technology45 nm32 nm22 nm
Die size567 mm
Number of transistors1.2 billion2.1 billion4.2 billion
Maximum cores8812
2
567 mm
2
649 mm
2
Maximum SMT
threads per core
Maximum frequency4.25 GHz4.4 GHz4.35 GHz
L2 Cache256 KB per core256 KB per core512 KB per core
L3 Cache4 MB or 8 MB of
Memory supportDDR3DDR3DDR3 and DDR4
I/O busGX++GX++PCIe Gen3
2.3 Memory subsystem
The Power E870 can have up to two system nodes per system with each system node having
32 DDR3 CDIMM slots capable of supporting 16 GB, 32 GB, and 64 GB CDIMMs running at
speeds of 1600 MHz. This allows for a maximum system memory of 4 TB for the
64 DDR3 CDIMMs slots in a system comprised of two system nodes.
4 threads4 threads8 threads
FLR-L3 cache per core
with each core having
access to the full 32
MB of L3 cache,
on-chip eDRAM
10 MB of FLR-L3
cache per core with
each core having
access to the full 80
MB of L3 cache,
on-chip eDRAM
8 MB of FLR-L3 cache
per core with each
core having access to
the full 96 MB of L3
cache, on-chip
eDRAM
The Power E880 can have up to four system nodes per system with each system node having
32 DDR3 CDIMM slots capable of supporting 16 GB, 32 GB, 64 GB, and 128 GB CDIMMs
running at speeds of 1600 MHz. This allows for a maximum system memory of 16 TB for the
128 DDR3 CDIMMs slots of a system comprised of four system nodes.
Note: At the time of writing, the Power E880 supports only two system nodes per server.
This allows for a maximum of 8 TB of memory for the 64 DDR3 CDIMM slots.
The memory on the systems are Capacity Upgrade on Demand capable, allowing for the
purchase of additional memory capacity and dynamically activate it when needed. It is
required that at least 50% of the installed RAM capacity is active.
The Power E870 and E880 servers support an optional feature called Active Memory
Expansion. This allows the effective maximum memory capacity to be much larger than the
true physical memory. This feature runs innovative compression and decompression of
memory content by using a dedicated coprocessor present on each POWER8 processor to
Chapter 2. Architecture and technical overview 47
Page 62
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
provide memory expansion up to 125%, depending on the workload type and its memory
usage. As an example, a server with 256 GB of RAM physically installed can effectively be
expanded over 512 GB of RAM. This approach can enhance virtualization and server
consolidation by allowing a partition to do more work with the same physical amount of
memory or allowing a server to run more partitions and do more work with the same physical
amount of memory.
2.3.1 Custom DIMM
Custom DIMMs (CDIMMs) are innovative memory DIMMs that house industry-standard
DRAM memory chips and a set of components that allow for higher bandwidth, lower latency
communications and increased availability. These components include:
By adopting this architecture for the memory DIMMs, several decisions and processes
regarding memory optimizations are run internally into the CDIMM. This saves bandwidth and
allows for faster processor-to-memory communications. This also allows for a more robust
RAS. For more information, see Chapter 4, “Reliability, availability, and serviceability” on
page 145.
The CDIMMs exists in two different form factors, a 152 SDRAM design named the Tall
CDIMM and an 80 SDRAM design named the Short CDIMM. Each design will be comprised
of multiple 4 GB SDRAM devices depending on its total capacity. The CDIMMs slots for the
Power E870 and Power E880 are tall CDIMMs slots. A filler is added to the short CDIMM
allowing it to properly latch into the same physical location of a tall CDIMM and allows for
proper airflow and ease of handling. Tall CDIMMs slots allow for larger DIMM sizes and the
adoption of future technologies more seamlessly.
48IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 63
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
A detailed diagram of the CDIMMs available for the Power E870 and Power E880 can be
seen in Figure 2-12.
Figure 2-12 Short CDIMM and Tall CDIMM details
The Memory Buffer is a L4 cache and is built on eDRAM technology (same as the L3 cache),
which has a lower latency than regular SRAM. Each CDIMM has 16 MB of L4 cache and a
fully populated Power E870 server has 1 GB of L4 Cache while a fully populated Power E880
has 2 GB of L4 Cache. The L4 Cache performs several functions that have direct impact on
performance and bring a series of benefits for the Power E870 and Power E880:
Reduces energy consumption by reducing the number of memory requests.
Increases memory write performance by acting as a cache and by grouping several
random writes into larger transactions.
Partial write operations that target the same cache block are gathered within the L4 cache
before being written to memory, becoming a single write operation.
Reduces latency on memory access. Memory access for cached blocks has up to 55%
lower latency than non-cached blocks.
2.3.2 Memory placement rules
For the Power E870 and Power E880, each memory feature code provides four CDIMMs.
Therefore a maximum of 8 memory feature codes per system node are allowed in order to fill
all the 32 DDR3 CDIMM slots.
All the memory CDIMMs are capable of capacity upgrade on demand and must have a
minimum of 50% of its physical capacity activated. For example, the minimum installed
Chapter 2. Architecture and technical overview 49
Page 64
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
memory for both servers is 256 GB of RAM, whereas they can have a minimum of 128 GB of
RAM active.
For the Power E870 the following memory options are orderable:
64 GB (4 X 16 GB) CDIMMs, 1600 MHz DDR3 DRAM (#EM8J)
128 GB (4 X 32 GB) CDIMMs, 1600 MHz DDR3 DRAM (#EM8K)
256 GB (4 X 64 GB) CDIMMs, 1600 MHz DDR3 DRAM (#EM8L)
For the Power E880 the following memory options are orderable:
64 GB (4 X 16 GB) CDIMMs, 1600 MHz DDR3 DRAM (#EM8J)
128 GB (4 X 32 GB) CDIMMs, 1600 MHz DDR3 DRAM (#EM8K)
256 GB (4 X 64 GB) CDIMMs, 1600 MHz DDR3 DRAM (#EM8L)
512 GB (4 X 128 GB) CDIMMs, 1600 MHz DDR3 DRAM (#EM8M)
Each processor has two memory controllers. These memory controllers must have at least a
pair of CDIMMs attached to it. This set of mandatory four CDIMMs is called memory quad. A
logical diagram of a POWER8 processor with its two memory quads can be seen on
Figure 2-13.
Figure 2-13 Logical diagram of a POWER8 processor and its two memory quads
The basic rules for memory placement follows:
Each feature code equals a set of four physical CDIMMs; a memory quad.
Each installed processor must have at least one memory quad populated which equals to
at least one feature code per installed processor.
A given processor can only have four or eight CDIMMs attached to it.
All the CDIMMs connected to the same POWER8 processor must be identical. However, it
is permitted to mix different CDIMM sizes between different POWER8 processors on a
system.
At least 50% of the installed memory must be activated via memory activation features.
The suggested approach is to install memory evenly across all processors in the system and
across all system nodes in a system. Balancing memory across the installed processors
allows memory access in a consistent manner and typically results in the best possible
performance for your configuration. You should account for any plans for future memory
upgrades when you decide which memory feature size to use at the time of the initial system
order.
50IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 65
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
POWER8
Processor
1
MC
1
MC
0
POWER8
Processor
3
MC
1
MC
0
POWER8
Processor
2
MC
1
MC
0
POWER8
Processor
0
MC
1
MC
0
P1-C52 (Quad 1)
P1-C51 (Quad 1)
P1-C50 (Quad 5)
P1-C49 (Quad 5)
P1-C48 (Quad 5)
P1-C47 (Quad 5)
P1-C46 (Quad 1)
P1-C45 (Quad 1)
P1-C44 (Quad 2)
P1-C43 (Quad 2)
P1-C42 (Quad 6)
P1-C41 (Quad 6)
P1-C40 (Quad 6)
P1-C39 (Quad 6)
P1-C38 (Quad 2)
P1-C37 (Quad 2)
P1-C36 (Quad 3)
P1-C35 (Quad 3)
P1-C34 (Quad 7)
P1-C33 (Quad 7)
P1-C32 (Quad 7)
P1-C31 (Quad 7)
P1-C30 (Quad 3)
P1-C29 (Quad 3)
P1-C28 (Quad 4)
P1-C27 (Quad 4)
P1-C26 (Quad 8)
P1-C25 (Quad 8)
P1-C24 (Quad 8)
P1-C23 (Quad 8)
P1-C22 (Quad 4)
P1-C21 (Quad 4)
Location (Memory Quad)
Front
Rear
A physical diagram with the location codes of the memory CDIMMs of a system node, as well
as their grouping as memory quads can be seen on Figure 2-14 on page 51.
Figure 2-14 System node physical diagram with location codes for CDIMMs
Each system node has eight memory quads identified by the different colors on Figure 2-14.
The location codes for the slots on each memory quad are:
Quad 1: P1-C45, P1-C46, P1-C51, and P1-C52
Quad 2: P1-C37, P1-C38, P1-C43, and P1-C44
Quad 3: P1-C29, P1-C30, P1-C35, and P1-C36
Quad 4: P1-C21, P1-C22, P1-C27, and P1-C28
Quad 5: P1-C47, P1-C48, P1-C49, and P1-C50
Quad 6: P1-C39, P1-C40, P1-C41, and P1-C42
Quad 7: P1-C31, P1-C32, P1-C33, and P1-C34
Quad 8: P1-C23, P1-C24, P1-C25, and P1-C26
Chapter 2. Architecture and technical overview 51
Page 66
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
Table 2-5 shows the CDIMM plugging order for a Power E870 or E880 with a single system
node:
Table 2-5 Optimal CDIMM memory quad placement for a single system node
System Node 1
Processor 0Processor 2Processor 3Processor 1
Quad 1Quad 5Quad 2Quad 6Quad 3Quad 7Quad 4Quad 8
15 2637 4 8
Notes:
Memory quads 1-4 must be populated.
Memory quads on the same processor must be populated with CDIMMs of the same capacity.
Table 2-6 shows the CDIMM plugging order for a Power E870 or E880 with two system
nodes:
Table 2-6 Optimal CDIMM memory quad placement for two system nodes
System Node 1
Processor 0Processor 2Processor 3Processor 1
Quad 1Quad 5Quad 2Quad 6Quad 3Quad 7Quad 4Quad 8
19 2113134 15
System Node 2
Processor 0Processor 2Processor 3Processor 1
Quad 1Quad 5Quad 2Quad 6Quad 3Quad 7Quad 4Quad 8
5106127148 16
Notes:
Memory quads 1-8 must be populated.
Memory quads on the same processor must be populated with CDIMMs of the same capacity.
52IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 67
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
Table 2-7 shows the CDIMM plugging order for a Power E880 with a three system nodes:
Table 2-7 Optimal CDIMM memory quad placement for three system nodes
System Node 1
Processor 0Processor 2Processor 3Processor 1
Quad 1Quad 5Quad 2Quad 6Quad 3Quad 7Quad 4Quad 8
1132163194 22
System Node 2
Processor 0Processor 2Processor 3Processor 1
Quad 1Quad 5Quad 2Quad 6Quad 3Quad 7Quad 4Quad 8
5146177208 23
System Node 3
Processor 0Processor 2Processor 3Processor 1
Quad 1Quad 5Quad 2Quad 6Quad 3Quad 7Quad 4Quad 8
915101811211224
Notes:
Memory quads 1-12 must be populated.
Memory quads on the same processor must be populated with CDIMMs of the same capacity.
Chapter 2. Architecture and technical overview 53
Page 68
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
Table 2-8 shows the CDIMM plugging order for a Power E880 with a four system nodes:
Table 2-8 Optimal CDIMM memory quad placement for four system nodes
System Node 1
Processor 0Processor 2Processor 3Processor 1
Quad 1Quad 5Quad 2Quad 6Quad 3Quad 7Quad 4Quad 8
1172213254 29
System Node 2
Processor 0Processor 2Processor 3Processor 1
Quad 1Quad 5Quad 2Quad 6Quad 3Quad 7Quad 4Quad 8
5186227268 30
System Node 3
Processor 0Processor 2Processor 3Processor 1
Quad 1Quad 5Quad 2Quad 6Quad 3Quad 7Quad 4Quad 8
919102311271231
Processor 0Processor 2Processor 3Processor 1
Quad 1Quad 5Quad 2Quad 6Quad 3Quad 7Quad 4Quad 8
1320142415281632
Notes:
Memory quads 1-16 must be populated.
Memory quads on the same processor must be populated with CDIMMs of the same capacity.
2.3.3 Memory activation
All the memory CDIMMs are capable of capacity upgrade on demand and must have a
minimum of 50% of its physical capacity activated. For example, the minimum installed
memory for both Power E870 and Power E880 is 256 GB of RAM, whereas they can have a
minimum of 128 GB of RAM active.
There are two activation types that can be used to accomplish this:
Static memory activations: Memory activations that are exclusive for a single server
Mobile memory activations: Memory activations that can be moved from server to server
in a power enterprise pool.
Both types of memory activations can co-reside in the same system, as long as at least 25%
of the memory activations are static. This leads to a maximum of 75% of the memory
activations as mobile.
System Node 4
54IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 69
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
1024 GB
512 GB
768 GB
256 GB
Installed memor y
Minimum active
Minimum Static active
Maximum Mobile active
50% of total memory
75% of installed memory
25% of installed memory
Without Power Enterprise Pools
With Power Enterprise Pools
On Figure 2-15 on page 55 there is an example of the minimum required activations for a
system with 1 TB of installed memory.
Figure 2-15 Example of the minimum required activations for a system with 1 TB of installed memory
2.3.4 Memory throughput
The granularity for static memory activation is 1 GB, while for mobile memory activation the
granularity is 100 GB. On Table 2-9 there is a list of the feature codes that can be used to
achieve the desired number of activations:
Table 2-9 Static and mobile memory activation feature codes
Feature codeFeature descriptionAmount of memoryType of activation
EMA51 GB Memory activation1 GBStatic
EMA6100 GB Memory activation100 GBStatic
EMA9100 GB Mobile memory activation100 GBMobile
Static memory activations can be converted to mobile memory activations after system
installation. In order to enable mobile memory activations, the systems must be part of a
power enterprise pool and have feature code #EB35 configured. For more information on
power enterprise pools, please see 2.4.2, “Power enterprise pools and mobile capacity on
demand (Mobile CUoD)” on page 62.
The peak memory and I/O bandwidths per system node have increased over 300% compared
to previous POWER7 servers, providing the next generation of data intensive applications
with a platform capable of handling the needed amount of data.
Power E870
Table 2-10 shows the maximum bandwidth estimates for a single core on the Power E870
system.
Chapter 2. Architecture and technical overview 55
Page 70
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
Table 2-10 Power E870 single core bandwidth estimates
Single corePower E870Power E870
1 core @ 4.024 GHz1 core @ 4.190 GHz
L1 (data) cache193.15 GBps201.12 GBps
L2 cache193.15 GBps201.12 GBps
L3 cache257.54 GBps268.16 GBps
The bandwidth figures for the caches are calculated as follows:
L1 cache: In one clock cycle, two 16-byte load operations and one 16-byte store operation
can be accomplished. The value varies depending on the clock of the core and the formula
is as follows:
– 4.024 GHz Core: (2 * 16 B + 1 * 16 B) * 4.024 GHz = 193.15 GBps
– 4.190 GHz Core: (2 * 16 B + 1 * 16 B) * 4.190 GHz = 201.12 GBps
L2 cache: In one clock cycle, one 32-byte load operation and one 16-byte store operation
can be accomplished. The value varies depending on the clock of the core and the formula
is as follows:
– 4.024 GHz Core: (1 * 32 B + 1 * 16 B) * 4.024 GHz = 193.15 GBps
– 4.190 GHz Core: (1 * 32 B + 1 * 16 B) * 4.190 GHz = 201.12 GBps
L3 cache: In one clock cycle, one 32-byte load operation and one 32-byte store operation
can be accomplished. The value varies depending on the clock of the core and the formula
is as follows:
– 4.024 GHz Core: (1 * 32 B + 1 * 32 B) * 4.024 GHz = 257.54 GBps
– 4.190 GHz Core: (1 * 32 B + 1 * 32 B) * 4.190 GHz = 286.16 GBps
For each system node of a Power E870 populated with four processors and all its memory
CDIMMs filled, the overall bandwidths are shown in Table 2-11.
Table 2-11 Power E870 system node bandwidth estimates
System node bandwidthsPower E870Power E870
32 cores @ 4.024 GHz40 cores @ 4.190 GHz
L1 (data) cache6,181 GBps8,045 GBps
L2 cache6,181 GBps8,045 GBps
L3 cache8,241 GBps10,726 GBps
Total Memory 922 GBps922 GBps
PCIe Interconnect252.064 GBps252.064 GBps
Intra-node buses
(two system nodes)
922 GBps922 GBps
PCIe Interconnect: Each POWER8 processor has 32 PCIe lanes running at 7.877Gbps
full-duplex. The bandwidth formula is calculated as follows:
Rounding: The bandwidths listed here may appear slightly differently in other materials
due to rounding of some figures.
56IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 71
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
For the entire Power E870 system populated with two system nodes, the overall bandwidths
are what is shown in Table 2-12 on page 57.
Table 2-12 Power E870 total bandwidth estimates
Total bandwidthsPower E870Power E870
64 cores @ 4.024 GHz80 cores @ 4.190 GHz
L1 (data) cache12,362 GBps16,090 GBps
L2 cache12,362 GBps16,090 GBps
L3 cache16,484 GBps21,453 GBps
Total Memory 1,844 GBps1,844 GBps
PCIe Interconnect504.128 GBps504.128 GBps
Inter-node buses (two
system nodes)
Intra-node buses (two
system nodes)
307 GBps307 GBps
1,844 GBps1,844 GBps
Power E880
Table 2-13 shows the maximum bandwidth estimates for a single core on the Power E880
system.
Table 2-13 Power E880 single core bandwidth estimates
Single corePower E880
1 core @ 4.350 GHz
L1 (data) cache208.80 GBps
L2 cache208.80 GBps
L3 cache278.40 GBps
The bandwidth figures for the caches are calculated as follows:
L1 cache: In one clock cycle, two 16-byte load operations and one 16-byte store operation
can be accomplished. The value varies depending on the clock of the core and the formula
is as follows:
– 4.350 GHz Core: (2 * 16 B + 1 * 16 B) * 4.350 GHz = 208.80 GBps
L2 cache: In one clock cycle, one 32-byte load operation and one 16-byte store operation
can be accomplished. The value varies depending on the clock of the core and the formula
is as follows:
– 4.350 GHz Core: (1 * 32 B + 1 * 16 B) * 4.350 GHz = 208.80 GBps
L3 cache: In one clock cycle, one 32-byte load operation and one 32-byte store operation
can be accomplished. The value varies depending on the clock of the core and the formula
is as follows:
– 4.350 GHz Core: (1 * 32 B + 1 * 32 B) * 4.350 GHz = 278.40 GBps
For each system node of a Power E880 populated with four processors and all its memory
CDIMMs filled, the overall bandwidths are shown in Table 2-14 on page 58.
Chapter 2. Architecture and technical overview 57
Page 72
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
Table 2-14 Power E880 system node bandwidth estimates
System node bandwidthsPower E880
32 cores @ 4.350 GHz
L1 (data) cache6,682 GBps
L2 cache6,682 GBps
L3 cache8,909 GBps
Total Memory 922 GBps
PCIe Interconnect252.064 GBps
Intra-node buses
(two system nodes)
922 GBps
PCIe Interconnect: Each POWER8 processor has 32 PCIe lanes running at 7.877 Gbps
full-duplex. The bandwidth formula is calculated as follows:
Rounding: The bandwidths listed here may appear slightly differently in other materials
due to rounding of some figures.
For the entire Power E880 system populated with four system nodes, the overall bandwidths
are shown in Table 2-15.
Table 2-15 Power E880 total bandwidth estimates
Total bandwidthsPower E880
128 cores @ 4.350 GHz
L1 (data) cache26,726 GBps
L2 cache26,726 GBps
L3 cache35,635 GBps
Total Memory 3,688 GBps
PCIe Interconnect1008.256 GBps
Inter-node buses (four
system nodes)
Intra-node buses (four
system nodes)
Note: At the time of writing, the Power E880 supports only two system nodes per server.
2.3.5 Active Memory Mirroring
The Power E870 and Power E880 systems have the ability to provide mirroring of the
hypervisor code across multiple memory CDIMMs. If a CDIMM that contains the hypervisor
code develops an uncorrectable error, its mirrored partner will enable the system to continue
to operate uninterrupted.
58IBM Power Systems E870 and E880 Technical Overview and Introduction
307 GBps
3,688 GBps
Page 73
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
Active Memory Mirroring (AMM) is included with all Power E870 and Power E880 systems at
no additional charge. It can be enabled, disabled, or re-enabled depending on the user’s
requirements.
The hypervisor code logical memory blocks will be mirrored on distinct CDIMMs to allow for
more usable memory. There is no specific CDIMM that hosts the hypervisor memory blocks
so the mirroring is done at the logical memory block level, not at the CDIMM level. To enable
the AMM feature it is mandatory that the server has enough free memory to accommodate
the mirrored memory blocks.
Besides the hypervisor code itself, other components that are vital to the server operation are
also mirrored:
Hardware page tables (HPTs), responsible for tracking the state of the memory pages
assigned to partitions
Translation control entities (TCEs), responsible for providing I/O buffers for the
partition’s communications
Memory used by the hypervisor to maintain partition configuration, I/O states, virtual I/O
information, and partition state
It is possible to check whether the Active Memory Mirroring option is enabled and change its
current status through HMC, under the Advanced Tab on the CEC Properties panel
(Figure 2-16).
Figure 2-16 CEC Properties panel on an HMC
After a failure on one of the CDIMMs containing hypervisor data occurs, all the server
operations remain active and flexible service processor (FSP) will isolate the failing CDIMMs.
Systems stay in the partially mirrored state until the failing CDIMM is replaced.
There are components that are not mirrored because they are not vital to the regular server
operations and require a larger amount of memory to accommodate its data:
Advanced Memory Sharing Pool
Memory used to hold the contents of platform dumps
Chapter 2. Architecture and technical overview 59
Page 74
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
Partition data: Active Memory Mirroring will not mirror partition data. It was designed to
mirror only the hypervisor code and its components, allowing this data to be protected
against a DIMM failure
With AMM, uncorrectable errors in data that are owned by a partition or application are
handled by the existing Special Uncorrectable Error handling methods in the hardware,
firmware, and operating system.
2.3.6 Memory Error Correction and Recovery
The memory has error detection and correction circuitry is designed such that the failure of
any one specific memory module within an ECC word can be corrected without any other
fault.
In addition, a spare DRAM per rank on each memory port provides for dynamic DRAM device
replacement during runtime operation. Also, dynamic lane sparing on the DMI link allows for
repair of a faulty data lane.
Other memory protection features include retry capabilities for certain faults detected at both
the memory controller and the memory buffer.
Memory is also periodically scrubbed to allow for soft errors to be corrected and for solid
single-cell errors reported to the hypervisor, which supports operating system deallocation of
a page associated with a hard single-cell fault.
For more details on Memory RAS, see 4.3.10, “Memory protection” on page 153.
2.3.7 Special Uncorrectable Error handling
Special Uncorrectable Error (SUE) handling prevents an uncorrectable error in memory or
cache from immediately causing the system to terminate. Rather, the system tags the data
and determines whether it will ever be used again. If the error is irrelevant, it does not force a
checkstop. If the data is used, termination can be limited to the program/kernel or hypervisor
owning the data, or freeze of the I/O adapters controlled by an I/O hub controller if data is to
be transferred to an I/O device.
2.4 Capacity on Demand
Several types of Capacity on Demand (CoD) offerings are optionally available on the
Power 870 and Power E880 servers to help meet changing resource requirements in an
on-demand environment, by using resources that are installed on the system but that are not
activated.
2.4.1 Capacity Upgrade on Demand (CUoD)
Power E870 and Power E880 systems include a number of active processor cores and
memory units. They can also include inactive processor cores and memory units. Active
processor cores or memory units are processor cores or memory units that are already
available for use on your server when it comes from the manufacturer. Inactive processor
cores or memory units are processor cores or memory units that are included with your
server, but not available for use until you activate them. Inactive processor cores and memory
60IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 75
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
0
10
20
30
40
50
60
70
80
Without CuOD
With CuOD
Time
Cores
Projected
Core Activations
units can be permanently activated by purchasing an activation feature called Capacity
Upgrade on Demand (CUoD) and entering the provided activation code on your server.
With the CUoD offering, you can purchase additional static processor or memory capacity
and dynamically activate them when needed, without requiring you to restart your server or
interrupt your business. All the static processor or memory activations are restricted to a
single server.
Capacity Upgrade on Demand can have several applications to allow for a more flexible
environment. One of its benefits is to allow for a given company to reduce the initial
investment on a system. Traditional projects using other technologies require that the a
system is acquired with all the resources available to support the whole lifecycle of the
project. This might incur in costs that would only be necessary on later stages of the project,
usually with impacts on software licensing costs and software maintenance.
By using Capacity Upgrade on Demand a company could start with a system with enough
installed resources to support the whole project lifecycle but only with enough active
resources necessary for the initial project phases. More resources could be added along with
the project, adjusting the hardware platform with the project needs. This would allow for a
company to reduce the initial investment in hardware and only acquire software licenses that
are needed on each project phase, reducing the Total Cost of Ownership and Total Cost of
Acquisition of the solution. Figure 2-17 shows a comparison between two scenarios: a fully
activated system versus a system with CUoD resources being activated along with the project
timeline.
Figure 2-17 Active cores scenarios comparison during a project lifecycle
The Table 2-16 on page 62 lists the static processor activation features that are available for
the Power E870 and Power E880.
Chapter 2. Architecture and technical overview 61
Page 76
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
Table 2-16 Power Systems Capacity Upgrade on Demand static processor activation features
The Table 2-17 lists the static memory activation features that are available for the
Power E870 and Power E880.
Table 2-17 Power Systems Capacity Upgrade on Demand static memory activation features
SystemDescriptionFeature code
Power E870, Power E880Activation of 1 GB DDR3 POWER8 memory#EMA5
Power E870, Power E880Activation of 100 GB DDR3 POWER8 memory#EMA6
2.4.2 Power enterprise pools and mobile capacity on demand (Mobile CUoD)
While static activations are valid for a single system, some customers could benefit from
moving processor and memory activations among different servers due to workload
rebalance or disaster recovery.
IBM power enterprise pools, is a technology for dynamically sharing processor and memory
activations among a group (or pool) of IBM Power Systems servers. Using Mobile Capacity on
Demand (CoD) activation codes, the systems administrator can perform tasks without
contacting IBM.
Two types of power enterprise pools are available:
Power 770 (9117-MMD) and Power E870 (9119-MME) class systems
Power 780 (9117-MHD), Power 795 (9119-FHB), and Power E870 (9119-MHE) class
systems
Each pool type can support systems with different clock speeds or processor generations.
The basic rules for the Mobile Capacity on Demand follows:
The Power 770 and Power 780 systems require a minimum of four static processor
activations.
The Power 870 and Power 880 require a minimum of eight static processor activations.
The Power 795 requires a minimum of 24 static processor activations or 25% of the
installed processor capacity whichever is bigger.
For all systems, 25% of the installed processor capacity must have static activations.
All the systems in a pool must be managed by the same HMC or by the same pair of
redundant HMCs. If redundant HMCs are used, the HMCs must be connected to a network so
that they can communicate with each other. The HMCs must have at least 2 GB of memory.
An HMC can manage multiple power enterprise pools and can also manage systems that are
not part of a power enterprise pool. Systems can belong to only one power enterprise pool at
a time. Powering down an HMC does not limit the assigned resources of participating
systems in a pool but does limit the ability to perform pool change operations.
62IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 77
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
After a power enterprise pool is created, the HMC can be used to perform the following
functions:
Mobile CoD processor and memory resources can be assigned to systems with inactive
resources. Mobile CoD resources remain on the system to which they are assigned until
they are removed from the system.
New systems can be added to the pool and existing systems can be removed from the
pool.
New resources can be added to the pool or existing resources can be removed from the
pool.
Pool information can be viewed, including pool resource assignments, compliance, and
history logs.
In order for the Mobile activation features to be configured it is necessary that a a power
enterprise pool is registered with IBM as well as the systems which are going to be included
as members of the pool. Also it is necessary that the systems have the feature code #EB35
for mobile enablement configured and the required contracts must be in place.
The Table 2-18 lists the mobile processor activation features that are available for the
Power E870 and Power E880.
Table 2-18 Mobile processor activation features
SystemDescriptionCUoD mobile processor core
Power E8701-Core Mobile activation#EP2S
Power E8801-Core Mobile activation#EP2T
The Table 2-19 lists the mobile memory activation features that are available for the
Power E870 and Power E880.
Table 2-19 Mobile memory activation features
SystemDescriptionFeature code
Power E870, Power E880100 GB Mobile memory activation#EMA4
For more information on power enterprise pools, please see Redpaper “power enterprise
pools on IBM Power Systems”.
With the elastic CoD offering, you can temporarily activate and deactivate processor cores
and memory units to help meet the demands of business peaks such as seasonal activity,
period-end, or special promotions. Elastic CoD was previously called On/Off CoD. When you
order an elastic CoD feature, you receive an enablement code that allows a system operator
to make requests for additional processor and memory capacity in increments of one
processor day or 1 GB memory day. The system monitors the amount and duration of the
activations. Both prepaid and post-pay options are available.
activation feature
Charges are based on usage reporting that is collected monthly. Processors and memory
may be activated and turned off an unlimited number of times, when additional processing
resources are needed.
Chapter 2. Architecture and technical overview 63
Page 78
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
This offering provides a system administrator an interface at the HMC to manage the
activation and deactivation of resources. A monitor that resides on the server records the
usage activity. This usage data must be sent to IBM on a monthly basis. A bill is then
generated based on the total amount of processor and memory resources utilized, in
increments of Processor and Memory (1 GB) Days.
New to both Power E870 and Power E880 are 90-day temporary elastic CoD processor and
memory enablement features. These features enable a system to temporarily activate all
inactive processor and memory CoD resources for a maximum of 90 days before ordering
another temporary elastic enablement feature code is required.
Before using temporary capacity on your server, you must enable your server. To enable, an
enablement feature (MES only) must be ordered and the required contracts must be in place.
If a Power E870 or Power E880 server uses the IBM i operating system in addition to any
other supported operating system on the same server, the client must inform IBM which
operating system caused the temporary elastic CoD processor usage so that the correct
feature can be used for billing.
The features that are used to order enablement codes and support billing charges on the
Power E870 and Power E880 are described in 1.4, “System features” on page 9 and 1.4.7,
“Memory features” on page 16.
The elastic CoD process consists of three steps: enablement, activation, and billing.
Enablement
Before requesting temporary capacity on a server, you must enable it for elastic CoD. To
do this, order an enablement feature and sign the required contracts. IBM will generate an
enablement code, mail it to you, and post it on the web for you to retrieve and enter on the
target server.
A
processor enablement code allows you to request up to 360 processor days of
temporary capacity. If the 360 processor-day limit is reached, place an order for another
processor enablement code to reset the number of days that you can request back to 360.
A
memory enablement code lets you request up to 999 memory days of temporary
capacity. If you reach the limit of 999 memory days, place an order for another memory
enablement code to reset the number of allowable days you can request back to 999.
Activation requests
When elastic CoD temporary capacity is needed, use the HMC menu for On/Off CoD.
Specify how many inactive processors or gigabytes of memory are required to be
temporarily activated for some number of days. You are billed for the days requested,
whether the capacity is assigned to partitions or remain in the shared processor pool.
At the end of the temporary period (days that were requested), you must ensure that the
temporarily activated capacity is available to be reclaimed by the server (not assigned to
partitions), or you are billed for any unreturned processor days.
Billing
The contract, signed by the client before receiving the enablement code, requires the
elastic CoD user to report billing data at least once a month (whether or not activity
occurs). This data is used to determine the proper amount to bill at the end of each billing
period (calendar quarter). Failure to report billing data for use of temporary processor or
memory capacity during a billing quarter can result in default billing equivalent to 90
processor days of temporary capacity.
64IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 79
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
For more information about registration, enablement, and usage of elastic CoD, visit the
following location:
http://www.ibm.com/systems/power/hardware/cod
2.4.4 Utility Capacity on Demand (Utility CoD)
Utility CoD automatically provides additional processor performance on a temporary basis
within the shared processor pool.
With Utility CoD, you can place a quantity of inactive processors into the server’s shared
processor pool, which then becomes available to the pool's resource manager. When the
server recognizes that the combined processor utilization within the shared processor pool
exceeds 100% of the level of base (purchased and active) processors that are assigned
across uncapped partitions, then a Utility CoD processor minute is charged and this level of
performance is available for the next minute of use.
If additional workload requires a higher level of performance, the system automatically allows
the additional Utility CoD processors to be used, and the system automatically and
continuously monitors and charges for the performance needed above the base (permanent)
level.
Registration and usage reporting for utility CoD is made using a public website and payment
is based on reported usage. Utility CoD requires PowerVM Standard Edition or PowerVM
Enterprise Edition to be active.
If a Power E870 or Power E880 server uses the IBM i operating system in addition to any
other supported operating system on the same server, the client must inform IBM which
operating system caused the temporary Utility CoD processor usage so that the correct
feature can be used for billing.
For more information regarding registration, enablement, and use of Utility CoD, visit the
following location:
A standard request for Trial CoD requires you to complete a form including contact information
and vital product data (VPD) from your Power E870 or Power E880 system with inactive CoD
resources.
A standard request activates two processors or 4 GB of memory (or both two processors and
4 GB of memory) for 30 days. Subsequent standard requests can be made after each
purchase of a permanent processor activation. An HMC is required to manage Trial CoD
activations.
An
exception request for Trial CoD requires you to complete a form including contact
information and VPD from your Power E870 or Power E880 system with inactive CoD
resources. An exception request will activate all inactive processors or all inactive memory (or
all inactive processor and memory) for 30 days. An exception request can be made only one
time over the life of the machine. An HMC is required to manage Trial CoD activations.
To request either a Standard or an Exception Trial, visit the following location:
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
I/O Backplane
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
P8 (0)
PHB 1
PHB 0
SMP
X Bus
SMP
A Bus
Memory
Controller 0
Memory
Controller 1
P8 (2)
PHB 1
PHB 0
SMP
X Bus
SMP
A Bus
Memory
Controller 0
Memory
Controller 1
P8 (3)
PHB 1
PHB 0
SMP
X Bus
SMP
A Bus
Memory
Controller 0
Memory
Controller 1
P8 (1)
Memory
Controller 0
Memory
Controller 1
PHB 1
PHB 0
SMP
X Bus
SMP
A Bus
2.4.6 Software licensing and CoD
For software licensing considerations with the various CoD offerings, see the most recent
revision of the Power Systems Capacity on Demand User’s Guide:
http://www.ibm.com/systems/power/hardware/cod
2.5 System bus
This section provides additional information related to the internal buses.
2.5.1 PCI Express Gen3
The internal I/O subsystem on the Power E870 and Power E880 is connected to the PCIe
Controllers on a POWER8 processor in the system. Each POWER8 processor module has
two buses that have 16 PCIe lanes each (for a total of 32 PCIe lanes) running at 7.877 Gbps
full-duplex and provides 31.508 GBps of I/O connectivity to the PCIe slots. A diagram with the
connections can be seen in Figure 2-18.
Figure 2-18 System nodes PCIe slots directly attached to PCIe controllers on POWER8 chips
Besides the slots directly attached to the processors PCI Gen3 controllers, the systems also
66IBM Power Systems E870 and E880 Technical Overview and Introduction
allow for additional PCIe adapters on external PCIe Expansion Drawers and disks on external
drawers connected through PCIe SAS adapters.
Figure 2-19 on page 67 shows a diagram with the I/O connectivity options available for the
Power E870 and Power E880. The system nodes allow for eight PCIe Gen3 x16 slots.
Additional slots can be added by attaching PCIe Expansion Drawers and SAS disks can be
Page 81
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
PCIe Gen3 x16
PCIe Adapter for
Expansion Dr awer
PCIe Gen3 x16
PCIe Gen3 x16
PCIe Gen3 x16
PCIe Gen3 x16
PCIe Gen3 x16
SAS Adapter
System Node
EXP24S SFF Gen2 Drawer
PCIe Expansion Drawer
P
C
I
e
G
e
n
3
x
8
P
C
I
e
G
e
n
3
x
8
P
C
I
e
G
e
n
3
x
8
P
C
I
e
G
e
n
3
x
8
P
C
I
e
G
e
n
3
x
8
P
C
I
e
G
e
n
3
x
8
P
C
I
e
G
e
n
3
x
8
P
C
I
e
G
e
n
3
x
8
P
C
I
e
G
e
n
3
x
8
P
C
I
e
G
e
n
3
x
8
P
C
I
e
G
e
n
3
x
8
S
A
S
A
d
a
p
t
e
r
Disks
PCIe Adapters
EXP24S SFF Gen2 Drawer
CPU
1
CPU
0
CPU
2
CPU
3
attached to EXP24S SFF Gen2 Drawers. The EXP24S can be either attached to SAS
adapters on the system nodes or on the PCIe Expansion Drawer.
For a list of adapters and their supported slots, please see 2.7, “PCI adapters” on page 70.
I
2.5.2 Service Processor Bus
Figure 2-19 I/O connectivity options available for Power E870 and Power E880
Disk support: There is no support for disks directly installed on the system nodes and
PCIe Expansion Drawers. If directly attached SAS disk are required, they need to be
installed in a SAS disk drawer and connected to a supported SAS controller in one of the
PCIe slots.
For more information on PCIe Expansion Drawers, please see 2.9.1, “PCIe Gen3 I/O
expansion drawer” on page 77.
The redundant service processor bus connectors are located on the rear of the control unit
and the system nodes. All the service processor (SP) communication between the control unit
and the system nodes flows though these cables.
Unlike the previous generations where a given pair of enclosures would host the service
processors, on Power E870 and Power E880 as a standard, redundant service processor
cards are installed on the control unit as well as redundant clock cards.
The cables used to provide communications between the control units and system nodes
depend on the amount of system nodes installed. When a system node is added, a new set of
cables is also added.
Chapter 2. Architecture and technical overview 67
Page 82
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
The cables necessary for each system node have been grouped under a single feature code,
allowing for an easier configuration. Each cable set includes a pair of FSP cables, a pair of
clock cables, and when applicable SMP cables and UPIC cables. Table 2-20 shows a list of
the feature codes available:
Table 2-20 Features for cable sets
Feature codeDescription
ECCASystem node to system control unit cable set for drawer 1
ECCBSystem node to system control unit cable set for drawer 2
ECCCSystem node to system control unit cable set for drawer 3
ECCDSystem node to system control unit cable set for drawer 4
Cable sets feature codes are incremental and depend on the number of installed drawers as
follows:
1 system node: #ECCA
2 system nodes: #ECCA and #ECCB
3 system nodes: #ECCA, #ECCB, and #ECCC
4 system nodes: #ECCA, #ECCB, #ECCC, and #ECCD
Note: As the time of writing, up to two system nodes are supported for Power E870 and
Power E880.
System connection topology is shown in 2.1, “Logical diagrams” on page 32.
2.6 Internal I/O Subsystem
The internal I/O subsystem resides on the I/O planar, which supports eight PCIe Gen3 x16
slots. All PCIe slots are hot-pluggable and enabled with enhanced error handling (EEH). In
the unlikely event of a problem, EEH-enabled adapters respond to a special data packet that
is generated from the affected PCIe slot hardware by calling system firmware, which
examines the affected bus, allows the device driver to reset it, and continues without a system
reboot. For more information about RAS on the I/O buses, see 4.3.11, “I/O subsystem
availability and Enhanced Error Handling” on page 154.
Table 2-21 lists the slot configuration of a Power E870 and Power E880 system nodes.
Table 2-21 Slot configuration and capabilities
SlotLocation codeSlot TypeCAPI capable
Slot 1P1-C1PCIe Gen3 x16NoYes
Slot 2P1-C2PCIe Gen3 x16YesYes
a
SRIOV capable
Slot 3P1-C3PCIe Gen3 x16NoYes
Slot 4P1-C4PCIe Gen3 x16YesYes
Slot 5P1-C5PCIe Gen3 x16NoYes
Slot 6P1-C6PCIe Gen3 x16YesYes
Slot 7P1-C7PCIe Gen3 x16NoYes
68IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 83
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
I/O Backplane
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
PCIe Gen3 x16 slot
P8 (0)
PHB 1
PHB 0
SMP
X Bus
SMP
A Bus
Memory
Controller 0
Memory
Controller 1
P8 (2)
PHB 1
PHB 0
SMP
X Bus
SMP
A Bus
Memory
Controller 0
Memory
Controller 1
P8 (3)
PHB 1
PHB 0
SMP
X Bus
SMP
A Bus
Memory
Controller 0
Memory
Controller 1
P8 (1)
Memory
Controller 0
Memory
Controller 1
PHB 1
PHB 0
SMP
X Bus
SMP
A Bus
P1-C1
P1-C2
P1-C3
P1-C4
P1-C5
P1-C6
P1-C7
P1-C8
Front
Rear
System Node top view
SlotLocation codeSlot TypeCAPI capable
a
SRIOV capable
Slot 8P1-C8PCIe Gen3 x16YesYes
a. At the time of writing there are no supported CAPI adapters for the Power E870 and E880
system unit.
The physical location of the slots can be seen on Figure 2-20.
2.6.1 Blind-swap cassettes
Figure 2-20 System node top view and PCIe slot location codes
The Power E870 and Power E880 use a next generation blind-swap cassette to manage the
installation and removal of PCIe adapters. This mechanism requires an interposer card that
allows the PCIe adapters to plug in vertically to the system, allows more airflow through the
cassette, and allows for faster hot swap procedures. Cassettes can be installed and removed
without removing the system nodes or PCIe expansion drawers from the rack.
Chapter 2. Architecture and technical overview 69
Page 84
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
FSP 1FSP 2
HMC1
HMC2
HMC1
HMC2
USB
2.6.2 System ports
The system nodes do not have integrated ports. All networking and storage for the virtual
machines must pre provided via PCIe adapters installed in standard PCIe slots.
The system control unit has one USB Port dedicated to the DVD drive and 4 Ethernet ports
used for HMC communications. There is no serial port so an HMC is mandatory for system
management. The FSP’s virtual console will be on the HMC.
The location of the USB and HMC Ethernet ports can be seen on Figure 2-21
Figure 2-21 Physical location of the USB and HMC ports on the system control unit
The connection and usage of the DVD can be seen on detail on 2.8.1, “DVD” on page 76.
2.7 PCI adapters
This section covers the types and functions of the PCI cards supported by IBM Power E870
and IBM Power E880 systems.
2.7.1 PCI Express (PCIe)
PCIe uses a serial interface and allows for point-to-point interconnections between devices
(using a directly wired interface between these connection points). A single PCIe serial link is
a dual-simplex connection that uses two pairs of wires, one pair for transmit and one pair for
receive, and can transmit only one bit per cycle. These two pairs of wires are called a
PCIe link can consist of multiple lanes. In such configurations, the connection is labelled as
x1, x2, x8, x12, x16, or x32, where the number is effectively the number of lanes.
The PCIe interfaces supported on this server are PCIe Gen3, capable of 16 GBps simplex
(32 GBps duplex) on a single x16 interface. PCIe Gen3 slots also support previous
generations (Gen2 and Gen1) adapters, which operate at lower speeds, according to the
following rules:
Place x1, x4, x8, and x16 speed adapters in same connector size slots first, before mixing
adapter speed with connector slot size.
lane. A
Adapters with smaller speeds are allowed in larger sized PCIe connectors but larger
speed adapters are not compatible in smaller connector sizes (i.e. a x16 adapter cannot
go in an x8 PCIe slot connector).
70IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 85
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
IBM POWER8 processor-based servers can support two different form factors of PCIe
adapters:
PCIe low profile (LP) cards, which are used with system node PCIe slots.
PCIe full height and full high cards are used in the PCIe Gen3 I/O expansion drawer
(#EMX0)
Low-profile PCIe adapter cards are supported only in low-profile PCIe slots, and full-height
and full-high cards are supported only in full-high slots.
Before adding or rearranging adapters, use the System Planning Tool to validate the new
adapter configuration. For more information, see the System Planning Tool website:
If you are installing a new feature, ensure that you have the software that is required to
support the new feature and determine whether there are any existing update prerequisites to
install. To do this, use the IBM prerequisite website:
https://www-912.ibm.com/e_dir/eServerPreReq.nsf
The following sections describe the supported adapters and provide tables of orderable
feature numbers. The tables indicate operating system support (AIX, IBM i, and Linux) for
each of the adapters.
2.7.2 LAN Adapters
To connect the Power E870 and Power E880 servers to a local area network (LAN), you can
use the LAN adapters that are supported in the PCIe slots of the system. Table 2-22 lists the
available LAN adapters.
Table 2-22 Available LAN adapters
Feature
Code
5260576FPCIe2 LP 4-port 1 GbE AdapterCECAIX, IBM i,
52875287PCIe2 2-port 10 GbE SR AdapterI/O drawerAIX, Linux
571757174-Port 10/100/1000 Base-TX PCI Express
576757672-Port 10/100/1000 Base-TX Ethernet PCI
576857682-Port Gigabit Ethernet-SX PCI Express
5769576910 Gigabit Ethernet-SR PCI Express AdapterI/O drawerAIX, IBM i,
5772576E10 Gigabit Ethernet-LR PCI Express AdapterI/O drawerAIX, IBM i,
CCINDescriptionPlacementOS support
Linux
I/O drawerAIX, Linux
Adapter
I/O drawerAIX, IBM i,
Express Adapter
I/O drawerAIX, IBM i,
Adapter
Linux
Linux
Linux
Linux
5899576FPCIe2 4-port 1 GbE AdapterI/O drawerAIX, IBM i,
EC28EC27PCIe2 2-Port 10 GbE RoCE SFP+ AdapterI/O drawerAIX, Linux
EC29EC29PCIe2 LP 2-Port 10 GbE RoCE SR AdapterCECAIX, Linux
Linux
Chapter 2. Architecture and technical overview 71
Page 86
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
EC30EC29PCIe2 2-Port 10 GbE RoCE SR AdapterI/O drawerAIX, Linux
EC3A57BDPCIe3 LP 2-Port 40 GbE NIC RoCE QSFP+
EC3B57B6PCIe3 2-Port 40 GbE NIC RoCE QSFP+
EN0S2CC3PCIe2 4-Port (10Gb+1 GbE) SR+RJ45
EN0U2CC3PCIe2 4-port (10Gb+1 GbE) Copper
EN0W2CC4PCIe2 2-port 10/1 GbE BaseT RJ45 AdapterI/O drawerAIX, Linux
EN0X2CC4PCIe2 LP 2-port 10/1 GbE BaseT RJ45
CCINDescriptionPlacementOS support
Adapter
Adapter
Adapter
SFP+RJ45 Adapter
Adapter
2.7.3 Graphics accelerator adapters
Table 2-23 lists the available graphics accelerator adapters. An adapter can be configured to
operate in either 8-bit or 24-bit color modes. The adapter supports both analog and digital
monitors.
CECAIX, Linux
I/O drawerAIX, Linux
I/O drawerAIX, Linux
I/O drawerAIX, Linux
CECAIX, Linux
Table 2-23 Available graphics accelerator adapters
Feature
Code
57485748POWER GXT145 PCI Express Graphics
EC41PCIe2 LP 3D Graphics AdapterCECLinux
2.7.4 SAS adapters
Table 2-24 lists the SAS adapters that are available for Power E870 and Power E880
systems.
Table 2-24 Available SAS adapters
Featur
e Code
590157B3PCIe Dual-x4 SAS AdapterI/O drawerAIX, IBM i,
591357B5CIe2 1.8GB Cache RAID SAS Adapter Tri-port
ESA357BBPCIe2 1.8GB Cache RAID SAS Adapter
CCINDescriptionPlacementOS support
I/O drawerAIX, Linux
Accelerator
CCINDescriptionPlacementOS support
Linux
I/O drawerAIX, IBM i,
6Gb
I/O drawerAIX, IBM i,
Tri-port 6Gb CR
Linux
Linux
EJ0J57B4PCIe3 RAID SAS Adapter Quad-port 6Gb x8I/O drawerAIX, IBM i,
72IBM Power Systems E870 and E880 Technical Overview and Introduction
Linux
Page 87
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
Featur
e Code
EJ0L57CEPCIe3 12GB Cache RAID SAS Adapter
EJ0M57B4PCIe3 LP RAID SAS ADAPTERCECAIX, IBM i,
EJ1057B4 PCIe3 SAS Tape/DVD Adapter Quad-port 6Gb x8I/O drawerAIX, IBM i,
EJ1157B4PCIe3 LP SAS Tape/DVD Adapter Quad-port
CCINDescriptionPlacementOS support
Quad-port 6Gb x8
6Gb x8
2.7.5 Fibre Channel adapter
The systems support direct or SAN connection to devices that use Fibre Channel adapters.
Table 2-25 summarizes the available Fibre Channel adapters, which all have LC connectors.
If you are attaching a device or switch with an SC type fiber connector, then an
LC-SC 50 Micron Fiber Converter Cable (#2456) or an LC-SC 62.5 Micron Fiber Converter
Cable (#2459) is required.
Table 2-25 Available Fibre Channel adapters
Feature
Code
CCINDescriptionPlacementOS support
I/O drawerAIX, IBM i,
Linux
Linux
Linux
CECAIX, IBM i,
Linux
5273577DPCIe LP 8Gb 2-Port Fibre Channel AdapterCECAIX, IBM i,
Linux
57295729PCIe2 8Gb 4-port Fibre Channel AdapterI/O drawerAIX, Linux
5735577DGigabit PCI Express Dual Port Fibre Channel
Adapter
577457744 Gigabit PCI Express Dual Port Fibre
Channel Adapter
EN0A577FPCIe2 16Gb 2-port Fibre Channel AdapterI/O drawerAIX, Linux
EN0B577FPCIe2 LP 16Gb 2-port Fibre Channel AdapterCECAIX, IBM i,
EN0YEN0YPCIe2 LP 8Gb 4-port Fibre Channel AdapterCECAIX, Linux
I/O drawerAIX, IBM i,
Linux
I/O drawerAIX, IBM i,
Linux
Linux
Chapter 2. Architecture and technical overview 73
Page 88
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
2.7.6 Fibre Channel over Ethernet
Fibre Channel over Ethernet (FCoE) allows for the convergence of Fibre Channel and
Ethernet traffic onto a single adapter and a converged fabric.
Figure 2-22 compares existing Fibre Channel and network connections and FCoE
connections.
Figure 2-22 Comparison between existing Fibre Channel and network connections and FCoE
connections
Table 2-26 lists the available FCoE adapters. They are high-performance Converged Network
Adapters (CNAs) using SR optics. Each port can simultaneously provide network interface
card (NIC) traffic and Fibre Channel functions.
Table 2-26 Available FCoE adapters
Feature
Code
EN0H2B93PCIe2 4-port (10Gb FCoE & 1 GbE) SR&RJ45I/O drawerAIX, Linux
EN0J2B93PCIe2 LP 4-port (10Gb FCoE & 1 GbE)
EN0K2CC1PCIe2 4-port (10Gb FCoE & 1 GbE)
EN0L2CC1PCIe2 LP 4-port(10Gb FCoE & 1 GbE)
CCINDescriptionPlacementOS support
SR&RJ45
SFP+Copper&RJ45
SFP+Copper&RJ45
For more information about FCoE, see An Introduction to Fibre Channel over Ethernet, and Fibre Channel over Convergence Enhanced Ethernet, REDP-4493.
2.7.7 Asynchronous and USB adapters
CECAIX, Linux
I/O drawerAIX, Linux
CECAIX, Linux
Asynchronous PCI adapters provide connection of asynchronous EIA-232 or RS-422
devices.
Each system control unit enclosure can have one slim-line bay which can support one DVD
drive (#EU13). The #EU13 DVD is cabled to a USB PCIe adapter located in either the system
node or in a PCIe Gen3 I/O drawer.
74IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 89
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
Table 2-27 lists the available asynchronous and USB adapters.
Table 2-27 Available asynchronous and USB adapters
Feature
Code
578557D24 Port Async EIA-232 PCIe AdapterI/O drawerAIX, Linux
EN272 Port Async EIA-232 PCIe AdapterI/O drawerAIX, IBM i, Linux
EC45PCIe2 LP 4-Port USB 3.0 AdapterCECAIX, IBM i, Linux
EC46PCIe2 4-Port USB 3.0 AdapterI/O drawerAIX, IBM i, Linux
CCINDescriptionPlacementOS support
2.7.8 InfiniBand host channel adapter
The InfiniBand architecture (IBA) is an industry-standard architecture for server I/O and
inter-server communication. It was developed by the InfiniBand Trade Association (IBTA) to
provide the levels of reliability, availability, performance, and scalability necessary for present
and future server systems with levels significantly better than can be achieved by using
bus-oriented I/O structures.
InfiniBand (IB) is an open set of interconnect standards and specifications. The main IB
specification is published by the InfiniBand Trade Association and is available at the following
location:
http://www.infinibandta.org/
InfiniBand is based on a switched fabric architecture of serial point-to-point links, where these
IB links can be connected to either host channel adapters (HCAs), used primarily in servers,
or to target channel adapters (TCAs), used primarily in storage subsystems.
The InfiniBand physical connection consists of multiple byte lanes. Each individual byte lane
is a four-wire, 2.5, 5.0, or 10.0 Gbps bidirectional connection. Combinations of link width and
byte-lane speed allow for overall link speeds from 2.5 Gbps to 120 Gbps. The architecture
defines a layered hardware protocol, and also a software layer to manage initialization and
the communication between devices. Each link can support multiple transport services for
reliability and multiple prioritized virtual communication channels.
Table 2-28 lists the available InfiniBand adapters.
Table 2-28 Available InfiniBand adapters
Feature
Code
528558E2PCIe2 2-Port 4X IB QDR Adapter 40GbI/O drawerAIX, Linux
CCINDescriptionPlacementOS support
2.7.9 Cryptographic Coprocessor
The Cryptographic Coprocessor cards provide both cryptographic coprocessor and
cryptographic accelerator functions in a single card.
The IBM PCIe Cryptographic Coprocessor adapter highlights the following features:
Integrated Dual processors that operate in parallel for higher reliability
Supports IBM Common Cryptographic Architecture or PKCS#11 standard
Chapter 2. Architecture and technical overview 75
Page 90
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
Ability to configure adapter as coprocessor or accelerator
Support for smart card applications using Europay, MasterCard and Visa
Cryptographic key generation and random number generation
PIN processing: generation, verification, translation
Encrypt and Decrypt using AES and DES keys
See the following site for the most recent firmware and software updates:
http://www.ibm.com/security/cryptocards/
Table 2-29 lists the cryptographic adapter that is available for the server.
Table 2-29 Available cryptographic adapters
Feature
Code
CCINDescriptionPlacementOS support
EJ28PCIe Crypto Coprocessor Gen3 BSC
2.8 Internal Storage
The system nodes for Power E870 and Power E880 do not allow for physical storage. All
storage must be provided externally via I/O expansion drawers or SAN. At the time of writing,
the only external I/O expansion drawer that can be used in Power E870 and Power E880 is
EXP24S, attached via SAS ports to a SAS PCIe adapter installed either in a system node
drawer or in a PCIe expansion slot located in an I/O expansion drawer.
The control unit has a DVD drive which is connected to an external USB port on the rear of
the unit. In order to use the DVD drive, at least on PCIe USB adapter must be installed and
connected via an USB cable to the DVD drive.
2.8.1 DVD
There is one DVD media bay per system, located on the front of the control unit. This media
bay allows for a DVD (#EU13) to be installed on the control unit and it enables the USB port
on the rear of the control unit.
The USB port must be connect via an USB cable to a USB PCIe adapter, installed in one of
the available PCIe slots on the system nodes or I/O expansion drawers. A diagram of the DVD
connection can be seen on Figure 2-23 on page 77.
4765-001
I/O drawerAIX, IBM i,
Linux
The basic rules for DVD drives on these system are:
Include DVD drive with #EC13
Include the PCI USB adapter #EC45
Include a USB cable male-male with the proper length. As a suggestion, #EBK4 is a 1.6m
cable that allows enough length for the adapters on the first two system nodes to be
connected to the USB DVD port.
This architecture allows for a more flexible infrastructure where the DVD drive is completely
independent of another components and can be freely moved between partition.
76IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 91
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
Control Unit – Front View
DVD Drive (#EU13)
Control Unit – Rear View
USB Port
PCIe USB
Adapter (#EC45)
USB Cable 1.6m
Male-Male (#EBK4)
System Node
PCIe Expansion
Drawer
It is also important to note that several daily functions where a DVD drive was needed can
now be executed by using other methods like Virtual Media Repository on the Virtual I/O
Server or the Remote Virtual I/O Server install on HMC.
2.9 External I/O subsystems
2.9.1 PCIe Gen3 I/O expansion drawer
Figure 2-23 DVD drive physical location on control unit and suggested cabling
This section describes the PCIe Gen3 I/O expansion drawer that can be attached to the
Power E870 and Power E880.
The PCIe Gen3 I/O expansion drawer is a 4U high, PCI Gen3-based and rack mountable I/O
drawer. It offers two PCIe FanOut Modules (#EMXF) that each provide six PCIe slots.
The physical dimensions of the drawer are 444.5 mm (17.5 in.) wide by 177.8 mm (7.0 in.)
high by 736.6 mm (29.0 in.) deep for use in a 19-inch rack.
A PCIe x16 to Optical CXP converter adapter (#EJ07) and 2.0 M (#ECC6) or 10.0 M (#ECC8)
CXP 16X Active Optical cables (AOC) connect the system node to a PCIe FanOut module in
the I/O expansion drawer. One feature #ECC6 or one #ECC8 ships two AOC cables.
Chapter 2. Architecture and technical overview 77
Page 92
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
Fan-out Module
6 PCIe G3 Slots
4 x8 & 2 x16
4 x8 & 2 x16
Fan-out Module
6 PCIe G3 Slots
4 x8 & 2 x16
4 x8 & 2 x16
Dual Power Cords
PCIe Optical Interface to CEC
PCIe Optical Interface to CEC
Concurrent repair and add/removal of PCIe adapter cards is done by HMC guided menus or
by operating system support utilities.
A blind swap cassette (BSC) is used to house the full high adapters which go into these slots.
The BSC is the same BSC as used with the previous generation server's
#5802/5803/5877/5873 12X attached I/O drawers.
Figure 2-24 shows the back view of the PCIe Gen3 I/O expansion drawer.
Figure 2-24 Rear view of the PCIe Gen3 I/O expansion drawer
I/O drawers are connected to the adapters in the system node with data transfer cables:
2M Optical Cable Pair for PCIe3 Expansion Drawer (#ECC6)
10M Optical Cable Pair for PCIe3 Expansion Drawer (#ECC8)
Cable lengths: Use the 2-meter cables for intra-rack installations. Use the 10-meter
cables for inter-rack installations.
Two PCIe3 Optical Cable Adapter for PCIe3 Expansion Drawer (#EJ07) are required to
connect the expansion drawer, one for each of the two PCIe3 6-slot fanout modules in the
drawer. The top port of the fan out module must be cabled to the top port of the #EJ07 port.
Likewise, the bottom two ports must be cabled together.
Connect an active optical cable to connector T1 on the PCIe3 optical cable adapter in your
server.
78IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 93
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
Connect the other end of the optical cable to connector T1 on one of the PCIe3 6-slot
fanout modules in your expansion drawer.
Connect another cable to connector T2 on the PCIe3 optical cable adapter in your server.
Connect the other end of the cable to connector T2 on the PCIe3 6-slot fanout module in
your expansion drawer.
Repeat above 4 steps for the other PCIe3 6-slot fanout module in the expansion drawer.
Drawer connections: Each fanout module in a PCIe3 Expansion Drawer cannot be
connected to two different systems. However it can be connected to different system nodes
in a system.
Figure 2-25 shows connector locations for the PCIe Gen3 I/O expansion drawer.
Figure 2-25 Connector locations for the PCIe Gen3 I/O expansion drawer
Chapter 2. Architecture and technical overview 79
Page 94
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
General rules for the PCI Gen3 I/O expansion drawer configuration
The PCIe3 optical cable adapter can be in any of the PCIe adapter slots in the Power E870
and Power E880 system node. However, we recommend that you use the PCIe adapter slot
priority information while selecting slots for installing PCIe3 Optical Cable Adapter (#EJ07).
Table 2-30 shows PCIe adapter slot priorities in the Power E870 and Power E880 system.
80IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 95
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
Left FanOut
Right FanOut
Left FanOut
Right FanOut
Slot 6
Slot 7
Slot 5
Slot 4
Slot 3
Slot 2
Slot 1
System node
I/O Draw er
I/O Draw er
Slot 8
Figure 2-27 shows example of a single system node and two PCI Gen3 I/O expansion
drawers.
Figure 2-27 Example of a single system node and two I/O drawers
Chapter 2. Architecture and technical overview 81
Page 96
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
Left FanO ut
Right FanOut
Left FanOut
Right FanOut
Slot 6
Slot 7
Slot 5
Slot 4
Slot 3
Slot 2
Slot 1
Slot 6
Slot 7
Slot 5
Slot 4
Slot 3
Slot 2
Slot 1
System node
System node
I/O Drawer
I/O Drawer
Slot 8
Slot 8
Figure 2-28 shows an example of two system nodes and two PCI Gen3 I/O expansion
drawers.
Figure 2-28 Example of two system nodes and two I/O drawers
82IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 97
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
Left FanOut
Right FanOut
Left FanOut
Right FanOut
Left FanOut
Right FanOut
Left FanOut
Right FanOut
Slot 6
Slot 7
Slot 5
Slot 4
Slot 3
Slot 2
Slot 1
Slot 6
Slot 7
Slot 5
Slot 4
Slot 3
Slot 2
Slot 1
System node
System node
I/O Drawer
I/O Drawer
I/O Drawer
I/O Drawer
Slot 8
Slot 8
Figure 2-29 shows an example of two system node and four PCI Gen3 I/O expansion
drawers.
Figure 2-29 Example of two system nodes and four I/O drawers
2.9.3 PCIe Gen3 I/O expansion drawer SPCN cabling
There is no system power control network (SPCN) used to control and monitor the status of
power and cooling within the I/O drawer. SPCN capabilities are integrated in the optical
cables.
Chapter 2. Architecture and technical overview 83
Page 98
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
2.10 External disk subsystems
This section describes the following external disk subsystems that can be attached to the
Power E870 and Power E880 system:
EXP24S SFF Gen2-bay drawer for high-density storage (#5887)
IBM System Storage
Note: The EXP30 Ultra SSD Drawer (#EDR1 or #5888), the EXP12S SAS Disk Drawer
(#5886), and the EXP24 SCSI Disk Drawer (#5786) are not supported on the Power E870
and Power E880 system.
2.10.1 EXP24S SFF Gen2-bay Drawer
The EXP24S SFF Gen2-bay Drawer (#5887) is an expansion drawer with twenty-four 2.5-inch
form-factor SAS bays. The EXP24S supports up to 24 hot-swap SFF-2 SAS hard disk drives
(HDDs) or solid-state drives (SSDs). It uses only 2 EIA of space in a 19-inch rack. The
EXP24S includes redundant ac power supplies and uses two power cords.
To maximize configuration flexibility and space utilization, the system node of Power E870
and Power E880 system does not have integrated SAS bays or integrated SAS controllers.
PCIe SAS adapters and the EXP24S can be used to provide direct access storage.
To further reduce possible single points of failure, EXP24S configuration rules consistent with
previous Power Systems are used. IBM i configurations require the drives to be protected
(RAID or mirroring). Protecting the drives is highly recommended, but not required for other
operating systems. All Power operating system environments that are using SAS adapters
with write cache require the cache to be protected by using pairs of adapters.
84IBM Power Systems E870 and E880 Technical Overview and Introduction
Page 99
Draft Document for Review October 14, 2014 10:19 am5137ch02.fm
GROUP 1
GROUP 1GROUP 2
GROUP 1GROUP 2GROUP 3GROUP 4
MODE 1
MODE 2
MODE 4
1 x 24 disks
2 x 12 disks
4 x 6 disks
IBM EXP24S front view
With AIX, Linux, and VIOS, you can order the EXP24S with four sets of six bays, two sets of
12 bays, or one set of 24 bays (mode 4, 2, or 1). With IBM i, you can order the EXP24S as
one set of 24 bays (mode 1). Figure 2-30 on page 85 shows the front of the unit and the
groups of disks on each mode.
Figure 2-30 EXP24S front view with location codes and disk groups depending on its mode of
operation
Mode setting is done by IBM Manufacturing and there is no option provided to change the
mode after it is shipped. The stickers indicate whether the enclosure is set to mode 1, mode
2, or mode 4. They are attached to the lower-left shelf of the chassis (A) and the center
support between the enclosure services manager modules (B).
Chapter 2. Architecture and technical overview 85
Page 100
5137ch02.fmDraft Document for Review October 14, 2014 10:19 am
Figure 2-31 on page 86 shows the mode stickers.
Figure 2-31 Mode sticker locations at the rear of the 5887 disk drive enclosure
86IBM Power Systems E870 and E880 Technical Overview and Introduction
Loading...
+ hidden pages
You need points to download manuals.
1 point = 1 manual.
You can buy points or you can get point for every manual you upload.