4405TOC.fmDraft Document for Review March 26, 2009 5:10 pm
viIBM Power 570 Technical Overview and Introduction
Draft Document for Review March 26, 2009 5:10 pm4405spec.fm
Notices
This information was developed for products and services offered in the U.S.A.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area. Any
reference to an IBM product, program, or service is not intended to state or imply that only that IBM product,
program, or service may be used. Any functionally equivalent product, program, or service that does not
infringe any IBM intellectual property right may be used instead. However, it is the user's responsibility to
evaluate and verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not give you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, Armonk, NY 10504-1785 U.S.A.
The following paragraph does not apply to the United Kingdom or any other country where such
provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION
PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR
IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT,
MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of
express or implied warranties in certain transactions, therefore, this statement may not apply to you.
This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.
Any references in this information to non-IBM Web sites are provided for convenience only and do not in any
manner serve as an endorsement of those Web sites. The materials at those Web sites are not part of the
materials for this IBM product and use of those Web sites is at your own risk.
IBM may use or distribute any of the information you supply in any way it believes appropriate without incurring
any obligation to you.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.
This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to the names and addresses used by an actual business
enterprise is entirely coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the sample
programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore,
cannot guarantee or imply reliability, serviceability, or function of these programs.
HACMP™
IntelliStation®
IBM Systems Director Active Energy
Manager™
IBM®
Micro-Partitioning™
OpenPower®
Power Architecture®
PowerPC®
PowerVM™
Predictive Failure Analysis®
POWER™
POWER Hypervisor™
POWER4™
POWER5™
POWER5+™
POWER6™
Redbooks®
RS/6000®
System i™
System i5™
System p™
System p5™
System x™
System z™
System Storage™
Tivoli®
TotalStorage®
Workload Partitions Manager™
1350™
The following terms are trademarks of other companies:
ABAP, SAP NetWeaver, SAP, and SAP logos are trademarks or registered trademarks of SAP AG in Germany
and in several other countries.
Oracle, JD Edwards, PeopleSoft, Siebel, and TopLink are registered trademarks of Oracle Corporation and/or
its affiliates.
InfiniBand, and the InfiniBand design marks are trademarks and/or service marks of the InfiniBand Trade
Association.
Flex, and Portable Document Format (PDF) are either registered trademarks or trademarks of Adobe Systems
Incorporated in the United States, other countries, or both.
Java, and all Java-based trademarks are trademarks of Sun Microsystems, Inc. in the United States, other
countries, or both.
Internet Explorer, Microsoft, and the Windows logo are trademarks of Microsoft Corporation in the United
States, other countries, or both.
Intel, Intel logo, Intel Inside logo, and Intel Centrino logo are trademarks or registered trademarks of Intel
Corporation or its subsidiaries in the United States, other countries, or both.
UNIX is a registered trademark of The Open Group in the United States and other countries.
Linux is a trademark of Linus Torvalds in the United States, other countries, or both.
Other company, product, or service names may be trademarks or service marks of others.
viiiIBM Power 570 Technical Overview and Introduction
Draft Document for Review March 26, 2009 5:10 pm4405pref.fm
Preface
This IBM® Redpaper is a comprehensive guide covering the IBM Power™ 570 server
supporting AIX, IBM i, and Linux for Power operating systems. The goal of this paper is to
introduce the major innovative Power 570 offerings and their prominent functions, including
the following.
Unique modular server packaging
New POWER6™ processors available at frequencies of 4.2, 4.4, and 5.0 GHz.
The POWER6 processor available at frequencies of 3.5, 4.2, and 4.7 GHz.
The specialized POWER6 DDR2 memory that provides greater bandwidth, capacity, and
reliability.
The 1 Gb or 10 Gb Integrated Virtual Ethernet adapter that brings native hardware
virtualization to this server
PowerVM™ virtualization including PowerVM Live Partition Mobility
Redundant service processors to achieve continuous availability
Professionals wishing to acquire a better understanding of IBM System p products should
read this Redpaper. The intended audience includes:
Clients
Sales and marketing professionals
Technical support professionals
IBM Business Partners
Independent software vendors
This Redpaper expands the current set of IBM Power Systems documentation by providing a
desktop reference that offers a detailed technical description of the 570 system.
This Redpaper does not replace the latest marketing materials and tools. It is intended as an
additional source of information that, together with existing materials, may be used to
enhance your knowledge of IBM server solutions.
The team that wrote this paper
This paper was produced by a team of specialists from around the world working at the
International Technical Support Organization, Austin Center.
Giuliano Anselmi works with passion for IBM devoted to RS/6000® and pSeries® systems
since 15 years, having a very deep knowledge of the related hardware, and solutions. He
used to be a pSeries Systems Product Engineer for 7 years, supporting Web Server Sales
Organization, IBM Sales, Business Partners, Technical Support Organizations. In 2004, he
joined Field Technical Sales Support group and has been accredited as IT specialist in 2007.
Currently he plays the role of system architect in IBM STG and supporting the General
Business division.
YoungHoon Cho is a System p Product Engineer at the pSeries post-sales Technical
Support Team in IBM Korea. He has seven years of experience working on RS/6000 and
4405pref.fmDraft Document for Review March 26, 2009 5:10 pm
System p products. He is an IBM Certified Specialist in System p and AIX® 5L™. He provides
second line support to field engineers with technical support on System p, and system
management.
Gregor Linzmeier is an IBM Advisory IT Specialist for IBM System p workstation and entry
servers as part of the Systems and Technology Group in Mainz, Germany supporting IBM
sales, Business Partners, and clients with pre-sales consultation and implementation of
client/server environments. He has worked for more than 15 years as an infrastructure
specialist for RT, RS/6000, IBM IntelliStation® POWER™, and AIX in large CATIA
client/server projects. Actual engagements are AIX Thin Server, Partition migration and
Green IT.
Marcos Quezada is a Senior Accredited IT Specialist in Argentina. He has 10 years of IT
experience as a UNIX systems' pre-sales specialist and as a Web Project Manager. He holds
a degree on Informatics Engineering from Fundación Universidad de Belgrano. His areas of
expertise include IBM RS/6000, IBM eServer™ pSeries/p5 and Power Systems servers
under the AIX operating system and pre-sales support of IBM Software, SAP® and Oracle®
solutions architecture running on IBM UNIX Systems with focus on competitive accounts..
John T Schmidt is an Accredited IT Specialist for IBM and has over 7 years experience with
IBM and System p. He has a degree in Electrical Engineering from the University of Missouri
- Rolla and an MBA from Washington University in St. Louis. He is currently working in the
United States as a presales Field Technical Sales Specialist for System p in St. Louis, MO.
Guido Somers is a Cross Systems Certified IT Specialist working for IBM Belgium. He has
13 years of experience in the Information Technology field, ten years of which were within
IBM. He holds degrees in Biotechnology, Business Administration, Chemistry, and
Electronics, and did research in the field of Theoretical Physics. His areas of expertise include
AIX, Linux®, system performance and tuning, logical partitioning, virtualization, HACMP™,
SAN, IBM System p servers, as well as other IBM hardware offerings. He currently works as a
Client IT Architect for Infrastructure and Global ISV Solutions in the e-Business Solutions
Technical Support (eTS) organization. He is also the author of the second edition of
Integrated Virtualization Manager on IBM System p5™, REDP-4061, and the PowerVM Live
Partition Mobility on IBM System p, SG24-740.
The project that produced this publication was managed by:
Scott Vetter, PMP
Thanks to the following people for their contributions to this project:
George Ahrens, Ron Arroyo, Brad Behle, Nick Bofferding, Martha Broyles, Pat Buckland,
Curtis Eide, Chris Eisenmann, Michael S. Floyd, Chris Francois, Andrew J. Geissler,
Gordon Grout, Volker Haug, Daniel J. Henderson, Tenley Jackson, Robert G. Kovacs,
Hye-Young McCreary, Bill Mihaltse, Jim A. Mitchell, Thoi Nguyen, Amartey Pearson,
Cale Rath, Todd Rosedahl, Terry Schardt, Julissa Villarreal, Brian Warner, Christine I. Wang.
IBM US
Bruno Digiovani
IBM Argentina
Become a published author
Join us for a two- to six-week residency program! Help write a book dealing with specific
products or solutions, while getting hands-on experience with leading-edge technologies. You
xIBM Power 570 Technical Overview and Introduction
Draft Document for Review March 26, 2009 5:10 pm4405pref.fm
will have the opportunity to team with IBM technical professionals, Business Partners, and
Clients.
Your efforts will help increase product acceptance and customer satisfaction. As a bonus, you
will develop a network of contacts in IBM development labs, and increase your productivity
and marketability.
Find out more about the residency program, browse the residency index, and apply online at:
ibm.com/redbooks/residencies.html
Comments welcome
Your comments are important to us!
We want our papers to be as helpful as possible. Send us your comments about this paper or
other IBM Redbooks® in one of the following ways:
Use the online Contact us review Redbooks form found at:
ibm.com/redbooks
Send your comments in an e-mail to:
redbooks@us.ibm.com
Mail your comments to:
IBM Corporation, International Technical Support Organization
Dept. HYTD Mail Station P099
2455 South Road
Poughkeepsie, NY 12601-5400
Preface xi
4405pref.fmDraft Document for Review March 26, 2009 5:10 pm
xiiIBM Power 570 Technical Overview and Introduction
Draft Document for Review May 22, 2009 3:10 pm4405ch01 General description.fm
1
Chapter 1.General description
The innovative IBM Power 570 mid-range server with POWER6 and available POWER6+ f
processor cards delivers outstanding price/performance, mainframe-inspired reliability and
availability features, flexible capacity upgrades, and innovative virtualization technologies to
enable management of growth, complexity, and risk.
The Power 570 leverages your exisiting investments by supporting AIX, IBM i, and Linux for
Power, and x86 Linux applications on a single server. It is available in 2-, 4-, 8-, 12-, and
16-core and 32-core configurations. As with the p5 570, the POWER6-based 570s modular
symmetric multiprocessor (SMP) architecture is constructed using 4U (EIA units), 4-core or
8-core building block modules (also referred to as nodes, or CECs). Each of these nodes
supports four POWER6 3.5, 4.2 or 4.7 GHz dual-core processors, and new POWER6 4.2
GHz dual-core processors, or POWER6+ 4.4, and 5.0 GHz four-core processors along with
cache, memory, media, disks, I/O adapters, and power and cooling to create a balanced,
extremely high-performance rack-mount system.
This design allows up to four modules to be configured in a 19-inch rack as a single SMP
server, allowing clients to start with what they need and grow by adding additional building
blocks. A fully configured 570 server may consist of 32 processor cores, 768 GB of DDR2
memory, four media bays, integrated ports for attaching communications devices, 24 mixed
PCI-X and PCI Express adapter slots, and 24 internal SAS (Serial Attached SCSI) drives
accommodating up to 7.2 TB of internal disk storage.
The 64-bit POWER6 processors in this server are integrated into a dual-core single chip
module and a dual-core dual chip module, with 32 MB of L3 cache, 8 MB of L2 cache, and 12
DDR2 memory DIMM slots. The unique DDR2 memory uses a new memory architecture to
provide greater bandwidth and capacity. This enables operating at a higher data rate for large
memory configurations. Each new processor card can support up to 12 DDR2 DIMMs
running at speeds of up to 667 MHz.
As with the POWER5™ processor, simultaneous miltithreading enabling two threads to be
executed at the same time on a single processor core is a standard feature of POWER6
technology. Introduced with the POWER6 processor design is hardware decimal
floating-point support improving the performance of the basic mathematical calculations of
financial transactions that occur regularly on today’s business computers. The POWER6
4405ch01 General description.fmDraft Document for Review May 22, 2009 3:10 pm
processor also includes an AltiVec SIMD accelerator, which helps to improve the performance
of high performance computing (HPC) workloads.
All Power Systems servers can utilize logical partitioning (LPAR) technology implemented
using System p virtualization technologies, the operating system (OS), and a hardware
management console (HMC). Dynamic LPAR allows clients to dynamically allocate many
system resources to application partitions without rebooting, allowing up to 16 dedicated
processor partitions on a fully configured system.
In addition to the base virtualization that is standard on every System p server, two optional
virtualization features are available on the server: PowerVM Standard Edition (formerly
Advanced POWER Virtualization (APV) Standard) and PowerVM Enterprise Edition (formerly
APV Enterprise).
PowerVM Standard Edition includes IBM Micro-Partitioning™ and Virtual I/O Server (VIOS)
capabilities. Micro-partitions can be defined as small as 1/10th of a processor and be
changed in increments as small as 1/100th of a processor. Up to 160 micro-partitions may be
created on a 16-core 570 system. VIOS allows for the sharing of disk and optical devices and
communications and Fibre Channel adapters. Also included is support for Multiple Shared
Processor Pools and Shared Dedicated Capacity.
PowerVM Enterprise Edition includes all features of PowerVM Standard Edition plus Live
Partition Mobility, newly available with POWER6 systems. It is designed to allow a partition to
be relocated from one server to another while end users are using applications running in the
partition.
Other features introduced with POWER6 processor-based technology include an Integrated
Virtual Ethernet adapter standard with every system, the Processor Instruction Retry feature
automatically monitoring the POWER6 processor and, if needed, restarting the processor
workload without disruption to the application, and a new HMC (Hardware Management
Console) graphical user interface offering enhanced systems control.
2IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 22, 2009 3:10 pm4405ch01 General description.fm
1.1 System specifications
Table 1-1 lists the general system specifications of a single Central Electronics Complex
(CEC) enclosure.
Table 1-1 System specifications
DescriptionRange (operating)
Operating temperature5 to 35 degrees C (41 to 95 F)
Relative humidity8% to 80%
Maximum wet bulb23 degrees C (73 F)
Noise level with 3.5 GHz processors FC 5620: 7.1 bels
with 3.5 GHz processors FC 5620 and acoustic rack doors:
6.7 bels
with 4.2 GHz processors FC 5622: 7.1 bels
with 4.2 GHz processors FC 5622 and acoustic rack doors:
6.7 bels
with 4.7 GHz processors FC 7380: 7.4 bels
with 4.7 GHz processors FC 7380 and acoustic rack doors:
6.9 bels
Operating voltage200 to 240 V ac 50/60 Hz
Maximum power consumption1400 watts (maximum)
Maximum power source loading1.428 kVA (maximum)
Maximum thermal output4778 BTU
Maximum altitude3,048 m (10,000 ft)
a. British Termal Unit (BTU)
1.2 Physical package
The system is available only in a rack-mounted form factor. It is a modular-built system
utilizing between one and four building block enclosures. Each of these CEC drawer building
blocks is packaged in a 4U
building block are shown in Table 1-2.
Table 1-2 Physical packaging of CEC drawer
DimensionOne CEC drawer
Height174 mm (6.85 in)
Width483 mm (19.0 in.)
Depth 824 mm (32.4 in.) from front of Bezel to rear of Power Supply
674 mm (25.6 in.) front rack rail mounting surface to I/O Adapter Bulkhead
793 mm (31.2 in.) front rack rail mounting surface to rear of Power Supply
a
/hr (maximum)
1
rack-mounted enclosure. The major physical attributes for each
Weight63.6 kg (140 lb.)
To help ensure the installation and serviceability in non-IBM, industry-standard racks, review
the vendor’s installation planning information for any product-specific installation
requirements.
1
One Electronic Industries Association Unit (1U) is 44.45 mm (1.75 in.).
Chapter 1. General description 3
4405ch01 General description.fmDraft Document for Review May 22, 2009 3:10 pm
Figure 1-1 shows system views.
Figure 1-1 Views of the system
1.3 System features
The full system configuration is made of four CEC building blocks. It features:
2-, 4-, 8-, 12-, 16-, and 32-core configurations utilizing the POWER6 chip on up to eight
Up to 192 GB DDR2 memory per enclosure, 768 GB DDR2 max per system. Available
memory features are 667 MHz, 533 MHz, or 400 MHz depending on memory density.
Up to 6 SAS DASD disk drives per enclosure, 24 max per system.
4IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 22, 2009 3:10 pm4405ch01 General description.fm
6 PCI slots per enclosure: 4 PCIe, 2 PCI-X; 24 PCI per system: 16 PCIe, 8 PCI-X.
Up to 2 GX+ adapters per enclosure; 8 per system
One hot-plug slim-line media bay per enclosure, 4 max per system.
The external processor fabric bus in this system is modular. For a multiple-drawer server
configuration, a processor fabric cable or cables, and a service interface cable are required.
Cable features are available for connecting pairs of drawers, three drawer stacks, and four
drawer stacks. With this modular approach, a separate cable is required to connect each
drawer to each other drawer in a multi-enclosure stack (See 2.2.1 and 2.4.2).
The service processor (SP), which is described in 2.14.1, “Service processor” on page 70.
Each system includes the following native ports:
Choice of integrated (IVE) I/O options -- one per enclosure.
– 2-port 1 Gigabit Integrated Virtual Ethernet controller with two system ports
(10/100/1000 twisted pair).
– 4-port 1 Gigabit Integrated Virtual Ethernet controller with one system port
(10/100/1000 twisted pair).
– 2-port 10 Gigabit Integrated Virtual Ethernet controller (SR optical) with one system
port.
Two USB ports per enclosure.
Two system (serial) ports per enclosure. Only the ports in the base enclosure are active,
and only when an HMC is not attached.
Two HMC ports per enclosure. The HMC must be attached to CEC enclosure 1 (and CEC
enclosure 2 to support redundant Service Processors).
Two SPCN ports per enclosure.
In addition, each building block features one internal SAS controller, redundant
hot-swappable cooling fans, redundant power supplies, and redundant processor voltage
regulators.
1.3.1 Processor card features
Each of the four system enclosures has two processor sockets and can contain two
POWER6/POWER6+ dual-core 64-bit processor card features, or two POWER6 dual-core
dual-chip processor card features. They are configured as dual cores on a single chip module
or dual chip module with 32 MB of L3 cache, 8 MB of L2 cache, and 12 DDR2 memory DIMM
slots.
The POWER6 processor is available at frequencies of 3.5, 4.2, or 4.7 GHz. The POWER6+
processor is available at frequencies of 4.2, 4.4, and 5.0 GHz.Each system must have a
minimum of two active processors. A system with one enclosure may have one or two
processor cards installed. A system with two, three, or four enclosures must have two
processor cards in each enclosure. When two or more processor cards are installed in a
system, all cards must have the same feature number.
All processor card features are available only as Capacity on Demand (CoD). The initial order
of the system must contain the feature code (FC) related to the desired processor card, and it
must contain the processor activation feature code. The types of CoD supported are:
Capacity Upgrade on Demand (CUoD) allows you to purchase additional permanent
processor or memory capacity and dynamically activate them when needed.
Chapter 1. General description 5
4405ch01 General description.fmDraft Document for Review May 22, 2009 3:10 pm
Utility CoD autonomically provides additional processor performance on a temporary
basis within the shared processor pool in one minute increments. It adds additional cores
to allow greater parellel operation, and can increase the effective L2 cache of the shared
processor pool
On/Off CoD enables processors or memory to be temporarily activated in full-day
increments as needed.
Trial CoD (exception) offers a one-time, no-additional-charge 30-day trial that allows you to
explore the uses of all inactive processor capacity on your server.
Trial CoD (standard) offers a one-time 2-core activation for 30 days.
Capacity Backup (IBM i only) offers a 1 license entitlement to a backup system on a
temparary basis.
Table 1-3 contains the feature codes for processor cards at the time of writing.
Table 1-3 Processor card and CoD feature codes
Processor card FCDescription
5620
5670
5640
5650
5622
5672
5641
5653
7380
5403
5404
5656
7951On/Off Processor Enablement. This feature can be ordered to
1.3.2 Memory features
Processor card feature codes 7380, 5620, and 5622 have 12 memory DIMM slots and must
be populated with POWER6 DDR2 Memory DIMMs. Each processor card feature must have
a minimum of four DIMMs installed. This includes inactive processor card features present in
the system. Table 1-4 shows the memory feature codes that are available at the time of
writing.
One Processor Activation for Processor FC 5622
Utility Billing for FC 5622-100 processor minutes
On/Off Processor Day Billing for FC 5621 or FC 5622
One Processor Activation for Processor FC 7380
Utility Billing for FC 7380-100 processor minutes
On/Off Processor Day Billing for FC 7380
enable your server for On/Off Capacity on Demand. Once
enabled, you can request processors on a temporary basis. You
must sign an On/Off Capacity on Demand contract before you
order this feature
All memory card features are available only as Capacity on Demand and support the same
CoD options described for processors (with the exception of Utility CoD).
Memory Trial CoD (exception) offers a one-time, no-additional-charge 30-day trial that
allows you to explore the uses of all memory capacity on your server.
Memory Trial CoD (standard) offers a one-time 4 GB activation for 30 days.
All POWER6 memory features must be purchased with sufficient permanent memory
activation features so that each memory feature is at least 50% active, except memory feature
code 8129 which must be purchased with Activation feature code 5681 for 100% activation.
6IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 22, 2009 3:10 pm4405ch01 General description.fm
Memory feature codes 5692, 5693, 5694, and 5695 can be mixed on the same POWER6
processor card. Memory feature codes 5696 and 8129 may not be mixed with any other
memory feature on a single processor card. A processor card with memory feature 5696 or
8129 can be mixed in the same CEC enclosure with a processor card containing other
POWER6 memory features. For all processors and all system configurations, if memory
features in a single system have different frequencies, all memory in the system will function
according to the lowest frequency present. Memory features 5696 and 8129 cannot be used
on processor card feature code 5620.
For all processors and all system configurations, if memory features in a single system have
different frequencies, all memory in the system will function according to the lowest frequency
present
1.3.3 Disk and media features
Each system building block features one SAS DASD controller with six hot-swappable
3.5-inch SAS disk bays and one hot-plug, slim-line media bay per enclosure. Only the new
SAS DASD hard disks are supported internally. The older SCSI DASD hard files can be
attached, but must be located in a remote I/O drawer. In a full configuration with four
connected building blocks, the combined system supports up to 24 disk bays
Table 1-5 shows the disk drive feature codes that each bay can contain.
Table 1-5 Disk drive feature code description
Feature codeDescription
364673 GB 15 K RPM SAS Disk Drive
3647146 GB 15 K RPM SAS Disk Drive
3648300 GB 15 K RPM SAS Disk Drive
In a full configuration with four connected, the combined system supports up to four media
devices with Media Enclosure and Backplane feature 5629.
Any combination of the following DVD-ROM and DVD-RAM drives can be installed:
FC 5756 IDE Slimline DVD-ROM Drive
Chapter 1. General description 7
4405ch01 General description.fmDraft Document for Review May 22, 2009 3:10 pm
FC 5757 IBM 4.7 GB IDE Slimline DVD-RAM Drive
1.3.4 I/O drawers
The system has seven I/O expansion slots per enclosure, including one dedicated GX+ slot.
The other 6 slots support PCI adapters. There are 3 PCIe 8X long slots and 1 PCIe 8X short
slot. The short PCIe slot may also be used for a second GX+ adapter. The remaining 2 slots
PCI-X long slots. If more PCI slots are needed, such as to extend the number of LPARs, up to
20 I/O drawers on a RIO-2 interface (7311-D11 or 7311-D20) ,and up to 32 I/O drawers on a
12X Channel interface (7314-G30) can be attached.
The adapters that are used in the GX expansion slots are concurrently maintainable on
systems with firmware level FM320_xxx_xxx, or later. If the GX adapter were to fail, the card
could be replaced with a working card without powering down the system.
7311 Model D11 I/O drawer
The 7311 Model D11 I/O drawer features six long PCI-X slots. Blind-swap cassettes are (FC
7862) are utilized. Two 7311 Model D11 I/O drawers fit side-by-side in the 4U enclosure (FC
7311) mounted in a 19-inch rack, such as the IBM 7014-T00 or 7014-T42.
The 7311 Model D11 I/O drawer offers a modular growth path for systems with increasing I/O
requirements. A fully configured system supports 20 attached 7311 Model D11 I/O drawers.
The combined system supports up to 128 PCI-X adapters and 16 PCIe adapters. In a full
configuration, Remote I/O expansion cards (FC 1800 - GX Dual Port RIO-2) are required.
The I/O drawer has the following attributes:
4U rack-mount enclosure (FC 7311) that can hold one or two D11 drawers
Default redundant hot-plug power and cooling devices
Two RIO-2 and two SPCN ports
7311 Model D11 I/O drawer physical package
Because the 7311 Model D11 I/O drawer must be mounted into the rack enclosure (FC 7311),
these are the physical characteristics of one I/O drawer or two I/O drawers side-by-side:
One 7311 Model D11 I/O drawer
– Width: 223 mm (8.8 in.)
– Depth: 711 mm (28.0 in.)
– Height: 175 mm (6.9 in.)
– Weight: 19.6 kg (43 lb.)
Two I/O drawers in a 7311 rack-mounted enclosure have the following characteristics:
– Width: 445 mm (17.5 in.)
– Depth: 711 mm (28.0 in.)
– Height: 175 mm (6.9 in.)
– Weight: 39.1 kg (86 lb.)
7311 Model D20 I/O drawer
The 7311 Model D20 I/O drawer is a 4U full-size drawer, which must be mounted in a rack. It
features seven hot-pluggable PCI-X slots and, optionally up to 12 hot-swappable disks
arranged in two 6-packs. Redundant concurrently maintainable power and cooling is an
8IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 22, 2009 3:10 pm4405ch01 General description.fm
optional feature (FC 6268). The 7311 Model D20 I/O drawer offers a modular growth path for
systems with increasing I/O requirements. When fully configured with 20 attached
7311 Model D20 drawers, the combined system supports up to 148 PCI-X adapters, 16 PCIe
adapters, and 264 hot-swappable disks. In a full configuration, Remote I/O expansion cards
(FC 1800 - GX Dual Port RIO-2) are required.
PCI-X and PCI cards are inserted into the slots from the top of the I/O drawer. The installed
adapters are protected by plastic separators, which are designed to prevent grounding and
damage when adding or removing adapters.
Note: The 7311 Model D20 I/O drawer initial order, or an existing 7311 Model D20 I/O
drawer that is migrated from another pSeries system, must have the RIO-2 ports available
(FC 6417).
7311 Model D20 I/O drawer physical package
The I/O drawer has the following physical characteristics:
Width: 482 mm (19.0 in.)
Depth: 610 mm (24.0 in.)
Height: 178 mm (7.0 in.)
Weight: 45.9 kg (101 lb.)
Figure 1-2 shows the different views of the 7311-D20 I/O drawer.
Chapter 1. General description 9
4405ch01 General description.fmDraft Document for Review May 22, 2009 3:10 pm
Adapters
Service
Access
I/O
Drawer
FrontRear
Operator panel
Power supply 2
Power supply 1
RIO ports
1 2 3 45 6 7
8 9 A B C D 8 9 A B C D
SCSI disk locations and IDs
Reserved ports
SPCN ports
Rack indicator
PCI-X slots
Figure 1-2 7311-D20 I/O drawer views
Note: The 7311 Model D10, and the 7311 Model D11, or the 7311 Model D20 I/O drawers
are designed to be installed by an IBM service representative.
7314 Model G30 PCI-X I/O drawer
The 7314 Model G30 I/O Drawer is a rack-mountable expansion cabinet that can be attached
to selected IBM System p host servers with IBM POWER6 technology. It is a half-rack width
drawer that allows up to two G30 drawers to fit side-by-side in enclosure FC 7314 in the same
4 EIA units of vertical space in a 19-inch rack. Each Model G30 Drawer gives you six
full-length PCI-X, 64-bit, 3.3V, PCI-X DDR adapter slots that can run at speeds up to 266
MHz.
The 7314 Model G30 I/O drawer offers a modular growth path for selected POWER6
systems. It attaches to the host system using IBMs12X Channel Interface technology. The
Dual-Port 12X Channel Attach Adapters available for the Model G30 allow higher-speed data
transfer rates for remote I/O drawers. A single 12X Channel I/O loop can support up to four
G30 I/O drawers.
When fully configured, the system supports up to 32 Model G30 I/O Drawers attached to GX
adapters (FC 1802 GX Dual Port - 12X Channel Attach) available for the GX+ slots. The
combined system supports up to 200 PCI-X adapters and 12 PCIe adapters.
The I/O drawer has the following attributes:
4 EIA unit rack-mount enclosure (FC 7314) holding one or two G30 drawers.
Six PCI-X DDR slots: 64-bit, 3.3V, 266 MHz. Blind-swap.
Redundant hot-swappable power and cooling units.
10IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 22, 2009 3:10 pm4405ch01 General description.fm
– short run: cables between this adapter and a host system may not exceed 3.0 Meters
in length. Cables between two I/O drawers may not exceed 1.5 Meters if both I/O
drawers include this short run adapter. Cables between two I/O drawers may not
exceed 3.0 Meters if either of the I/O drawers includes this short run adapter.
– long run: this adapter includes the repeater function and can support longer cable
loops allowing drawers to be located in adjacent racks. 12X Cables up to 8 Meters in
length can be attached to this adapter. The required 12X Cables are ordered under a
separate feature number.
Six blind-swap cassettes.
The I/O drawer physical characteristics are shown in Table 1-6.
Table 1-6 7314 G30 I/O Drawer specifications
DimensionOne G30 drawerMounting enclosure
Height172 mm (6.8 in.)176 mm (6.9 in.)
Width224 mm (8.8 in.)473 mm (18.6 in.)
Depth800 mm (31.5 in.)800 mm (31.5 in.)
Weight20 kg (44 lb.)45.9 kg (101 lb.) max with 2 G30 drawers
Note: 12X Channel I/O drawers cannot be mixed in a single I/O loop with RIO-2 drawers. A
host system can support both RIO-2 and 12X Channel data transfer loops as long as the
system supports both technologies and has the capability to support two or more
independent remote I/O loops. See 2.10.6, “7314 Model G30 I/O drawer” on page 54 and
2.10.5, “7311 I/O drawer and SPCN cabling” on page 54 for more information.
I/O drawers and usable PCI slot
The different I/O drawer model types can be intermixed on a single server within the
appropriate I/O loop. Depending on the system configuration, the maximum number of I/O
drawers supported is different. If both 7311 and 7314 drawers are being used, the total
number of I/O drawers allowed will be the values shown for the 7314-G30, assuming enough
GX slots are available to configure the required RIO-2 and 12x channel adapters. For either
attachment technology, up to four I/O drawers are supported in a loop.
Table 1-7 summarizes the maximum number of I/O drawers supported and the total number
of PCI slots available when expansion consists of a single drawer type.
Table 1-7 Maximum number of I/O drawers supported and total number of PCI slots
System
drawers/cores
Max RIO-2
drawers
a
Max
12X Ch
drawers
Total number of slots
a
D11D20G30
PCI-XPCIePCI-XPCIePCI-XPCIe
1 drawer /
2-core
1 drawer /
4-core
2 drawers /
8-core
44264304284
88503
1216767
b
583
b
887
Chapter 1. General description 11
b
b
503
1006
c
c
4405ch01 General description.fmDraft Document for Review May 22, 2009 3:10 pm
System
drawers/cores
3 drawers /
12-core
4 drawers /
16-core
a. Up to four I/O drawers are supported in a loop
b. One PCIe slot is reserved for the Remote I/O expansion card
c. One PCIe slot per CEC drawer is reserved for the 12X channel attach expansion card.
Max RIO-2
drawers
162410211
203212815
a
Max
12X Ch
drawers
Total number of slots
a
D11D20G30
PCI-XPCIePCI-XPCIePCI-XPCIe
1.3.5 Hardware Management Console models
The Hardware Management Console (HMC) is required for this system. It provides a set of
functions that are necessary to manage the system, including Logical Partitioning, Capacity
on Demand, inventory and microcode management, and remote power control functions.
Connection of an HMC disables the two integrated system ports.
Table 1-8 lists the HMC models available for POWER6 based systems at the time of writing.
They are preloaded with the required Licensed Machine Code Version 7 (FC 0962) to support
POWER6 systems, in addition to POWER5 and POWER5+ systems.
b
b
11811
14815
b
b
1509
20012
c
c
Existing HMC models 7310 can be upgraded to Licensed Machine Code Version 7 to support
environments that may include POWER5, POWER5+, and POWER6 processor-based
servers. Version 7 is not available for the 7315 HMCs. Licensed Machine Code Version 6
(FC 0961) is not available for 7042 HMCs, and Licensed Machine Code Version 7 (FC 0962)
is not available on new 7310 HMC orders.
Table 1-8 POWER6 HMC models available
Type-modelDescription
7042-C06IBM 7042 Model C06 desktop Hardware Management Console
7042-CR4IBM 7042 Model CR4 rack-mount Hardware Management Console
Note: POWER5 and POWER5+ processor-based servers must have firmware SF240 or
later installed before being managed by a 7042 HMC or 7310 HMC with FC 0962 installed.
1.4 System racks
The system is designed to be installed in a 7014-T00 or -T42 rack. The 7014 Model T00 and
T42 are 19-inch racks for general use with IBM System p rack-mount servers. An existing T00
or T42 rack can be used if sufficient space and power are available. The system is not
supported in the 7014-S25 or the S11.
Note: The B42 rack is also supported.
FC 0469 Customer Specified Rack Placement provides the client the ability to specify the
physical location of the system modules and attached expansion modules (drawers) in the
12IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 22, 2009 3:10 pm4405ch01 General description.fm
racks. The client’s input is collected and verified through the marketing configurator (eConfig).
The client’s request is reviewed by eConfig for safe handling by checking the weight
distribution within the rack. The manufacturing plant provides the final approval for the
configuration. This information is then used by IBM Manufacturing to assemble the system
components (drawers) in the rack according to the client’s request.
If a system is to be installed in a non-IBM rack or cabinet, it must be ensured that the rack
conforms to the EIA
2
standard EIA-310-D (see 1.4.7, “OEM rack” on page 19).
Note: It is the client’s responsibility to ensure that the installation of the drawer in the
preferred rack or cabinet results in a configuration that is stable, serviceable, safe, and
compatible with the drawer requirements for power, cooling, cable management, weight,
and rail security.
1.4.1 IBM 7014 Model T00 rack
The 1.8-meter (71-in.) Model T00 is compatible with past and present IBM System p systems.
The T00 rack has the following features:
36 EIA units (36 U) of usable space.
Optional removable side panels.
Optional highly perforated front door.
Optional side-to-side mounting hardware for joining multiple racks.
Standard business black or optional white color in OEM format.
Increased power distribution and weight capacity.
protection with modular rear brace, concrete floor bolt-down hardware, and bolt-in steel
front filler panels.
Support for both AC and DC configurations.
The rack height is increased to 1926 mm (75.8 in.) if a power distribution panel is fixed to
the top of the rack.
Up to four power distribution units (PDUs) can be mounted in the PDU bays (see
Figure 1-3 on page 15), but others can fit inside the rack. See 1.4.3, “The AC power
distribution unit and rack content” on page 14.
An optional rack status beacon (FC 4690). This beacon is designed to be placed on top of
a rack and cabled to servers, such as a Power 570 and other components inside the rack.
Servers can be programmed to illuminate the beacon in response to a detected problem
or changes in the system status.
A rack status beacon junction box (FC 4693) should be used to connect multiple servers to
the beacon. This feature provides six input connectors and one output connector for the
rack. To connect the servers or other components to the junction box or the junction box to
the rack, status beacon cables (FC 4691) are necessary. Multiple junction boxes can be
linked together in a series using daisy chain cables (FC 4692).
Weights:
– T00 base empty rack: 244 kg (535 lb.)
– T00 full rack: 816 kg (1795 lb.)
2
Electronic Industries Alliance (EIA). Accredited by American National Standards Institute (ANSI), EIA provides a
forum for industry to develop standards and publications throughout the electronics and high-tech industries.
Chapter 1. General description 13
4405ch01 General description.fmDraft Document for Review May 22, 2009 3:10 pm
1.4.2 IBM 7014 Model T42 rack
The 2.0-meter (79.3-inch) Model T42 addresses the client requirement for a tall enclosure to
house the maximum amount of equipment in the smallest possible floor space. The features
that differ in the Model T42 rack from the Model T00 include:
42 EIA units (42 U) of usable space (6 U of additional space).
The Model T42 supports AC only.
Weights:
– T42 base empty rack: 261 kg (575 lb.)
– T42 full rack: 930 kg (2045 lb.)
Optional Rear Door Heat eXchanger (FC 6858)
Improved cooling from the Rear Door Heat eXchanger enables clients to more densely
populate individual racks, freeing valuable floor space without the need to purchase additional
air conditioning units. The Rear Door Heat eXchanger features:
Water-cooled heat exchanger door designed to dissipate heat generated from the back of
computer systems before it enters the room.
An easy-to-mount rear door design that attaches to client-supplied water, using industry
standard fittings and couplings.
Up to 15 KW (approximately 50,000 BTUs/hr) of heat removed from air exiting the back of
a fully populated rack.
One year limited warranty.
Physical specifications
The following are the general physical specifications
Approximate height1945.5 mm (76.6 in.)
Approximate width635.8 mm (25.03 in.)
Approximate depth back door only1042.0 mm (41.0 in.)
Approximate depth back door and front1098.0 mm (43.3 in.)
Approximate depth sculptured style front door 1147.0 mm (45.2 in.)
Approximate weight 31.9 kg (70.0 lb.)
Client responsibilities
Clients must ensure the following:
Secondary water loop (to building chilled water)
Pump solution (for secondary loop)
Delivery solution (hoses and piping)
Connections: Standard 3/4-inch internal threads
1.4.3 The AC power distribution unit and rack content
For rack models T00 and T42, 12-outlet PDUs are available. These include PDUs Universal
UTG0247 Connector (FC 9188 and FC 7188) and Intelligent PDU+ Universal UTG0247
Connector (FC 5889 and FC 7109).
Four PDUs can be mounted vertically in the back of the T00 and T42 racks. See Figure 1-3
for the placement of the four vertically mounted PDUs. In the rear of the rack, two additional
14IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 22, 2009 3:10 pm4405ch01 General description.fm
PDUs can be installed horizontally in the T00 rack and three in the T42 rack. The four vertical
mounting locations will be filled first in the T00 and T42 racks. Mounting PDUs horizontally
consumes 1 U per PDU and reduces the space available for other racked components. When
mounting PDUs horizontally, we recommend that you use fillers in the EIA units occupied by
these PDUs to facilitate proper air-flow and ventilation in the rack.
Figure 1-3 PDU placement and PDU view
For detailed power cord requirements and power cord feature codes, see IBM System p5,
eServer p5 and i5, and OpenPower Planning, SA38-0508. For an online copy, see the IBM
Systems Hardware Information Center. You can find it at:
http://publib.boulder.ibm.com/eserver/
Note: Ensure that the appropriate power cord feature is configured to support the power
being supplied.
The Base/Side Mount Universal PDU (FC 9188) and the optional, additional, Universal PDU
(FC 7188) and the Intelligent PDU+ options (FC 5889 and FC 7109) support a wide range of
country requirements and electrical power specifications. The PDU receives power through a
UTG0247 power line connector. Each PDU requires one PDU-to-wall power cord. Various
power cord features are available for different countries and applications by varying the
PDU-to-wall power cord, which must be ordered separately. Each power cord provides the
unique design characteristics for the specific power requirements. To match new power
requirements and save previous investments, these power cords can be requested with an
initial order of the rack or with a later upgrade of the rack features.
The PDU has 12 client-usable IEC 320-C13 outlets. There are six groups of two outlets fed by
six circuit breakers. Each outlet is rated up to 10 amps, but each group of two outlets is fed
from one 15 amp circuit breaker.
Note: Based on the power cord that is used, the PDU can supply from 4.8 kVA to 19.2 kVA.
The total kilovolt ampere (kVA) of all the drawers plugged into the PDU must not exceed
the power cord limitation.
Chapter 1. General description 15
4405ch01 General description.fmDraft Document for Review May 22, 2009 3:10 pm
The Universal PDUs are compatible with previous models.
Note: Each system drawer to be mounted in the rack requires two power cords, which are
not included in the base order. For maximum availability it is highly recommended to
connect power cords from the same system to two separate PDUs in the rack. And to
connect each PDU to independent power sources.
1.4.4 Intelligent Power Distribution Unit (iPDU)
Energy consumption is becoming a large issue in computer-based businesses. The energy
required to power and cool computers can be a significant cost to a business – reducing profit
margins and consuming resources.
For all systems without an internal thermal and power consumption method the IBM
Intelligent Power Distribution Management (IPDU) provides a solution to measure and collect
power data. An iPDU (FC 5889) mounts in a rack and provides power outlets for the servers
to plug into.
The following list shows the characteristics of an iPDU:
Input connectorConnect power cord to this connector
Power outletsPower outlet for devices. There are nine or 12 power
outlets, depending on the model
RS232 serial connectorUpdate firmware
RJ45 console connectorProvides a connection using a DB9-to-RJ45 provided cable
to a notebook computer as a configuration console.
RJ45 Ethernet (LAN) connector Port to configure the iPDU through a LAN. Speed is 10/100
auto sensed.
When a configured iPDU is selected the following dialog panel (as in IBM Systems Director)
will appear as shown in Figure 1-4 on page 17.
16IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 22, 2009 3:10 pm4405ch01 General description.fm
RS232 serial
RJ-45 LAN
connector
Figure 1-4 Intelligent Power Distribution Unit
Individual outlet
RJ-45 console
connector
Outlet group
Input connector
In this panel, outlet names and outlet group names are shown. Each iPDU node will contain
either node group outlets or individual outlets. For further information of integration and IBM
Director functionality, see:
The system consists of one to four CEC enclosures. Each enclosure occupies 4U of vertical
rack space. The primary considerations that should be accounted for when mounting the
system into a rack are:
The Power 570 is designed to be placed at any location in the rack. For rack stability, it is
advisable to start filling a rack from the bottom.
For configurations with two, three, or four drawers, all drawers must be installed together in
the same rack, in a contiguous space of 8 U, 12 U, or 16 U within the rack. The uppermost
enclosure in the system is the base enclosure. This enclosure will contain the primary
active Service Processor and the Operator Panel.
Any remaining space in the rack can be used to install other systems or peripherals,
provided that the maximum permissible weight of the rack is not exceeded and the
installation rules for these devices are followed.
The 7014-T42 rack is constructed with a small flange at the bottom of EIA location 37.
When a system is installed near the top of a 7014-T42 rack, no system drawer can be
installed in EIA positions 34, 35, or 36. This is to avoid interference with the front bezel or
with the front flex cable, depending on the system configuration. A two-drawer system
cannot be installed above position 29. A three-drawer system cannot be installed above
position 25. A four-drawer system cannot be installed above position 21. (The position
number refers to the bottom of the lowest drawer.)
Chapter 1. General description 17
4405ch01 General description.fmDraft Document for Review May 22, 2009 3:10 pm
When a system is installed in an 7014-T00 or -T42 rack that has no front door, a Thin
Profile Front Trim Kit must be ordered for the rack. The required trim kit for the 7014-T00
rack is FC 6246. The required trim kit for the 7014-T42 rack is FC 6247.
The design of the 570 is optimized for use in a 7014-T00 or -T42 rack. Both the front cover
and the processor flex cables occupy space on the front left side of an IBM 7014 rack that
may not be available in typical non-IBM racks.
Acoustic Door features are available with the 7014-T00 and 7014-T42 racks to meet the
lower acoustic levels identified in the specification section of this document. The Acoustic
Door feature can be ordered on new T00 and T42 racks or ordered for the T00 and T42
racks that clients already own.
1.4.6 Useful rack additions
This section highlights useful additions to a rack.
IBM 7214 Model 1U2 SAS Storage Enclosure
IBM 7212 Model 102 IBM TotalStorage storage device enclosure
The IBM 7212 Model 102 is designed to provide efficient and convenient storage expansion
capabilities for selected System p servers. The IBM 7212 Model 102 is a 1 U rack-mountable
option to be installed in a standard 19-inch rack using an optional rack-mount hardware
feature kit. The 7212 Model 102 has two bays that can accommodate any of the following
storage drive features:
A Digital Data Storage (DDS) Gen 5 DAT72 Tape Drive provides a physical storage
capacity of 36 GB (72 GB with 2:1 compression) per data cartridge.
A VXA-2 Tape Drive provides a media capacity of up to 80 GB (160 GB with 2:1
compression) physical data storage capacity per cartridge.
A Digital Data Storage (DDS-4) tape drive provides 20 GB native data capacity per tape
cartridge and a native physical data transfer rate of up to 3 MBps that uses a 2:1
compression so that a single tape cartridge can store up to 40 GB of data.
A DVD-ROM drive is a 5 1/4-inch, half-high device. It can read 640 MB CD-ROM and
4.7 GB DVD-RAM media. It can be used for alternate IPL
3
(IBM-distributed CD-ROM
media only) and program distribution.
A DVD-RAM drive with up to 2.7 MBps throughput. Using 3:1 compression, a single disk
can store up to 28 GB of data. Supported DVD disk native capacities on a single
DVD-RAM disk are as follows: up to 2.6 GB, 4.7 GB, 5.2 GB, and 9.4 GB.
Flat panel display options
The IBM 7316-TF3 Flat Panel Console Kit can be installed in the system rack. This 1 U
console uses a 15-inch thin film transistor (TFT) LCD with a viewable area of
304.1 mm x 228.1 mm and a 1024 x 768 pels
Kit has the following attributes:
Flat panel color monitor
Rack tray for keyboard, monitor, and optional VGA switch with mounting brackets
4
resolution. The 7316-TF3 Flat Panel Console
IBM Travel Keyboard mounts in the rack keyboard tray (Integrated Trackpoint and
UltraNav)
3
Initial program load
4
Picture elements
18IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 22, 2009 3:10 pm4405ch01 General description.fm
IBM PS/2 Travel Keyboards are supported on the 7316-TF3 for use in configurations where
only PS/2 keyboard ports are available.
The IBM 7316-TF3 Flat Panel Console Kit provides an option for the USB Travel Keyboards
withUltraNav.The keyboard enables the 7316-TF3 to be connected to systems that do not
have PS/2keyboard ports. The USB Travel Keyboard can be directly attached to an available
integrated USB port or a supported USB adapter (FC 2738) on System p5 servers or
7310-CR3 and 7315-CR3 HMCs.
The Netbay LCM (Keyboard/Video/Mouse) Switch (FC 4202) provides users single-point
access and control of up to 64 servers from a single console. The Netbay LCM Switch has a
maximum video resolution of 1600 x 280 and mounts in a 1 U drawer behind the 7316-TF3
monitor. A minimum of one LCM feature (FC 4268) or USB feature (FC 4269) is required with
a Netbay LCM Switch (FC 4202). Each feature can support up to four systems. When
connecting to a Power 570, FC 4269 provides connection to the server USB ports.
When selecting the LCM Switch, consider the following information:
The KVM Conversion Option (KCO) cable (FC 4268) is used with systems with PS/2 style
keyboard, display, and mouse ports.
The USB cable (FC 4269) is used with systems with USB keyboard or mouse ports.
The switch offers four ports for server connections. Each port in the switch can connect a
maximum of 16 systems:
– One KCO cable (FC 4268) or USB cable (FC 4269) is required for every four systems
supported on the switch.
– A maximum of 16 KCO cables or USB cables per port can be used with the Netbay
Note: A server microcode update might be required on installed systems for boot-time
System Management Services (SMS) menu support of the USB keyboards. The update
might also be required for the LCM switch on the 7316-TF3 console (FC 4202). For
microcode updates, see the following URL:
We recommend that you have the 7316-TF3 installed between EIA 20 to 25 of the rack for
ease of use. The 7316-TF3 or any other graphics monitor requires the POWER GXT135P
graphics accelerator (FC 1980) to be installed in the server, or some other graphics
accelerator, if supported.
1.4.7 OEM rack
The system can be installed in a suitable OEM rack, provided that the rack conforms to the
EIA-310-D standard for 19-inch racks. This standard is published by the Electrical Industries
Alliance, and a summary of this standard is available in the publication IBM System p5, eServer p5 and i5, and OpenPower Planning, SA38-0508.
The key points mentioned in this documentation are as follows:
The front rack opening must be 451 mm wide + 0.75 mm (17.75 in. + 0.03 in.), and the
rail-mounting holes must be 465 mm + 0.8 mm (18.3 in. + 0.03 in.) apart on center
(horizontal width between the vertical columns of holes on the two front-mounting flanges
and on the two rear-mounting flanges). See Figure 1-5 on page 20 for a top view showing
the specification dimensions.
LCM Switch to connect up to 64 servers.
Chapter 1. General description 19
4405ch01 General description.fmDraft Document for Review May 22, 2009 3:10 pm
Figure 1-5 Top view of non-IBM rack specification dimensions
The vertical distance between the mounting holes must consist of sets of three holes
spaced (from bottom to top) 15.9 mm (0.625 in.), 15.9 mm (0.625 in.), and 12.67 mm
(0.5 in.) on center, making each three-hole set of vertical hole spacing 44.45 mm (1.75 in.)
apart on center. Rail-mounting holes must be 7.1 mm + 0.1 mm (0.28 in. + 0.004 in.) in
diameter. See Figure 1-6 and Figure 1-7 on page 21 for the top and bottom front
specification dimensions.
Figure 1-6 Rack specification dimensions, top front view
20IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 22, 2009 3:10 pm4405ch01 General description.fm
Figure 1-7 Rack specification dimensions, bottom front view
It might be necessary to supply additional hardware, such as fasteners, for use in some
manufacturer’s racks.
The system rack or cabinet must be capable of supporting an average load of 15.9 kg
(35 lb.) of product weight per EIA unit.
The system rack or cabinet must be compatible with drawer mounting rails, including a
secure and snug fit of the rail-mounting pins and screws into the rack or cabinet rail
support hole.
Note: The OEM rack must only support ac-powered drawers. We strongly recommend that
you use a power distribution unit (PDU) that meets the same specifications as the PDUs to
supply rack power. Rack or cabinet power distribution devices must meet the drawer power
requirements, as well as the requirements of any additional products that will be connected
to the same power distribution device.
Chapter 1. General description 21
4405ch01 General description.fmDraft Document for Review May 22, 2009 3:10 pm
22IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
2
Chapter 2.Architecture and technical
overview
This chapter discusses the overall system architecture represented by Figure 2-1, with its
major components described in the following sections. The bandwidths that are provided
throughout the section are theoretical maximums used for reference. You should always
obtain real-world performance measurements using production workloads.
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
2.1 The POWER6 processor
The POWER6 processor capitalizes on all the enhancements brought by the POWER5 chip.
Two of the enhancements of the POWER6 processor is the ability to do processor instruction
retry and alternate processor recovery. This significantly reduces exposure to both hard
(logic) and soft (transient) errors in the processor core.
Processor instruction retry
Soft failures in the processor core are transient errors. When an error is
encountered in the core, the POWER6 processor will first automatically retry the
instruction. If the source of the error was truly transient, the instruction will succeed
and the system will continue as before. On predecessor IBM systems, this error
would have caused a checkstop.
Alternate processor retry
Hard failures are more difficult, being true logical errors that will be replicated each
time the instruction is repeated. Retrying the instruction will not help in this situation
because the instruction will continue to fail. Systems with POWER6 processors
introduce the ability to extract the failing instruction from the faulty core and retry it
elsewhere in the system, after which the failing core is dynamically deconfigured
and called out for replacement. The entire process is transparent to the partition
owning the failing instruction. Systems with POWER6 processors are designed to
avoid what would have been a full system outage.
POWER6 single processor checkstopping
Another major advancement in POWER6 processors is single processor
checkstopping. A processor checkstop would result in a system checkstop. A new
feature in the 570 is the ability to contain most processor checkstops to the partition
that was using the processor at the time. This significantly reduces the probability of
any one processor affecting total system availability.
POWER6 cache availability
In the event that an uncorrectable error occurs in L2 or L3 cache, the system will be
able to dynamically remove the offending line of cache without requiring a reboot. In
addition POWER6 utilizes a L1/L2 cache design and a write-through cache policy
on all levels, helping to ensure that data is written to main memory as soon as
possible.
Figure 2-2 on page 25 shows a high-level view of the POWER6 processor.
24IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
POWER6
Fabric bus
controller
core
4.2 GHz
1.9 GHz
L2
4 MB
GX Bus
Controller
Alti
Vec
32 MB
L3 cache
POWER5
POWER6
Alti
core
Vec
1.9 GHz
4.2 GHz
L2
L3
4 MB
Ctrl
Memory
Controller
Memory+GX+ Bridge
Figure 2-2 POWER6 processor
The CMOS 11S0 lithography technology in the POWER6 processor uses a 65 nm fabrication
process, which enables:
Performance gains through faster clock rates from 3.5 GHz, 4.2 GHz up to 4.7 GHz.
Physical size of 341 mm.
The POWER6 processor consumes less power and requires less cooling. Thus, you can use
the POWER6 processor in servers where previously you could only use lower frequency
chips due to cooling restrictions.
The 64-bit implementation of the POWER6 design provides the following additional
enhancements:
Compatibility of 64-bit architecture
– Binary compatibility for all POWER and PowerPC® application code level
– Support of partition migration
– Support of virtualized partition memory
– Support of four page sizes : 4 KB, 64 KB, 16 MB, and 16 GB
High frequency optimization
– Designed to operate at maximum speed of 5 GHz
Superscalar core organization
– Simultaneous Multithreading: two threads
In order dispatch of five operations (single thread), seven operations (Simultaneous
Multithreading) to nine execution units:
• Two load or store operations
• Two fixed-point register-register operations
• Two floating-point operations
• One branch operation
Chapter 2. Architecture and technical overview 25
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
The POWER6 processor implements the 64-bit IBM Power Architecture® technology. Each
POWER6 chip incorporates two dual-threaded Simultaneous Multithreading processor cores,
a private 4 MB level 2 cache (L2) for each processor, a 36 MB L3 cache controller shared by
the two processors, integrated memory controller and data interconnect switch and support
logic for dynamic power management, dynamic configuration and recovery, and system
monitoring.
2.1.1 Decimal floating point
This section describes the behavior of the decimal floating-point processor, the supported
data types, formats, and classes, and the usage of registers.
The decimal floating-point (DFP) processor shares the 32 floating-point registers (FPRs) and
the floating-point status and control register (FPSCR) with the binary floating-point (BFP)
processor. However, the interpretation of data formats in the FPRs, and the meaning of some
control and status bits in the FPSCR are different between the BFP and DFP processors.
The DFP processor supports three DFP data formats:
DFP32 (single precision)
DFP64 (double precision)
DFP128 (quad precision)
Most operations are performed on the DFP64 or DFP128 format directly. Support for DFP32
is limited to conversion to and from DFP64. For some operations, the DFP processor also
supports operands in other data types, including signed or unsigned binary fixed-point data,
and signed or unsigned decimal data.
DFP instructions are provided to perform arithmetic, compare, test, quantum-adjustment,
conversion, and format operations on operands held in FPRs or FPR pairs.
Arithmetic instructions These instructions perform addition, subtraction, multiplication, and
division operations.
Compare instructionsThese instructions perform a comparison operation on the
numerical value of two DFP operands.
Test instructionsThese instructions test the data class, the data group, the
exponent, or the number of significant digits of a DFP operand.
Quantum-adjustment instructions
These instructions convert a DFP number to a result in the form
that has the designated exponent, which may be explicitly or
implicitly specified.
Conversion instructionsThese instructions perform conversion between different data
formats or data types.
Format instructionsThese instructions facilitate composing or decomposing a DFP
operand.
For example, SAP NetWeaver® 7.10 ABAP™ kernel introduces a new SAP ABAP data type
called
DECFLOAT to enable more accurate and consistent results from decimal floating point
computations. The decimal floating point (DFP) support by SAP NetWeaver leverages the
built-in DFP feature of POWER6 processors. This allows for highly simplified ABAP-coding
while increasing numeric accuracy and with a potential for significant performance
improvements.
26IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
2.1.2 AltiVec and Single Instruction, Multiple Data
IBM Semiconductor’s advanced Single Instruction, Multiple Data (SIMD) technology based on
the AltiVec instruction set is designed to enable exceptional general-purpose processing
power for high-performance POWER processors. This leading-edge technology is engineered
to support high-bandwidth data processing and algorithmic-intensive computations, all in a
single-chip solution
With its computing power, AltiVec technology also enables high-performance POWER
processors to address markets and applications in which performance must be balanced with
power consumption, system cost and peripheral integration.
The AltiVec technology is a well known environment for software developers who want to add
efficiency and speed to their applications. A 128-bit vector execution unit was added to the
architecture. This engine operates concurrently with the existing integer and floating-point
units and enables highly parallel operations, up to 16 operations in a single clock cycle. By
leveraging AltiVec technology, developers can optimize applications to deliver acceleration in
performance-driven, high-bandwidth computing.
The AltiVec technology is not comparable to the IBM POWER6 processor implementation,
using the Simultaneous Multithreading functionality.
2.2 IBM EnergyScale technology
IBM EnergyScale™ technology is featured on the IBM POWER6 processor-based systems. It
provides functions to help the user understand and control IBM server power and cooling
usage.
In this section we will describe IBM EnergyScale features and hardware and software
requirements
Power TrendingEnergyScale provides continuous power usage data collection. This
enables the administrators with the information to predict power
consumption across their infrastructure and to react to business and
processing needs. For example, an administrator could adjust server
consumption to reduce electrical costs. To collect power data for the
570 you need to power it through an Intelligent Power Distribution
Unit (iPDU). Other systems that support power trending collect the
information internally and do not require any additional hardware.
Power Saver ModePower Saver Mode reduces the voltage and frequency by a fixed
percentage. This percentage is predetermined to be within a safe
operating limit and is not user configurable. Under current
implementation this is a 14% frequency drop. When CPU utilization is
low, Power Saver Mode has no impact on performance. Power Saver
Mode can reduce the processor usage up to a 30%. Power Saver
Mode is not supported during boot or re-boot although it is a
persistent condition that will be sustained after the boot when the
system starts executing instructions. Power Saver is only supported
with 4.2 GHz processors and faster.
Power CappingPower Capping enforces a user specified limit on power usage.
Power Capping is not a power saving mechanism. It enforces power
caps by actually throttling the processor(s) in the system, degrading
performance significantly. The idea of a power cap is to set
something that should never be reached but frees up
margined
Chapter 2. Architecture and technical overview 27
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
power in the data center. The margined power is the amount of extra
power that is allocated to a server during its installation in a
datacenter. It’s based on the server environmental specifications that
usually are never reached. Since server specifications are always
based on maximum configurations and worst case scenarios.
Processor Core Nap The IBM POWER6 processor uses a low-power mode called Nap
that stops processor execution when there is no work to do on that
processor core, both threads are idle. Nap mode allows the hardware
to clock off most of the circuits inside the processor core. Reducing
active power consumption by turning off the clocks allows the
temperature to fall, which further reduces leakage (static) power of
the circuits causing a cumulative effect. Unlicensed cores are kept in
core Nap until they are licensed and return to core Nap whenever
they are unlicensed again.
EnergyScale for I/OIBM POWER6 processor-based systems automatically power off
pluggable, PCI adapter slots that are empty or not being used to save
approximately 14 watts per slot. System firmware automatically
scans all pluggable PCI slots at regular intervals looking for ones that
meet the criteria for being not in use and powers them off. This
support is available for all POWER6 processor-based servers, and
the expansion units that they support. Note that it applies to hot
pluggable PCI slots only.
2.2.1 Hardware and software requirements
This sections summarizes the supported systems and software user interfaces for
EnergyScale functions.
Table 2-1 EnergyScale systems support
Power
trending
7998-61XYYYYN
8203-E4AYYYYY
8204-E8AYYYYY
9117-MMA (<
4.20 GHz)
9117-MMA
(>= 4.2 GHz)
a. An iPDU is required for this support. The feature code 5889 is for a base iPDU in a rack while
the feature code 7109 is for additional iPDUs for the same rack. Supported racks are 7014-B42,
7014-S25, 7014-T00 and 7014-T42.
b. Only supported if GX Dual Port RIO-2 Attach (FC 1800) is not present.
Y - via iPDU
Y - via iPDU
The primary user interface for EnergyScale features on a POWER6 based system is IBM
Systems Director Active Energy Manager™ running within IBM Director
Power saver
mode
a
NNYY
a
b
Y
Power
capping
NYY
Processor
Nap
I/O
Table 2-2 on page 29 shows the ASMI, HMC and Active Energy Manager interface support.
28IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
In the 570, the POWER6 processors, associated L3 cache chip, and memory DIMMs are
packaged in processor cards. The 570 uses a dual-core processor module for a 2-core,
4-core, 8-core, 12-core, and 16-core configuration running at 3.5 GHz, 4.2 GHz, or 4.7 GHz.
The 570 has two processor sockets on the system planar. Each socket will accept a
processor card feature. A single CEC may have one or two processor cards installed. A
system with two, three, or four CEC must have two processor cards in each CEC.
Each processor can address all the memory on the processor card. Access to memory
behind another processor is accomplished through the fabric buses.
The 2-core 570 processor card contains a dual-core processor chip, a 36 MB L3 cache chip
and the local memory storage subsystem.
NYY
NNY
Figure 2-3 shows a layout view of a 570 processor card and associated memory.
L3
Memory dimms
Cache
Memory
controller
POWER6
AltivecAltivec
core
L2 Cache
4 MB
L3 ctrl
Fabric bus
Memory
controller
POWER6
core
L2 Cache
4 MB
Figure 2-3 The 570 processor card with DDR2 memory socket layout view
The storage structure for the POWER6 processor is a distributed memory architecture that
provides high-memory bandwidth, although each processor can address all memory and
Chapter 2. Architecture and technical overview 29
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
sees a single shared memory resource. They are interfaced to 12 memory slots, where as
each memory DIMM has its own memory buffer chip and are interfaced in a point-to-point
connection.
I/O connects to the 570 processor module using the GX+ bus. The processor module
provides a single GX+ bus. The GX+ bus provides an interface to I/O devices through the
RIO-2 connections or a 12X Channel attach connections.
2.3.1 Processor drawer interconnect cables
In combined systems that are made of more than one 570 building block, the connection
between processor cards in different building blocks is provided with a processor drawer
interconnect cable. Different processor drawer interconnect cables are required for the
different numbers of 570 building blocks that a combined system can be made of, as shown in
Figure 2-4.
Because of the redundancy and fault recovery built-in to the system interconnects, a drawer
failure does not represent a system failure. Once a problem is isolated and repaired, a system
reboot may be required to reestablish full bus speed, if the failure was specific to the
interconnects.
The SMP fabric bus that connects the processors of separate 570 building blocks is routed on
the interconnect cable that is routed external to the building blocks. The flexible cable
attaches directly to the processor cards, at the front of the 570 building block, and is routed
behind the front covers (bezels) of the 570 building blocks. There is an optimized cable for
each drawer configuration. Figure 2-4 illustrates the logical fabric bus connections between
the drawers, and shows the additional space required left of the bezels for rack installation.
Figure 2-4 Logical 570 building block connection
2.3.2 Processor clock rate
The 570 system features base 2-core, 4-core, 8-core, 12-core, and 16-core configurations
with the POWER6 processor running at 3.5 GHz, 4.2 GHz, and 4.7 GHz.
30IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
Note: Any system made of more than one processor card must have all processor cards
running at the same speed.
To verify the processor characteristics on a system running at 4.2 GHz, use one of the
following commands:
lsattr -El procX
Where X is the number of the processor, for example, proc0 is the first processor in the
system. The output from the command is similar to the following output (
this output, signifies that the value cannot be changed through an AIX command
interface):
frequency ..4208000000 ........Processor Speed False
The pmcycles command (available with AIX) uses the performance monitor cycle counter
and the processor real-time clock to measure the actual processor clock speed in MHz.
The following output is from a 4-core 570 system running at 4.2 GHz with simultaneous
multithreading enabled:
Cpu 0 runs at 4208 MHz
Cpu 1 runs at 4208 MHz
Cpu 2 runs at 4208 MHz
Cpu 3 runs at 4208 MHz
Cpu 4 runs at 4208 MHz
Cpu 5 runs at 4208 MHz
Cpu 6 runs at 4208 MHz
Cpu 7 runs at 4208 MHz
Note: The pmcycles command is part of the bos.pmapi fileset. Use the lslpp -l
bos.pmapi command to determine if it is installed on your system.
2.4 Memory subsystem
When you consider a 570 initial order, the memory controller is internal to the POWER6
processor and it interfaces any of the memory buffer chips within the pluggable fully buffered
DIMMs (12 slots available per processor card, as described in 1.3.2, “Memory features” on
page 6).
2.4.1 Fully buffered DIMM
Fully buffered DIMM is a memory technology which can be used to increase reliability, speed
and density of memory subsystems. Conventionally, data lines from the memory controller
have to be connected to data lines in every DRAM module. As memory width, as well as
access speed, increases, the signal degrades at the interface of the bus and the device. This
limits the speed or the memory density. The fully buffered DIMMs take a different approach
because directs signaling interface between the memory controller and the DRAM chips,
splitting it into two independent signaling interfaces with a buffer between them. The interface
Chapter 2. Architecture and technical overview 31
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
between the memory controller and the buffer is changed from a shared parallel interface to a
point-to-point serial interface (see Figure 2-5).
POWER6
chip
Memory
controller
Common clock
source
Figure 2-5 Fully buffered DIMMs architecture
The result of the fully buffered memory DIMMs implementation is an enhanced scalability and
throughput.
2.4.2 Memory placements rules
DRAM
DRAM
DRAM
DRAM
Memory
buffer
chip
DRAM
DRAM
DRAM
DRAM
DRAM
DRAM
DRAM
DRAM
Memory
buffer
chip
DRAM
DRAM
DRAM
DRAM
...
DRAM
DRAM
DRAM
DRAM
Memory
buffer
chip
DRAM
DRAM
DRAM
DRAM
The minimum memory capacity for a 570 initial order is 2 GB when a 3.5 GHz, 4.2 GHz, or
4.7 GHz system is configured with two processor-cores. FC 5620, FC 5622, and FC 7380
processor cards support up to 12 fully buffered DIMM slots and DIMMs must be installed in
quads. Then the quads are organized as follows:
First quad includes J0A, J0B, J0C, and J0D memory slots
Second quad includes J1A, J1B, J1C, and J1D memory slots
Third quad includes J2A, J2B, J2C, and J2D memory slots
See Figure 2-6 on page 33to locate any available quad.
32IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
J2B
J1B
J0B
L3 cache
POWER6
chip
Figure 2-6 Memory DIMM slots for FC 5620, FC 5622, and FC 7380
J2D
J1D
J0D
J2A
J1A
J0A
J2C
J1C
J0C
In addition to the quad placement rules, minimum memory required depends from the number
of processor-cores configured in the 570:
2 GB is the minimum memory required for a 2-core system
4 GB is the minimum memory required for a 4-core system
8 GB is the minimum memory required for an 8-core system
16 GB is the minimum memory required for a 16-core system
Every processor card in a 570 configuration requires a memory quad.
The maximum installable memory is 192 GB per any 570 drawer, thus a fully configured 570
supports up to 768 GB (48 GB per processor-core).
When configuring the memory in a 570, placing 2 memory features (8 DIMMs) on a single
processor card will provide the maximum available memory bandwidth. Adding the third
memory feature will provide additional memory capacity but will not increase memory
bandwidth. System performance that is dependent on memory bandwidth can be improved by
purchasing two smaller features per processor card as opposed to one large feature per
processor card. To achieve this, when placing an order, ensure the order has 2X memory
features for every processor card feature on the order.
2.4.3 Memory consideration for model migration from p5 570 to 570
A p5 570 (based on POWER5 or POWER5+ processor) can be migrated to a 570. Since the
570 supports only DDR2 memory, if the initial p5 570 server to migrate has DDR2 memory, it
can be migrated to the target 570 that requires FC 5621 processor card to accept it.
Additional memory can be also included in the model migration order.
In FC 5621 processor card, the memory controller interfaces to four memory buffer chips per
processor card with 8 memory slots available to be populated with available DDR2 memory
DIMMs migrated from the p5 570 server (see Figure 2-7 on page 34 for memory DIMM slots
reference).
Chapter 2. Architecture and technical overview 33
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
If the initial p5 570 server does not configure supported DDR2 memory, then the target 570
can be configured with FC 5622 processor card and the desired amount of memory must be
included in the model migration order.
J1D
J0D
J0C
chip
A
J1C
Memory buffer
J1A
J0A
J0B
J1B
chip
Memory buffer
C
chip
B
L3 cache
POWER6
chip
Memory buffer
chip
D
Memory buffer
Figure 2-7 Memory DIMM slots for FC 5621
Important: The process to migrate a p5 570 to a 570 requires analysis of the existing
p5 570 memory DIMMs. Contact an IBM service representative before issuing the
configuration upgrade order.
A 570 with FC 5621 processor cards can be expanded by purchasing additional 570
enclosures with FC 5622 processor cards. FC 5621 and FC 5622 cannot be mixed within the
same 570 enclosure but can be mixed in the same system. Maximum memory configurable
depends from the number of FC 5621 and FC 5622 processor cards available in the fully
combined 570 system.
2.4.4 OEM memory
OEM memory is not supported or certified for use in IBM System p servers. If the 570 is
populated with OEM memory, you could experience unexpected and unpredictable behavior,
especially when the system is using Micro-Partitioning technology.
All IBM memory is identified by an IBM logo and a white label that is printed with a barcode
and an alphanumeric string, as illustrated in Figure 2-8 on page 35.
34IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
Figure 2-8 IBM memory certification label
2.4.5 Memory throughput
The memory subsystem throughput is based on the speed of the memory. On processor,
there are four memory channels, each with single 2 byte read and 1 byte write. Memory
channels of POWER6 memory controller are connected to Memory buffers. The processor
chip has two POWER6 processors. The DDR2 bus allows double reads or writes per clock
cycle. If a 667 MHz memory feature is selected, the throughput is (4 x 2 x 2 x 2 x 667) + (4 x 1
x 2 x 2 x 667) or 32016 MBps or 32 GBps. These values are maximum theoretical
throughputs for comparison purposes only.
Table 2-3 provides the theoretical throughput values of 4.7 GHz processors and 667 Mhz
memory configuration.
Table 2-3 Theoretical throughput values
Memory Bandwidth
L1 (Data)75.2 GB/sec
L2 / Chip300.8 GB/sec
L3 / Chip37.6 GB/sec
Memory / Chip32 GB/sec
Inter-Node Buses (16-cores)75.2 GB/sec
Intra-Node Buses (16-cores)100.26 GB/sec
2.5 System buses
The following sections provide additional information related to the internal buses.
2.5.1 I/O buses and GX+ card
Each POWER6 processor provides a GX+ bus which is used to connect to an I/O subsystem
or Fabric Interface card. The processor card populating the first processor slot is connected to
the GX+ multifunctional host bridge chip which provides the following major interfaces:
One GX+ passthru bus:
GX+ passthru elastic interface runs at one half the frequency of the primary. It allows other
GX+ bus hubs to be connected into the system.
Chapter 2. Architecture and technical overview 35
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
Two 64-bit PCI-X 2.0 buses, one 64-bit PCI-X 1.0 bus, and one 32-bit PCI-X 1.0 bus
Four 8x PCI Express links
Two 10 Gbps Ethernet ports: Each port is individually configurable to function as two
1 Gb/s port
In a fully populated 570, there are two GX+ buses, one from each processor. Each 570 has 2
GX+ slots with a single GX+ bus. The GX+ multifunctional host bridge provide a dedicated
GX+ bus routed to the first GX+ slot through GX+ passthru bus. The second GX+ slot is not
active unless the second processor card is installed. It is not required for CoD cards to be
activated in order for the associated GX+ bus to be active.
Optional Dual port RIO-2 I/O Hub (FC 1800) and Dual port 12x Channel Attach (FC 1802)
adapters that are installed in the GX+ slots are used for external DASD and IO drawer
expansion. All GX+ cards are Hot-Pluggable.
Table 2-4 provides I/O bandwidth of 4.7 GHz processors configuration.
Table 2-4 I/O bandwidth
I/OBandwidth
Total I/O62.6 GB/sec (16-cores)
Primary GX Bus9.4 GB/sec (per node)
GX Bus Slot 14.7 GB/sec (per node)
GX Bus Slot 26.266 GB/sec (per node)
2.5.2 Service processor bus
The Service Processor (SP) flex cable is at the rear of the system and is used for SP
communication between the system drawers. The SP cable remains similar to the p5 570 in
that there is a unique SP cable for each configuration, but the p5 570 SP cables cannot be
used for 570. Although, SP function is implemented in system drawer 1 and system drawer 2,
Service interface card is required in every system drawer for signal distribution functions
inside the system drawer. There is a unique SP cable for each drawer as Figure 2-9 on
page 37 shows.
36IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
Figure 2-9 SP Flex™ cables
FC 5657: 2-drawer SP cable
FC 5658: 3-drawer SP cable
FC 5660: 4-drawer SP cable.
2.6 Internal I/O subsystem
The internal I/O subsystem resides on the system planar which supports a mixture of both
PCIe and PCI-X slots. All PCIe or PCI-X slots are hot pluggable and Enhanced Error
Handling (EEH) enabled. In the unlikely event of a problem, EEH-enabled adapters respond
to a special data packet generated from the affected PCIe or PCI-X slot hardware by calling
system firmware, which will examine the affected bus, allow the device driver to reset it, and
continue without a system reboot.
Table 2-5 display slot configuration of 570.
Table 2-5 Slot configuration of a 570
Slot#DescriptionLocation codePHBMax Card Size
Slot 1PCIe x8P1-C1PCIe PHB0Long
Slot 2PCIe x8P1-C2PCIe PHB1Long
Slot 3PCIe x8P1-C3PCIe PHB2Long
Slot 4PCI-X DDR,
64-bit, 266 MHz
Slot 5 PCI-X DDR,
64-bit, 266 MHz
P1-C4PCI-X PHB1Long
P1-C5PCI-X PHB3Long
Chapter 2. Architecture and technical overview 37
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
Slot#DescriptionLocation codePHBMax Card Size
Slot 6PCIe x8
Slot 7GX+P1-C9
Adapter slots P1-C6 and P1-C8 share the same physical space in a system enclosure. When
a GX+ adapter is installed in GX slot P1-C8, PCIe slot P1-C6 cannot be used.
The 570 uses generation 3, blind swap cassettes to manage the installation and removal of
adapters. Cassettes can be installed and removed without removing the drawer from the rack.
2.6.1 System ports
Although each system drawer is equipped with an Integrated Virtual Ethernet adapter (IVE)
that has up to two serial ports, only the serial ports which are located on the system drawer
(non-IVE) communicate with the service processor. They are called system ports. In
operating system environment, the system ports become host virtual system ports and are
not general RS232 serial port, but rather are limited use ports available for specifically
supported functions.
The use of the integrated system ports on a 570 is limited to serial connected TTY console
functionality and IBM approved call-home modems. These system ports do not support other
general serial connection uses, such as UPS, HACMP heartbeat, printers, mice, track balls,
space balls, etc. If you need serial port function, optional PCI adapters which are described in
2.8.6, 2.8.6, “Asynchronous PCI adapters” on page 47 are available.
GX+
P1-C6
P1-C8
PCIe PHB3Short
If an HMC is connected, a
under AIX), and you can also connect a modem to the HMC. The system ports are not usable
in this case. Either the HMC ports or the system ports can be used, but not both.
Configuration of the system ports, including basic ports settings (baud rate, etc.), modem
selection, and call-home, can be accomplished with the Advanced Systems Management
Interface (ASMI).
Note: The 570 must have an HMC. In normal operation, the system ports are for service
representatives only.
virtual serial console is provided by the HMC (logical device vsa0
2.7 Integrated Virtual Ethernet adapter
The POWER6 processor-based servers extend the virtualization technologies introduced in
POWER5 by offering the Integrated Virtual Ethernet adapter (IVE). IVE, also named Host
Ethernet Adapter (HEA) in other documentation, enables an easy way to manage the sharing
of the integrated high-speed Ethernet adapter ports. It is a standard set of features that are
part of every POWER6 processor-based server.
Integrated Virtual Ethernet adapter is a 570 standard feature but you can select from different
options. The IVE comes from a general market requirements for improved performance and
virtualization for Ethernet. It offers:
Either two 10 Gbps Ethernet ports or four 1 Gbps ports or two 1 Gbps integrated ports
A low cost Ethernet solution for low-end andmid-range System p servers
38IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
Virtual Ethernet resources without the Virtual I/O Server
Designed to operate at media speeds
The IVE is a physical Ethernet adapter that is connected directly to the GX+ bus instead of
being connected to a PCIe or PCI-X bus, either as an optional or integrated PCI adapter. This
provides IVE high throughput, and low latency. IVE also includes special hardware features to
provide logical Ethernet adapters that can communicate to logical partitions (LPAR) reducing
the use of POWER Hypervisor™ (PHYP).
IVE design provides a direct connection for multiple LPARs to share its resources. This allows
LPARs to access external networks through the IVE without having to go through an Ethernet
bridge on another logical partition, such as a Virtual I/O Server. Therefore, this eliminates the
need to move packets (using virtual Ethernet) between partitions and then through a shared
Ethernet adapter (SEA) to an Ethernet port. LPARs can share IVE ports with improved
performance.
Using Virtual I/O Ser ver Shared Ethernet Adapt erUsing Integrated V irtual Ethernet
Partition
Packet
Forwarder
Network Adapters
AIXHosting
Virtual
Ethernet
Driver
Virtual Ethernet Switch
Hypervisor
AIXLinux
Virtual
Ethernet
Driver
Virtual
Ethernet
Driver
LAN, WAN, ...
AIXAIXLinux
Virtual
Ethernet
Driver
Virtual
Ethernet
Driver
Integrated Virtual Et hernet
Virtual
Ethernet
Driver
Figure 2-10 Integrated Virtual Ethernet compared to Virtual I/O Server Shared Ethernet Adapter
IVE supports 2 or 4 Ethernet ports running at 1 Gbps and 2 ports running at 10 Gbps
depending on the IVE feature ordered. In the case of a 570 server, clients that are using
1 Gbps connection bandwidth in their IT infrastructure could move up to 10 Gbps
infrastructure by adding a new 570 enclosure with the Integrated 2-ports 10 Gbps virtual
Ethernet.
After any IBM System p initial order, you must use the MES
1
process to make any changes in
the system configuration.
For more information on IVE features read Integrated Virtual Ethernet Technical Overview and Introduction, REDP-4340.
2.7.1 Physical ports and system integration
The following sections discuss the physical ports and the features available at the time of
writing on a 570.
Each 570 enclosure can have unique Integrated Virtual Ethernet adapters, so a fully
configured 570 server can be comprised of several different IVE feature codes.
1
MES stands for Miscellaneous Equipment Shipment. It is the IBM process for IBM system upgrade
Chapter 2. Architecture and technical overview 39
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
The following feature codes are available, at the time of writing, for each 570 enclosure:
Figure 2-11 shows the major components of the Integrated Virtual Ethernet adapter hardware
and additional system ports, according to the different feature codes.
FC 5636 2-ports 1 Gbps assembly
FC 5636 2-ports 1 Gbps assembly
VPD chip
VPD chip
2-ports 1 Gbps
2-ports 1 Gbps
Virtual Ethernet
Virtual Ethernet
Serial
Serial
Port 2/
Port 2 /
UPS
UPS
Serial
Serial
Port
Port
FC 5639 4-ports 1 Gbps assembly
FC 5639 4-ports 1 Gbps assembly
VPD chip
VPD chip
4-ports 1 Gbps
4-ports 1 Gbps
Virtual Ethernet
Virtual Ethernet
Serial
Serial
Port 2 /
Port 2 /
UPS
UPS
FC 5637 2-ports 10 Gbps assembly
FC 5637 2-ports 10 Gbps assembly
VPD chip
VPD chip
2-ports 10 Gbps
2-ports 10 Gbps
Virtual Ethernet
Virtual Ethernet
Serial
Serial
Port 2 /
Port 2 /
UPS
UPS
Figure 2-11 Integrated Virtual Ethernet feature codes and assemblies
Any IVE feature code located in the first enclosure of a 570 also includes the System VPD
(Vital Product Data) Chip and system (serial) port (1 or 2 depending on the feature code).
The IVE feature code is installed by manufacturing. Similar to other integrated ports, the
feature does not have hot-swappable or hot-pluggable capability and must be serviced by a
trained IBM System Service Representative.
Figure 2-12 on page 41 shows the rear view of a basic 570 in a state of disassembly with
some necessary components and covers removed to highlight the connection of the feature
code assembly into the server enclosure I/O subsystem system board.
2
10 Gbps SR (short range) is designed to support short distances over deployed multi-mode fiber cabling, it has a
range of between 26 m and 82 m depending on cable type. It also supports 300 m operation over new, 50 µm
2000 MHz·km multi-mode fiber (using 850 nm).
40IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
Figure 2-12 Integrated Virtual Ethernet adapter connection on System p 570 I/O system board
2.7.2 Feature code port and cable support
All the IVE feature codes have different connectivity options and different cable support (see
Figure 2-13).
FC 5636FC 5639FC 5637
2 Ethernet ports RJ45
10/100/1000 Mbps
2 system ports
Figure 2-13 IVE physical port connectors according to IVE feature codes
4 Ethernet ports RJ45
10/100/1000 Mbps
1 system port
2 Ethernet ports SR
10 Gbps
1 system port
FC 5636 and FC 5639 supports:
1 Gbps connectivity
10 Mbps and 100 Mbps connectivity
RJ-45 connector
Use the Ethernet cables that meet Cat 5e
3
cabling standards, or higher, for best performance.
Chapter 2. Architecture and technical overview 41
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
FC 5637 supports:
Only 10 Gbps SR connectivity
62.5 micron multi-mode fiber cable type
– LC physical connector type
– 33 meters maximum range
2.7.3 IVE subsystem
Figure 2-14 shows a high level-logical diagram of the IVE.
P5IOC2
System
memory
Figure 2-14 IVE system placement
Every POWER6 processor-based server I/O subsystem contains the P5IOC2 chip. It is a
dedicated controller that acts as the primary bridge for all PCI buses and all internal I/O
devices. IVE major hardware components reside inside the P5IOC2 chip.
The IVE design provides a great improvement of latency for short packets. Messaging
applications such as distributed databases require low latency communication for
synchronization and short transactions. The methods used to achieve low latency include:
GX+ bus attachment
Immediate data in descriptors (reduce memory access)
Direct user space per-connection queueing (OS bypass)
Designed for up to 3-times throughput improvement over current 10 Gbps solutions
POWER6 chip
GX interface
IVE
2 x 10 Gbps or
4 x 1 Gbps Ethernet
Provide additional acceleration functions in order to reduce host code path length. These
include header / data split to help with zero-copy stacks
Provide I/O virtualization support so that all partitions of the system can natively take
advantage of the above features
Allows one 10 Gbps port to replace up to 10 dedicated PCI 1 Gbps adapters in a
partitioned system
One of the key design goals of the IVE is the capability to integrate up to two 10 Gbps
Ethernet ports or four 1 Gbps Ethernet ports into the P5IOC2 chip, with the effect of a low
cost Ethernet solution for low-end and mid-range server platforms. Any 10 Gbps, 1 Gbps,
100 Mbps or 10 Mbps speeds share the same I/O pins and do not require additional hardware
3
Category 5 cable, commonly known as Cat 5, is a twisted pair cable type designed for high signal integrity.
Category 5 has been superseded by the Category 5e specification.
42IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
or feature on top of the IVE card assembly itself. Another key goal is the support of all the
state-of-art NIC functionality provided by leading Ethernet NIC vendors.
IVE offers the following functions with respect to virtualization:
Up to 32 logical ports identified by MAC address
Sixteen MAC addresses are assigned to each IVE port group.
Each logical port can be owned by a separate LPAR
Direct data path to LPAR
Function enablement per LPAR
Default send and receive queues per LPAR
Ethernet MIB and RMON counters per LPAR
VLAN filtering per logical port (4096 VLANs * 32 Logical Ports)
Internal layer 2 switch for LPAR to LPAR data traffic
Multicast / Broadcast redirection to Multicast / Broadcast manager
IVE relies exclusively on the system memory and CEC processing cores to implement
acceleration features. There is not a requirement of dedicated memory, thus reducing the cost
of this solution and providing maximum flexibility. IVE Ethernet MACs and acceleration
features consume less than 8 mm
2
of logic in CMOS 9SF technology.
IVE does not have flash memory for its open firmware but it is stored in the Service Processor
flash and then passed to POWER Hypervisor (PHYP) control. Therefore flash code update is
done by PHYP.
2.8 PCI adapters
Peripheral Component Interconnect Express PCIe uses a serial interface and allows for
point-to-point interconnections between devices using directly wired interface between these
connection points. A single PCIe serial link is a dual-simplex connection using two pairs of
wires, one pair for transmit and one pair for receive, and can only transmit one bit per cycle. It
can transmit at the extremely high speed of 2.5 Gbps, which equates to a burst mode of 320
MBps on a single connection. These two pairs of wires is called a lane. A PCIe link may be
comprised of multiple lanes. In such configurations, the connection is labeled as x1, x2, x8,
x12, x16 or x32, where the number is effectively the number of lanes.
IBM offers PCIe adapter options for the 570, as well as PCI and PCI-extended (PCI-X)
adapters. All adapters support Extended Error Handling (EEH). PCIe adapters use a different
type of slot than PCI and PCI-X adapters. If you attempt to force an adapter into the wrong
type of slot, you may damage the adapter or the slot. A PCI adapter can be installed in a
PCI-X slot, and a PCI-X adapter can be installed in a PCI adapter slot. A PCIe adapter cannot
be installed in a PCI or PCI-X adapter slot, and a PCI or PCI-X adapter cannot be installed in
a PCIe slot. For a full list of the adapters that are supported on the systems and for important
information regarding adapter placement, see PCI Adapter Placement, SA76-0090. You can
find this publication at :
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
Before adding or rearranging adapters, use the System Planning Toll to validate the new
adapter configuration. See the System Planning Tool Web site at :
If you are installing a new feature, ensure that you have the software required to support the
new feature and determine whether there are any existing PTF prerequisites to install. To do
this, use the IBM Prerequisite Web site at :
http://www-912.ibm.com/e_dir/eServerPrereq.nsf
2.8.1 LAN adapters
To connect a 570 local area network (LAN), you can use Integrated 10/100/1000 dual-port
Virtual Ethernet with optional 10/100/1000 quad-port or dual-port 10 Gb Virtual Ethernet.
Table 2-6 lists the additional LAN adapters that are available.
Table 2-6 Available LAN adapters
Feature codeAdapter descriptionSlotSize
5700Gigabit Ethernet-SXPCI-XShort
570110/100/1000 Base-TX EthernetPCI-XShort
57062-port 10/100/1000 Base-TX PCI-XShort
57072-port Gigabit EthernetPCI-XShort
57174-port 1 Gb Ethernet PCIe 4xPCIeShort
a
5718
a
5719
572110 Gb Ethernet - Short ReachPCI-XShort
572210 GB Ethernet - Long ReachPCI-X Short
57404-port 10/100/1000 EthernetPCI-XShort
57672-port 1Gb Ethernet (UTP)PCIeShort
57682-port 1Gb Ethernet (Fiber)PCIeShort
a. Supported, but not available for a new configuration
10 Gigabit Ethernet-SRPCI-XShort
IBM 10 Gigabit Ethernet-SRPCI-XShort
2.8.2 SCSI and SAS adapters
To connect to external SCSI or SAS devices, the adapters that are provided in Table 2-7 are
available to be configured.
Table 2-7 Available SCSI adapters
Feature codeAdapter descriptionSlotSize
a
5712
Dual Channel Ultra320 SCSIPCI-XShort
5736DDR Dual Channel Ultra320 SCSIPCI-XShort
5900PCI-X DDR Dual -x4 SAS AdapterPCI-XShort
a. Supported, but not available for a new configuration
44IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
Table 2-8 on page 45 shows comparing Parallel SCSI to SAS.
Table 2-8 Comparing Parallel SCSI to SAS
Parallel SCSISAS
ArchitectureParallel, all devices connected
to shared bus
Performance320 Mb/s (Ultra320 SCSI),
performance degradeds as
devices added to shared bus
Scalability15 drivesOver 16,000 drives
CompatibilityIncompatible with all other drive
interfaces
Max. Cable Length12 meters total (must sum
lengths of all cables used on
bus)
Cable From FactorMultitude of conductors adds
bulk, cost
Hot PluggabilityNoYes
Device IdentificationManually set, user must ensure
no ID number conflicts on bus
TerminationManually set, user must ensure
proper installation and
functionality of terminators
Serial, point-to-point, discrete
signal paths
3 Gb/s, roadmap to 12 Gb/s,
performance maintained as
more devices added
Compatible with Serial ASA
(SATA)
8 meters per discrete
connection, total domain
cabling hundreds of meters
Compact connectors and
cabling save space, cost
Worldwide unique ID set at time
of manufacture
Discrete signal paths enable
device to include termination by
default
2.8.3 iSCSI
iSCSI is an open, standards-based approach by which SCSI information is encapsulated
using the TCP/IP protocol to allow its transport over IP networks. It allows transfer of data
between storage and servers in block I/O formats (that is defined by iSCSI protocol) and thus
enables the creation of IP SANs. iSCSI allows an existing network to transfer SCSI
commands and data with full location independence and defines the rules and processes to
accomplish the communication. The iSCSI protocol is defined in iSCSI IETF draft-20. For
more information about this standard, see:
http://tools.ietf.org/html/rfc3720
Although iSCSI can be, by design, supported over any physical media that supports TCP/IP
as a transport, today's implementations are only on Gigabit Ethernet. At the physical and link
level layers, iSCSI supports Gigabit Ethernet and its frames so that systems supporting iSCSI
can be directly connected to standard Gigabit Ethernet switches and IP routers. iSCSI also
enables the access to block-level storage that resides on Fibre Channel SANs over an IP
network using iSCSI-to-Fibre Channel gateways such as storage routers and switches. The
iSCSI protocol is implemented on top of the physical and data-link layers and presents to the
operating system a standard SCSI Access Method command set. It supports SCSI-3
commands and reliable delivery over IP networks. The iSCSI protocol runs on the host
initiator and the receiving target device. It can either be optimized in hardware for better
performance on an iSCSI host bus adapter (such as FC 5713 and FC 5714 supported in IBM
System p servers) or run in software over a standard Gigabit Ethernet network interface card.
IBM System p systems support iSCSI in the following two modes:
HardwareUsing iSCSI adapters (see “IBM iSCSI adapters” on page 46).
Chapter 2. Architecture and technical overview 45
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
Software Supported on standard Gigabit adapters, additional software (see
“IBM iSCSI software Host Support Kit” on page 46) must be installed.
The main processor is utilized for processing related to the iSCSI
protocol.
Initial iSCSI implementations are targeted at small to medium-sized businesses and
departments or branch offices of larger enterprises that have not deployed Fibre Channel
SANs. iSCSI is an affordable way to create IP SANs from a number of local or remote storage
devices. If Fibre Channel is present, which is typical in a data center, it can be accessed by
the iSCSI SANs (and vice versa) via iSCSI-to-Fibre Channel storage routers and switches.
iSCSI solutions always involve the following software and hardware components:
InitiatorsThese are the device drivers and adapters that reside on the client.
They encapsulate SCSI commands and route them over the IP
network to the target device.
Target sThe target software receives the encapsulated SCSI commands over
the IP network. The software can also provide configuration support
and storage-management support. The underlying target hardware
can be a storage appliance that contains embedded storage, and it
can also be a gateway or bridge product that contains no internal
storage of its own.
IBM iSCSI adapters
iSCSI adapters in IBM System p systems provide the advantage of increased bandwidth
through the hardware support of the iSCSI protocol. The 1 Gigabit iSCSI TOE (TCP/IP
Offload Engine) PCI-X adapters support hardware encapsulation of SCSI commands and
data into TCP and transports them over the Ethernet using IP packets. The adapter operates
as an iSCSI TOE. This offload function eliminates host protocol processing and reduces CPU
interrupts. The adapter uses a Small form factor LC type fiber optic connector or a copper
RJ45 connector.
Table 2-9 provides the orderable iSCSI adapters.
Table 2-9 Available iSCSI adapters
Feature codeDescriptionSlotSize
5713Gigabit iSCSI TOE on PCI-X on copper media adapterPCI-XShort
5714Gigabit iSCSI TOE on PCI-X on optical media adapter PCI-XShort
IBM iSCSI software Host Support Kit
The iSCSI protocol can also be used over standard Gigabit Ethernet adapters. To utilize this
approach, download the appropriate iSCSI Host Support Kit for your operating system from
the IBM NAS support Web site at:
http://www.ibm.com/storage/support/nas/
The iSCSI Host Support Kit on AIX and Linux acts as a software iSCSI initiator and allows
you to access iSCSI target storage devices using standard Gigabit Ethernet network
adapters. To ensure the best performance, enable the TCP Large Send, TCP send and
receive flow control, and Jumbo Frame features of the Gigabit Ethernet Adapter and the
iSCSI Target. Tune network options and interface parameters for maximum iSCSI I/O
throughput on the operating system.
46IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
2.8.4 Fibre Channel adapter
The 570 servers support direct or SAN connection to devices using Fibre Channel adapters.
Table 2-10 provides a summary of the available Fibre Channel adapters.
All of these adapters have LC connectors. If you are attaching a device or switch with an SC
type fibre connector, then an LC-SC 50 Micron Fiber Converter Cable (FC 2456) or an LC-SC
62.5 Micron Fiber Converter Cable (FC 2459) is required.
Supported data rates between the server and the attached device or switch are as follows:
Distances of up to 500 meters running at 1 Gbps, distances up to 300 meters running at
2 Gbps data rate, and distances up to 150 meters running at 4 Gbps. When these adapters
are used with IBM supported Fibre Channel storage switches supporting long-wave optics,
distances of up to 10 kilometers are capable running at 1 Gbps, 2 Gbps, and 4 Gbps data
rates.
Table 2-10 Available Fibre Channel adapters
Feature codeDescriptionSlotSize
a
5716
5758DDR 4 Gb single port Fibre Channel PCI-XShort
2 Gigabit Fibre Channel PCI-X AdapterPCI-XShort
5759DDR 4 Gb dual port Fibre Channel PCI-XShort
57731-port 4 Gb Fibre ChannelPCIeShort
57742-port 4 Gb Fibre ChannelPCIeShort
a. Supported, but not available for a new configuration
2.8.5 Graphic accelerators
The 570 support up to four graphics adapters. Table 2-11 provides the available graphic
accelerators. They can be configured to operate in either 8-bit or 24-bit color modes. These
adapters support both analog and digital monitors.
Table 2-11 Available Graphic accelerators
Feature codeDescriptionSlotSize
2849GXT135P Graphics AcceleratorPCI-XShort
5748GXT145 Graphics AcceleratorPCIeShort
Note: Both adapters are not hot-pluggable.
2.8.6 Asynchronous PCI adapters
Asynchronous PCI-X adapters provide connection of asynchronous EIA-232 or RS-422
devices. If you have a cluster configuration or high-availability configuration and plan to
connect the IBM System p servers using a serial connection, the use of the two system ports
is not supported. You should use one of the features listed in Table 2-12 on page 48.
Chapter 2. Architecture and technical overview 47
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
In many cases, the FC 5723 asynchronous adapter is configured to supply a backup HACMP
heartbeat. In these cases, a serial cable (FC 3927 or FC 3928) must be also configured. Both
of these serial cables and the FC 5723 adapter have 9-pin connectors.
2.8.7 Additional support for existing PCI adapters
The lists of the major PCI adapters that you can configure in a 570 when you build an
available configuration are described in 2.8.1, “LAN adapters” on page 44 through 2.8.6,
“Asynchronous PCI adapters” on page 47. The list of all the supported PCI adapters, with the
related support for additional external devices, is more extended.
If you would like to use PCI adapters you already own, contact your IBM service
representative to verify whether those adapters are supported.
2.9 Internal storage
The 570 internal disk subsystem is driven by the latest DASD interface technology Serial
Attached SCSI (SAS). This interface provides enhancements over parallel SCSI with its point
to point high frequency connections. The SAS controller has eight SAS ports, four of them are
used to connect to the DASD drives and one to a media device.
The DASD backplane implements two SAS port expanders that take four SAS ports from the
SAS controller and expands it to 12 SAS ports. These 12 ports allow for redundant SAS ports
to each of the six DASD devices.
The DASD backplane provides the following functions :
supports six 3.5 inches SAS DASD devices
contains two SAS port expanders for redundant SAS paths to the SAS devices
SAS passthru connection to medias backplane
2.9.1 Integrated RAID options
The 570 supports a 6-pack DASD backplane attached to the system planar. To support RAID
functionality a combination of additional adapters is required. At the time of writing RAID level
0 and 10 are supported using adapter FC 5900 or 5909 and FC 3650 or FC 3651 as
described in Table 2-13.
Table 2-13 Raid support
FeatureDescription
3650 + 5900 or 5909A hardware feature that occupies PCIe slot P1C3 provides a mini
48IBM Power 570 Technical Overview and Introduction
SAS 4x connector to the rear bulkhead and allows three of the
internal SAS drives ( disk 4,5,6) to be controlled using an external
SAS controller.
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
FeatureDescription
3651 + 5900 or 5909A hardware feature that occupies PCIe slot P1C3 provides a mini
SAS 4x connector to the rear bulkhead and allows all of the internal
SAS drives to be controlled using an external SAS controller.
2.9.2 Split backplane
As described in paragraph 2.9.1, “Integrated RAID options” on page 48 the same features are
required to split the 6-pack DASD backplane in two groups of three disks. Using feature 3650
and 5900 disk 4,5, and, 6 are managed using the external SAS controller. Disk 1, 2, and 3 are
managed by the internal SAS controller.
2.9.3 Internal media devices
Inside each CEC drawer in the 570 there is an optional media backplane with one media bay.
The internal IDE media bay in separate CEC drawers can be allocated or assigned to a
different partition. The media backplane inside each CEC drawer cannot be split between two
logical partitions.
2.9.4 Internal hot-swappable SAS drives
The 570 can have up to six hot-swappable disk drives plugged in the physical 6-pack disk
drive backplane. The hot-swap process is controlled by the virtual SAS Enclosure Services
(VSES), which is located in the logical 6-pack disk drive backplane. The 6-pack disk drive
backplanes can accommodate the devices listed in Table 2-14
Table 2-14 Hot-swappable disk options
Feature codeDescription
364673.4 GB 15,000 RPM SAS hot-swappable disk drive
3647146.8 GB 15,000 RPM SAS hot-swappable disk drive
3548300 GB 15,000 RPM SAS hot-swappable disk drive
Prior to the hot-swap of a disk drive in the hot-swappable-capable bay, all necessary
operating system actions must be undertaken to ensure that the disk is capable of being
deconfigured. After the disk drive has been deconfigured, the SAS enclosure device will
power-off the slot, enabling safe removal of the disk. You should ensure that the appropriate
planning has been given to any operating-system-related disk layout, such as the AIX Logical
Volume Manager, when using disk hot-swap capabilities. For more information, see Problem Solving and Troubleshooting in AIX 5L, SG24-5496.
Note: We recommend that you follow this procedure, after the disk has been deconfigured,
when removing a hot-swappable disk drive:
1. Release the tray handle on the disk assembly.
2. Pull out the disk assembly a little bit from the original position.
3. Wait up to 20 seconds until the internal disk stops spinning.
Now you can safely remove the disk from the DASD backplane.
Chapter 2. Architecture and technical overview 49
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
2.10 External I/O subsystems
This section describes the external I/O subsystems, which include the I/O drawers, the IBM
System Storage™ EXP 12S SAS drawer, as well as the 7311-D11, 7311-D20, 7311-G30,
7031-D24, and 7031-T24 deskside tower.
Table 2-15 provided an overview of all the supported I/O drawers.
Table 2-15 I/O drawer capabilities
DrawerDASDPCI SlotsRequirements for a 570
7311-D116 x PCI-XGX+ adapter card FC 1800
7311-D2012 x SCSI disk drive bays7 x PCI-XGX+ adapter card FC 1800
7314-G306 x PCI-X
DDR
266 MHz
7031-T24/D2424 x SCSI disk drive baysAny supported SCSI adapter
FC 588612 x SAS disk drive baysAny supported SAS adapter
GX+ adapter card FC 1802
Each POWER6 chip provides a GX+ bus which is used to connect to an I/O subsystem or
Fabric Interface card. In a fully populated 570 enclosure there are two GX+ buses, one from
each POWER6 chip. Each 570 enclosure has 2 GX+ slots with a single GX+ bus. The second
GX+ slot is not active unless the second CPU card is installed. If the second CPU card is
installed, then the second GX+ slot and associated bus is active and available.
The maximum number of attached remote I/O drawers depends on the number of system unit
enclosures in the system and the I/O attachment type. Each GX+ bus can be populated with a
GX+ adapter card that adds more RIO-G ports to connect external I/O drawers.
The GX+ adapter cards listed in Table 2-16 are supported at the time of writing.
Table 2-16 GX+ adapter card supported
Feature
code
FC 1800GX dual port RIO-2 attachCard provides two ports that support up to
GX+ adapter card descriptionGX+ adapter card I/O drawer support
four of the following I/O drawers:
7311-D10
7311-D11
7311-D20
FC 1802GX dual port 12X channel attachCard provides two 12X connections that
2.10.1 7311 Model D11 I/O drawers
The 7311-D11 provides six PCI-X slots supporting an enhanced blind-swap mechanism.
Drawers must have a RIO-2 adapter to connect to the server.
Each primary PCI-X bus is connected to a PCI-X-to-PCI-X bridge, which provides three slots
with Extended Error Handling (EEH) for error recovering. In the 7311 Model D11 I/O drawer,
50IBM Power 570 Technical Overview and Introduction
support up to four of the following I/O drawer:
7314-G30
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
slots 1 to 6 are PCI-X slots that operate at 133 MHz and 3.3 V signaling. Figure 2-15 on
page 51 shows a conceptual diagram of the 7311 Model D11 I/O drawer.
PCI-X Host
Bridge
RIO 2
133 MHz,
64-bit PCI-X
PCI-X Bridge
6
4
/
1
3
3
1
133 MHz, 64-bit PCI-X
6
6
4
4
/
/
1
1
3
3
3
3
2
3
PCI-X Bridge
6
6
4
4
/
/
1
1
3
3
3
3
4
56
6
4
/
1
3
3
Figure 2-15 Conceptual diagram of the 7311-D11 I/O drawer
7311 Model D11 features
This I/O drawer model provides the following features:
Six hot-plug 64-bit, 133 MHz, 3.3 V PCI-X slots, full length, enhanced blind-swap cassette
Default redundant hot-plug power and cooling
Two default remote (RIO-2) ports and two SPCN ports
7311 Model D11 rules and maximum support
Table 2-17 describes the maximum number of I/O drawer supported
Table 2-17 Maximum number of 7311 Model D11 I/O drawers supported
2.10.2 Consideration for 7311 Model D10 I/O drawer
It is not possible to configure the 7311 Model D10 I/O drawer in a 570 initial order. Clients who
decided to migrate p5 570 to 570 can decide to re-use 7311 Model D10 I/O drawers originally
connected to the p5 570. It requires the same connection of the 7311 Model D11 and can be
intermixed in the same loop
Chapter 2. Architecture and technical overview 51
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
2.10.3 7311 Model D20 I/O drawer
The 7311 Model D20 I/O drawer must have the RIO-2 loop adapter (FC 6417) to be
connected to the 570 system. The PCI-X host bridge inside the I/O drawer provides two
primary 64-bit PCI-X buses running at 133 MHz. Therefore, a maximum bandwidth of 1 GBps
is provided by each of the buses.
Figure 2-16 shows a conceptual diagram of the 7311 Model D20 I/O drawer subsystem.
PCI-X Host
Bridge
RIO 2
133 MHz,
64-bit PCI-X
PCI-X Bridge
6
6
4
4
/
/
1
1
3
3
3
3
1
2
133 M H z, 64 -bit PC I-X
PCI-X Bridge
6
6
4
/
1
3
3
3
6
4
4
/
/
1
1
3
3
3
3
4
5
6
4
/
1
3
3
6
6
4
/
1
3
3
7
Figure 2-16 Conceptual diagram of the 7311-D20 I/O drawer
7311 Model D20 internal SCSI cabling
A 7311 Model D20 supports hot-swappable SCSI Ultra320 disk drives using two 6-pack disk
bays for a total of 12 disks. Additionally, the SCSI cables (FC 4257) are used to connect a
SCSI adapter (any of various features) in slot 7 to each of the 6-packs, or two SCSI adapters,
one in slot 4 and one in slot 7. (See Figure 2-17.)
Connect the SCSI cable feature to the SCSI adapter
in rightmost slot (7) as shown below:
to 6-pack
backplanes
If a SCSI card is also placed in slot 4, wire as shown below:
to 6-pack
backplanes
Figure 2-17 7311 Model D20 internal SCSI cabling
Note: Any 6-packs and the related SCSI adapter can be assigned to a logical partition (for
example, the two partitions can be two Virtual I/O server partitions). If one SCSI adapter is
connected to both 6-packs, then both 6-packs can be assigned only to the same partition.
52IBM Power 570 Technical Overview and Introduction
SCSI cables FC 4257SCSI cables FC 4257
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
2.10.4 7311 Model D11 and Model D20 I/O drawers and RIO-2 cabling
As described in 2.10, “External I/O subsystems” on page 50, we can connect up to four I/O
drawers in the same loop, and up to 20 I/O drawers to the p5-570 system.
Each RIO-2 port can operate at 1 GHz in bidirectional mode and is capable of passing data in
each direction on each cycle of the port. Therefore, the maximum data rate is 4 GBps per I/O
drawer in double barrel mode (using two ports).
There is one default primary RIO-2 loop in any 570 building block. This feature provides two
Remote I/O ports for attaching up to four 7311 Model D11 or 7311 Model D20 I/O drawers or
7311 Model D10 to the system in a single loop. Different I/O drawer models can be used in
the same loop, but the combination of I/O drawers must be a total of four per single loop. The
optional RIO-2 expansion card may be used to increase the number of I/O drawers that can
be connected to one 570 building block, and the same rules of the default RIO-2 loop must be
considered. The method that is used to connect the drawers to the RIO-2 loop is important for
performance.
Figure 2-18 shows how you could connect four I/O drawers to one p5-570 building block. This
is a logical view; actual cables should be wired according to the installation instructions.
Cost OptimizedPerformance Optimized
PCI-X slots
I/O drawer #1
I/O drawer #2
I/O drawer #3
I/O drawer #4
Figure 2-18 RIO-2 cabling examples
I/O drawer #3
I/O drawer #4
I/O drawer #1
I/O drawer #2
Note: If you have 20 I/O drawers, although there are no restrictions on their placement, this
can affect performance.
RIO-2 cables are available in different lengths to satisfy different connection requirements:
Remote I/O cable, 1.2 m (FC 3146, for between D11 drawers only)
Remote I/O cable, 1.75 m (FC 3156)
Remote I/O cable, 2.5 m (FC 3168)
Remote I/O cable, 3.5 m (FC 3147)
Remote I/O cable, 10 m (FC 3148)
Chapter 2. Architecture and technical overview 53
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
2.10.5 7311 I/O drawer and SPCN cabling
SPCN4 is used to control and monitor the status of power and cooling within the I/O drawer.
SPCN is a loop: Cabling starts from SPCN port 0 on the 570 to SPCN port 0 on the first I/O
drawer. The loop is closed, connecting the SPCN port 1 of the I/O drawer back to port 1 of the
570 system. If you have more than one I/O drawer, you continue the loop, connecting the next
drawer (or drawers) with the same rule.
SPCN cabling examples
SPCN cabling examples
Primary drawer
Primary drawer
SPCN port0
SPCN port 0
SPCN port1
SPCN port 1
I/O drawer or secondary drawer
I/O drawer or secondary drawer
SPCN port0
SPCN port 0
SPCN port1
SPCN port 1
Primary drawer
Primary drawer
SPCN port0
SPCN port 0
SPCN port1
SPCN port 1
I/O drawer or secondarydrawer
I/O drawer or secondary drawer
SPCN port0
SPCN port 0
SPCN port1
SPCN port 1
I/O drawer or secondarydrawer
I/O drawer or secondary drawer
SPCN port0
SPCN port 0
SPCN port1
SPCN port 1
Figure 2-19 SPCN cabling examples
There are different SPCN cables to satisfy different length requirements:
SPCN cable drawer-to-drawer, 2 m (FC 6001)
SPCN cable drawer-to-drawer, 3 m (FC 6006)
SPCN cable rack-to-rack, 6 m (FC 6008)
SPCN cable rack-to-rack, 15 m (FC 6007)
SPCN cable rack-to-rack, 30 m (FC 6029)
2.10.6 7314 Model G30 I/O drawer
The 7314-G30 expansion unit is a rack-mountable, I/O expansion drawer that is designed to
be attached to the system unit using the InfiniBand® bus and InfiniBand cables. The
7314-G30 can accommodate 6 blind swap adapter cassettes. Cassettes can be installed and
removed without removing the drawer from the rack. The Figure 2-20 on page 55 shows the
back view of the expansion unit.
4
System Power Control Network
54IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
Figure 2-20 7311 Model G40 rear side
7314 Model G30 rules and maximum support
Table 2-18 describes the maximum number of I/O drawer supported
Table 2-18 Maximum number of 7314 Model G30 I/O drawers supported
Similar to the 7311 Model D10 and D11, up to two 7314 Model G30 drawers can be installed
in a unit enclosure (FC 7314). The unit enclosure requires to be installed in a 19" rack such as
the 7014-T00 or 7014-T42. The actual installation location in the rack will vary depending on
other rack content specify codes ordered with rack.
2.11 External disk subsystems
The 570 has internal hot-swappable drives. When the AIX operating system is installed in a
IBM System p server, the internal disks are usually used for the AIX rootvg volume group and
Chapter 2. Architecture and technical overview 55
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
paging space. Specific client requirements can be satisfied with the several external disk
possibilities that the 570 supports.
2.11.1 IBM System Storage EXP 12S (FC 5886)
The IBM System Storage EXP 12S is a high density rack-mountable disk drive enclosure for
supporting a total of twelve 3.5-inch disk drives on POWER6 systems only. Using
hot-swappable 300 GB SAS disk drives the EXP 12S drawer can provide 3.6 TB of disk
capacity. The expansion drawer provides redundant power device, cooling, and SAS
expanders all devices are hot-swappable. The SAS disks are front accessible, using the same
disk carrier and 3.5 inch SAS disk drives as used in IBM POWER6 Power Systems. As a two
unit, 19-inch rack-mountable disk enclosure supports 73 GB, 146 GB, and 300 GB
hot-swappable SAS disk drives.
The IBM System Storage EXP 12S drawer offers :
Modular SAS disk expansion drawer
Up to 12 3.5-inch SAS disk drives
Variety of supported connection options, from single attachment to a HACMP solution
Redundant hot-plug power and cooling with dual line cords
Redundant and hot-swappable SAS expanders
IBM System Storage EXP 12S drawer physical description
The EXP 12S drawer must be mounted in a 19-inch rack, such as the IBM 7014-T00 or
7014-T42. The EXP 12S drawer has the following attributes:
One drawer EXP 12S
– Width: 481.76 mm (18.97 in)
– Depth: 511.00 mm (20.12 in)
– Height: 87.36 mm ( 3.38 in)
– Weight: 18 kb (39.70 lp)
Connecting a EXP 12S drawer to a POWER6 system
To connect a EXP 12S SAS drawer to a POWER6 system an additional adapter is needed.
Table 2-19 provides the current list of available adapters.
Table 2-19 SAS adapters
FeatureDescription
5900PCI-X DDR dual SAS adapter
Depending on the required configuration a different set of cables are needed to connect the
EXP 12S drawer to the system or drawer. A list of cables are provided in Table 2-20.
Table 2-20 SAS connection cables
FeatureDescription
3652SAS cable (EE) drawer to drawer 1 meter
3652SAS cable (EE) drawer to drawer 3 meter
3654SAS cable (EE) drawer to drawer 3 meter
56IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
FeatureDescription
3691SAS cable (YO) adapter to SAS enclosure 1.5 meter
3692SAS cable (YO) adapter to SAS enclosure 3 meter
3693SAS cable (YO) adapter to SAS enclosure 6 meter
A typical base configuration is using a server machine and a single attached drawer as shown
in Figure 2-21.
2x
YO Cable
4x
#5900
#5900
4x
4x
4x
4x
ESMESMESMESM
2x
Figure 2-21 Base configuration of one SAS drawer
A maximum configuration using four EXP 12S drawers on one adapter feature is shown in
Figure 2-22.
YO Cable
2x
ESMESMESMESM
2x
4x
4x
#5900
#5900
4x
4x
4x
4x
YO Cable
2x
ESMESMESMESM
2x
Figure 2-22 Maximum attachment of EXP 12S on one adapter
EE Cable
2x
ESMESMESMESM
2x
EE Cable
2x
ESMESMESMESM
2x
2.11.2 IBM TotalStorage EXP24 Expandable Storage
The IBM TotalStorage® EXP24 Expandable Storage disk enclosure, Model D24 or T24, can
be purchased together with the 570 and will provide low-cost Ultra320 (LVD) SCSI disk
storage. This disk storage enclosure device provides more than 7 TB of disk storage in a 4 U
rack-mount (Model D24) or compact deskside (Model T24) unit. Whether high availability
Chapter 2. Architecture and technical overview 57
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
storage solutions or simply high capacity storage for a single server installation, the unit
provides a cost-effective solution. It provides 24 hot-swappable disk bays, 12 accessible from
the front and 12 from the rear. Disk options that can be accommodated in any of the four
six-packs disk drive enclosure are 73.4 GB, 146.8 GB, or 300 GB 10 K rpm or 36.4 GB,
73.4 GB, or 146.8 GB 15 K rpm drives. Each of the four six-packs disk drive enclosure might
be attached independently to an Ultra320 SCSI or Ultra320 SCSI RAID adapter. For high
available configurations, a dual bus repeater card (FC 5742) allows each six-pack to be
attached to two SCSI adapters, installed in one or multiple servers or logical partitions.
Optionally, the two front or two rear six-packs might be connected together to form a single
Ultra320 SCSI bus of 12 drives.
2.11.3 IBM System Storage N3000, N5000 and N7000
The IBM System Storage N3000 and N5000 line of iSCSI enabled storage offerings provide a
flexible way to implement a Storage Area Network over an Ethernet network. Flexible-Fibre
Channel and SATA disk drive capabilities allow for deployment in multiple solution
environments, including data compliant retention, nearline storage, disk-to-disk backup
scenarios, and high-performance mission-critical I/O intensive operations.
Newest member of the IBM System storage N series family are the N7000 systems. The
N7000 series is designed to deliver midrange to high-end enterprise storage and data
management capabilities.
See the following link for more information:
http://www.ibm.com/servers/storage/nas
2.11.4 IBM TotalStorage Storage DS4000 Series
The IBM System Storage DS4000™ line of Fibre Channel enabled Storage offerings provides
a wide range of storage solutions for your Storage Area Network. The IBM TotalStorage
DS4000 Storage server family consists of the following models: DS4100, DS4300, DS4500,
and DS4800. The Model DS4100 Express Model is the smallest model and scales up to
44.8 TB; the Model DS4800 is the largest and scales up to 89.6 TB of disk storage at the time
of this writing. Model DS4300 provides up to 16 bootable partitions, or 64 bootable partitions
if the turbo option is selected, that are attached with the Gigabit Fibre Channel Adapter
(FC 1977). Model DS4500 provides up to 64 bootable partitions. Model DS4800 provides
4 GB switched interfaces. In most cases, both the IBM TotalStorage DS4000 family and the
IBM System p5 servers are connected to a storage area network (SAN). If only space for the
rootvg is needed, the Model DS4100 is a good solution.
For support of additional features and for further information about the IBM TotalStorage
DS4000 Storage Server family, refer to the following Web site:
2.11.5 IBM System Storage DS6000 and DS8000 series
The IBM System Storage Models DS6000™ and DS8000™ are the high-end premier storage
solution for use in storage area networks and use POWER technology-based design to
provide fast and efficient serving of data. The IBM TotalStorage DS6000 provides enterprise
class capabilities in a space-efficient modular package. It scales to 57.6 TB of physical
storage capacity by adding storage expansion enclosures. The Model DS8000 series is the
flagship of the IBM DS family. The DS8000 scales to 1024 TB. However, the system
architecture is designed to scale to over one petabyte. The Model DS6000 and DS8000
systems can also be used to provide disk space for booting LPARs or partitions using
58IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
Micro-Partitioning technology. System Storage and the IBM Power servers are usually
connected together to a storage area network.
For further information about ESS, refer to the following Web site:
The Hardware Management Console (HMC) is a dedicated workstation that provides a
graphical user interface for configuring, operating, and performing basic system tasks for the
POWER6 processor-based (as well as the POWER5 and POWER5+ processor-based)
systems that function in either non-partitioned, partitioned, or clustered environments. In
addition the HMC is used to configure and manage partitions. One HMC is capable of
controlling multiple POWER5, POWER5+, and POWER6 processor-based systems.
At the time of writing, one HMC supports up to 48 POWER5, POWER5+ and POWER6
processor-based systems and up to 254 LPARs using the HMC machine code Version 7.3.
For updates of the machine code and HMC functions and hardware prerequisites, refer to the
following Web site:
POWER5, POWER5+ and POWER6 processor-based system HMCs require Ethernet
connectivity between the HMC and the server’s service processor, moreover if dynamic LPAR
operations are required, all AIX 5L, AIX V6, and Linux partitions must be enabled to
communicate over the network to the HMC. Ensure that at least two Ethernet ports are
available to enable public and private networks:
The HMC 7042 Model C06 is a deskside model with one integrated 10/100/1000 Mbps
Ethernet port, and two additional PCI slots.
The HMC 7042 Model CR4 is a 1U, 19-inch rack-mountable drawer that has two native
10/100/1000 Mbps Ethernet ports and two additional PCI slots.
Note: The IBM 2-Port 10/100/1000 Base-TX Ethernet PCI-X Adapter (FC 5706) should be
ordered to provide additional physical Ethernet connections.
For any logical partition in a server, it is possible to use a Shared Ethernet Adapter set in
Virtual I/O Server or Logical Ports of the Integrated Virtual Ethernet card, for a unique or
fewer connections from the HMC to partitions. Therefore, a partition does not require it’s own
physical adapter to communicate to an HMC.
It is a good practice to connect the HMC to the first HMC port on the server, which is labeled
as HMC Port 1, although other network configurations are possible. You can attach a second
HMC to HMC Port 2 of the server for redundancy (or vice versa). Figure 2-23 on page 60
shows a simple network configuration to enable the connection from HMC to server and to
enable Dynamic LPAR operations. For more details about HMC and the possible network
connections, refer to the Hardware Management Console V7 Handbook, SG24-7491.
Chapter 2. Architecture and technical overview 59
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
Figure 2-23 HMC to service processor and LPARs network connection
The default mechanism for allocation of the IP addresses for the service processor HMC
ports is dynamic. The HMC can be configured as a DHCP server, providing the IP address at
the time the managed server is powered on. If the service processor of the managed server
does not receive DHCP reply before time-out, predefined IP addresses will setup on both
ports. Static IP address allocation is also an option. You can configure the IP address of the
service processor ports with a static IP address by using the Advanced System Management
Interface (ASMI) menus.
Note: The service processor is used to monitor and manage the system hardware
resources and devices. The service processor offers the following connections:
Two Ethernet 10/100 Mbps ports
Both Ethernet ports are only visible to the service processor and can be used to attach
the server to an HMC or to access the Advanced System Management Interface
(ASMI) options from a client web browser, using the http server integrated into the
service processor internal operating system.
Both Ethernet ports have a default IP address
– Service processor Eth0 or HMC1 port is configured as 169.254.2.147 with netmask
255.255.255.0
– Service processor Eth1 or HMC2 port is configured as 169.254.3.147 with netmask
255.255.255.0
More information about the Service Processor can be found in 4.5.1, “Service processor”
on page 116.
Functions performed by the HMC include:
Creating and maintaining a multiple partition environment
Displaying a virtual operating system session terminal for each partition
Displaying a virtual operator panel of contents for each partition
60IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
Detecting, reporting, and storing changes in hardware conditions
Powering managed systems on and off
Acting as a service focal point
Generating or importing System Plans
The HMC provides both graphical and command line interface for all management tasks.
Remote connection to the HMC using a web browser (as of HMC Version 7, previous versions
required a special client program, called WebSM) or SSH are possible. The command line
interface is also available by using the SSH secure shell connection to the HMC. It can be
used by an external management system or a partition to perform HMC operations remotely.
2.12.1 High availability using the HMC
The HMC is an important hardware component. HACMP Version 5.4 high availability cluster
software can be used to execute dynamic logical partitioning operations or activate additional
resources (where available), thus becoming an integral part of the cluster.
If redundant HMC function is desired, the servers can be attached to two separate HMCs to
address availability requirements. All HMCs must have the same level of Hardware
Management Console Licensed Machine Code Version 7 (FC 0962) to manage POWER6
processor-based servers or an environment with a mixture of POWER5, POWER5+, and
POWER6 processor-based servers. The HMCs provide a locking mechanism so that only one
HMC at a time has write access to the service processor. Depending on your environment,
you have multiple options to configure the network. Figure 2-24 shows one possible high
available configuration.
LAN3 – Open network
HMC1HMC2
eth0eth1
LAN 1LAN 2
12
FSP
System ASystem B
LPAR A1
LPAR A2
LPAR A3
eth0eth1
12
FSP
LPAR B1
LPAR B2
LPAR B3
LAN1 –hardware management network for
first FSP ports (private)
LAN2 – hardware management network for
second FSP ports (private), separate
network hardware than LAN1
LAN3 - open network for HMC access and
dLPAR operations
Figure 2-24 Highly available HMC and network architecture
Note that only hardware management networks (LAN1 and LAN2) are highly available on the
above picture in order to keep simplicity. However, management network (LAN3) can be
made highly available by using a similar concept and adding more Ethernet adapters to
LPARs and HMCs.
Chapter 2. Architecture and technical overview 61
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
Redundant Service Processor connectivity
Redundant Service Processor function for managing the service processors when one fails is
supported on all systems that are operating with system firmware level FM320_xxx_xxx, or
later. This support is available for configurations with two or more CEC enclosures.
Redundant Service Processor function requires that the Hardware Management Console
(HMC) be attached to the Service Interface Card in both CEC enclosure 1 and CEC
enclosure 2. The Service Interface Card in these two enclosures must be connected using an
external Power Control cable (FC 6006 or similar). Figure 2-25 shows a redundant HMC and
redundant service processor connectivity configuration.
LAN3 – Open network
HMC1HMC2
eth0eth1
LAN 1LAN 2
12
FSP
CEC 1CEC 2
LPAR 1
LPAR 2
LPAR 3
Figure 2-25 Redundant HMC connection and Redundant Service Processor configuration
eth0eth1
12
FSP
LAN1 –hardware management network for
first FSP ports (private)
LAN2 – hardware management network for
second FSP ports (private), separate
network hardware than LAN1
LAN3 - open network for HMC access and
dLPAR operations
In a configuration with multiple systems or HMC’s, the customer is required to provide
switches or hubs to connect each HMC to the appropriate Service Interface Cards in each
system. One HMC should connect to the port labeled as HMC Port 1 on the first 2 CEC
drawers of each system, and a second HMC should be attached to HMC Port 2 on the first 2
CEC drawers of each system. This provides redundancy both for the HMCs and the service
processors.
For more details about redundant HMCs, refer to the Hardware Management Console V7 Handbook, SG24-7491.
2.12.2 Operating System Support
The POWER6-based IBM System 570 supports IBM AIX 5L Version 5.2, IBM AIX 5L Version
5.3, IBM AIX Version 6.1 and Linux distributions from SUSE and Red Hat.
Note: For specific technical support details, please refer to the support for IBM Web site:
http://www.ibm.com/systems/p/os
62IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
IBM AIX 5L
If installing AIX 5L on the 570, the following minimum requirements must be met:
AIX 5L for POWER V5.3 with 5300-07 Technology Level (APAR IY99422), CD#
LCD4-7463-09, DVD# LCD4-7544-05 or later
AIX 5L for POWER V5.3 with 5300-06 Technology Level with Service Pack 4 (APAR
IZ06992)
AIX 5L for POWER V5.2 with 5200-10 Technology Level (APAR IY94898), CD#
LCD4-1133-11
IBM periodically releases maintenance packages (service packs or technology levels) for the
AIX5L operating system. These packages can be ordered on CD-ROM or downloaded from:
The fixcentral Web site also provides information about how to obtain the CD-ROM.
You can also get individual operating system fixes and information about obtaining AIX 5L
service at this site. From AIX 5L V5.3 the Service Update Management Assistant, which
helps the administrator to automate the task of checking and downloading operating system
downloads, is part of the base operating system. For more information about the suma
command functionality, refer to:
AIX 5L is supported on the System p servers in partitions with dedicated processors (LPARs),
and shared-processor partitions (micro-partitions). When combined with one of the PowerVM
features, AIX 5L Version 5.3 can make use of all the existing and new virtualization features
such as micro-partitions, virtual I/O, virtual LAN, and PowerVM Live Partition Mobility, to
name a few.
IBM AIX V6.1
IBM is making available a new version of AIX, AIX V6.1 which will include significant new
capabilities for virtualization, security features, continuous availability features and
manageability. AIX V6.1 is the first generally available version of AIX V6.
Kernel exploitation of POWER6 Storage Keys for greater reliability
Robust journaled file system and Logical Volume Manager (LVM) software including
integrated file system snapshot
Tools for managing the systems environment -- System Management
Chapter 2. Architecture and technical overview 63
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
Interface Tool (SMIT) and the IBM Systems Director Console for AIX
Linux for System p systems
Linux is an open source operating system that runs on numerous platforms from embedded
systems to mainframe computers. It provides a UNIX-like implementation across many
computer architectures.
This section discusses two brands of Linux to be run in partitions. The supported versions of
Linux on System p servers are:
Novell SUSE Linux Enterprise Server V10 SP1 for POWER Systems or later
Red Hat Enterprise Linux Advanced Server V4.5 for POWER or later
Red Hat Enterprise Linux Advanced Server V5.1 for Power or later
The PowerVM features are supported in Version 2.6.9 and above of the Linux kernel. The
commercially available latest distributions from Red Hat, Inc. (RHEL AS 5) and Novell SUSE
Linux (SLES 10) support the IBM system p 64-bit architectures and are based on this 2.6
kernel series.
Clients wishing to configure Linux partitions in virtualized System p systems should consider
the following:
Not all devices and features supported by the AIX operating system are supported in
logical partitions running the Linux operating system.
Linux operating system licenses are ordered separately from the hardware. Clients can
acquire Linux operating system licenses from IBM, to be included with their System 570 or
from other Linux distributors.
For information about the features and external devices supported by Linux refer to:
For information about SUSE Linux Enterprise Server 10, refer to:
http://www.novell.com/products/server
For information about Red Hat Enterprise Linux Advanced Server 5, refer to:
http://www.redhat.com/rhel/features
Supported virtualization features
SLES 10, RHEL AS 4.5 and RHEL AS 5 support the following virtualization features:
Virtual SCSI, including for the boot device
Shared-processor partitions and virtual processors, capped and uncapped
Dedicated-processor partitions
Dynamic reconfiguration of processors
Virtual Ethernet, including connections through the Shared Ethernet Adapter in the Virtual
I/O Server to a physical Ethernet connection
Simultaneous multithreading (SMT)
SLES 10, RHEL AS 4.5, and RHEL AS 5 do not support the following:
Dynamic reconfiguration of memory
Dynamic reconfiguration of I/O slot
64IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
Note: IBM only supports the Linux systems of clients with a SupportLine contract covering
Linux. Otherwise, contact the Linux distributor for support.
i5/OS
At the time of writing i5/OS® is not supported.
2.13 Service information
The 570 is not a client setup server (CSU). Therefore, the IBM service representative
completes the system installation.
2.13.1 Touch point colors
Blue (IBM blue) or terra-cotta (orange) on a component indicates a touch point (for electronic
parts) where you can grip the hardware to remove it from or to install it into the system, to
open or to close a latch, and so on. IBM defines the touch point colors as follows:
Blue
This requires a shutdown of the system before the task can be
performed, for example, installing additional processors contained
in the second processor book.
Terra-cotta The system can remain powered on while this task is being
performed. Keep in mind that some tasks might require that other
steps have to be performed first. One example is deconfiguring a
physical volume in the operating system before removing a disk
from a 4-pack disk enclosure of the server.
Blue and terra-cotta
Important: It is important to adhere to the touch point colors on the system. Not doing so
can compromise your safety and damage the system.
2.13.2 Operator Panel
The service processor provides an interface to the control panel that is used to display server
status and diagnostic information. See Figure 2-26 on page 66 for operator control panel
physical details and buttons.
Terra-cotta takes precedence over this color combination, and the
rules for a terra-cotta-only touch point apply.
Chapter 2. Architecture and technical overview 65
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
Figure 2-26 Operator control panel physical details and button
Note: For servers managed by the HMC, use it to perform control panel functions.
Primary control panel functions
The primary control panel functions are defined as functions 01 to 20, including options to
view and manipulate IPL modes, server operating modes, IPL speed, and IPL type.
The following list describes the primary functions:
Function 01: Display the selected IPL type, system operating mode, and IPL speed
Function 02: Select the IPL type, IPL speed override, and system operating mode
Function 03: Start IPL
Function 04: Lamp Test
Function 05: Reserved
Function 06: Reserved
Function 07: SPCN functions
Function 08: Fast Power Off
Functions 09 to 10: Reserved
Functions 11 to 19: System Reference Code
Function 20: System type, model, feature code, and IPL type
All the functions mentioned are accessible using the Advanced System Management
Interface (ASMI), HMC, or the control panel.
Extended control panel functions
The extended control panel functions consist of two major groups:
Functions 21 through 49, which are available when you select Manual mode from Function
02.
Support service representative Functions 50 through 99, which are available when you
select Manual mode from Function 02, then select and enter the client service switch 1
(Function 25), followed by service switch 2 (Function 26).
66IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
Function 30 – CEC SP IP address and location
Function 30 is one of the Extended control panel functions and is only available when Manual
mode is selected. This function can be used to display the central electronic complex (CEC)
Service Processor IP address and location segment. The Table 2-21 shows an example of
how to use the Function 03.
Table 2-21 CEC SP IP address and location
Information on operator panelAction or description
3 0Use the increment or decrement buttons to scroll
to Function 30.
3 0 * *Press Enter to enter sub-function mode.
3 0 0 0Use the increment or decrement buttons to select
an IP address:
0 0 = Service Processor ETH0 or HMC1 port
0 1 = Service Processor ETH1 or HMC2 port
S P A : E T H 0 : _ _ _ T 5
1 9 2 . 1 6 8 . 2 . 1 4 7
3 0 * *Use the increment or decrement buttons to select
3 0Press Enter to exit sub-function mode.
2.14 System firmware
Server firmware is the part of the Licensed Internal Code that enables hardware, such as the
service processor. Depending on your service environment, you can download, install, and
manage your server firmware fixes using different interfaces and methods, including the
HMC, or by using functions specific to your operating system.
Note: Normally, installing the server firmware fixes through the operating system is a
nonconcurrent process.
Temporary and permanent firmware sides
The service processor maintains two copies of the server firmware:
One copy is considered the permanent or backup copy and is stored on the permanent
side, sometimes referred to as the
The other copy is considered the installed or temporary copy and is stored on the
temporary side, sometimes referred to as the
the server from the temporary side.
Press Enter to display the selected IP address.
sub-function exit.
p side.
t side. We recommend that you start and run
The copy actually booted from is called the activated level, sometimes referred to as
Note: The default value, from which the system boots, is temporary.
b.
Chapter 2. Architecture and technical overview 67
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
The following examples are the output of the lsmcode command for AIX and Linux, showing
the firmware levels as they are displayed in the outputs.
AIX:
The current permanent system firmware image is SF220_005.
The current temporary system firmware image is SF220_006.
The system is currently booted from the temporary image.
Linux:
system:SF220_006 (t) SF220_005 (p) SF220_006 (b)
When you install a server firmware fix, it is installed on the temporary side.
Note: The following points are of special interest:
The server firmware fix is installed on the temporary side only after the existing
contents of the temporary side are permanently installed on the permanent side (the
service processor performs this process automatically when you install a server
firmware fix).
If you want to preserve the contents of the permanent side, you need to remove the
current level of firmware (copy the contents of the permanent side to the temporary
side) before you install the fix.
However, if you get your fixes using the Advanced features on the HMC interface and
you indicate that you do not want the service processor to automatically accept the
firmware level, the contents of the temporary side are not automatically installed on the
permanent side. In this situation, you do not need to remove the current level of
firmware to preserve the contents of the permanent side before you install the fix.
You might want to use the new level of firmware for a period of time to verify that it works
correctly. When you are sure that the new level of firmware works correctly, you can
permanently install the server firmware fix. When you permanently install a server firmware
fix, you copy the temporary firmware level from the temporary side to the permanent side.
Conversely, if you decide that you do not want to keep the new level of server firmware, you
can remove the current level of firmware. When you remove the current level of firmware, you
copy the firmware level that is currently installed on the permanent side from the permanent
side to the temporary side.
System firmware download site
For the system firmware download site for the 570, go to:
In the main area of the firmware download site, select the correct machine type and model.
The 570 machine type and model is 9117-MMA (see Figure 2-27 on page 69)
68IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
Figure 2-27 IBM Microcode downloads site
Receive server firmware fixes using an HMC
If you use an HMC to manage your server and you periodically configure several partitions on
the server, you need to download and install fixes for your server and power subsystem
firmware.
How you get the fix depends on whether the HMC or server is connected to the Internet:
The HMC or server is connected to the Internet.
There are several repository locations from which you can download the fixes using the
HMC. For example, you can download the fixes from your service provider's Web site or
support system, from optical media that you order from your service provider, or from an
FTP server on which you previously placed the fixes.
Neither the HMC nor your server is connected to the Internet (server firmware only).
You need to download your new server firmware level to a CD-ROM media or FTP server.
For both of these options, you can use the interface on the HMC to install the firmware fix
(from one of the repository locations or from the optical media). The Change Internal Code
wizard on the HMC provides a step-by-step process for you to perform the procedure to install
the fix. Perform these steps:
1. Ensure that you have a connection to the service provider (if you have an Internet
connection from the HMC or server).
2. Determine the available levels of server and power subsystem firmware.
3. Create the optical media (if you do not have an Internet connection from the HMC or
server).
4. Use the Change Internal Code wizard to update your server and power subsystem
firmware.
5. Verify that the fix installed successfully.
Chapter 2. Architecture and technical overview 69
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
For a detailed description of each task, select System p information, support, and
troubleshooting → Fixes and upgrades → Getting fixes and upgrades from the IBM
Periodically, you need to install fixes for your server firmware. If you do not use an HMC to
manage your server, you must get your fixes through your operating system. In this situation,
you can get server firmware fixes through the operating system regardless of whether your
operating system is AIX or Linux.
To do this, complete the following tasks:
1. Determine the existing level of server firmware using the lsmcode command.
2. Determine the available levels of server firmware.
3. Get the server firmware.
4. Install the server firmware fix to the temporary side.
5. Verify that the server firmware fix installed successfully.
6. Install the server firmware fix permanently (optional).
Note: To view existing levels of server firmware using the lsmcode command, you need to
have the following service tools installed on your server:
AIX
You must have AIX diagnostics installed on your server to perform this task. AIX
diagnostics are installed when you install AIX on your server. However, it is possible to
deselect the diagnostics. Therefore, you need to ensure that the online AIX diagnostics
are installed before proceeding with this task.
The service processor is an embedded controller running the service processor internal
operating system. The service processor operating system contains specific programs and
device drivers for the service processor hardware. The host interface is a 32-bit PCI-X
interface connected to the Enhanced I/O Controller.
Service processor is used to monitor and manage the system hardware resources and
devices. The service processor offers the following connections:
Two Ethernet 10/100 Mbps ports
Both Ethernet ports are only visible to the service processor and can be used to attach the
p5-570 to a HMC or to access the Advanced System Management Interface (ASMI)
70IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
options from a client Web browser, using the HTTP server integrated into the service
processor internal operating system.
Both Ethernet ports have a default IP address:
– Service processor Eth0 or HMC1 port is configured as 169.254.2.147.
– Service processor Eth1 or HMC2 port is configured as 169.254.3.147.
2.14.2 Redundant service processor
A Service Interface card is required to be installed in every drawer. The card in the top drawer
provides the Service Processor function. Redundant Service Processor function for
managing the service processors when one fails is supported on all systems that are
operating with system firmware level FM320_xxx_xxx, or later. This support is available for
configurations with two or more CEC enclosures. The card in the second drawer provides a
Service Processor on standby capable of taking over the Service Processor function from the
primary drawer.
The SP Flash in the second drawer will be updated whenever an update is made to the SP
Flash in the top drawer. The Service Interface cards in drawers 3 and 4 do not use the
Service Processor function, and the FLASH code is not updated. Therefore Service Interface
cards from drawers 3 or 4 must NOT be moved into drawers 1 or 2.
Redundant Service Processor function requires that the Hardware Management Console
(HMC) be attached to the Service Interface Card in both CEC enclosure 1 and CEC
enclosure 2. The Service Interface Card in these two enclosures must be connected using an
external Power Control cable (FC 6006 or similar).
This section provides a brief overview of the different 570 hardware management user
interfaces available.
Advanced System Management Interface
The Advanced System Management Interface (ASMI) is the interface to the service processor
that enables you to set flags that affect the operation of the server, such as auto power
restart, and to view information about the server, such as the error log and vital product data.
This interface is accessible using a Web browser on a client system that is connected directly
to the service processor (in this case, a standard Ethernet cable or a crossed cable can be
both used) or through an Ethernet network. Using the
enables the ability to change the service processor IP addresses or to apply some security
policies and avoid the access from undesired IP addresses or range. The ASMI can also be
accessed using a terminal attached to the system service processor ports on the server, if the
server is not HMC managed. The service processor and the ASMI are standard on all IBM
System p servers.
network configuration menu, the ASMI
You might be able to use the service processor's default settings. In that case, accessing the
ASMI is not necessary.
Accessing the ASMI using a Web browser
The Web interface to the Advanced System Management Interface is accessible through, at
the time of writing, Microsoft® Internet Explorer® 6.0, Netscape 7.1, Mozilla Firefox, or
Chapter 2. Architecture and technical overview 71
4405ch02 Architecture and technical overview.fmDraft Document for Review May 28, 2009 1:59 pm
Opera 7.23 running on a PC or mobile computer connected to the service processor. The
Web interface is available during all phases of system operation, including the initial program
load and runtime. However, some of the menu options in the Web interface are unavailable
during IPL or runtime to prevent usage or ownership conflicts if the system resources are in
use during that phase.
Accessing the ASMI using an ASCII console
The Advanced System Management Interface on an ASCII console supports a subset of the
functions provided by the Web interface and is available only when the system is in the
platform standby state. The ASMI on an ASCII console is not available during some phases of
system operation, such as the initial program load and runtime.
Accessing the ASMI using an HMC
To access the Advanced System Management Interface using the Hardware Management
Console, complete the following steps:
1. Open Systems Management from the navigation pane.
2. From the work pane, select one or more managed systems to work with.
3. From the System Management tasks list, select Operations.
4. From the Operations task list, select Advanced System Management (ASM).
System Management Services
Use the System Management Services (SMS) menus to view information about your system
or partition and to perform tasks, such as changing the boot list or setting the network
parameters.
To start System Management Services, perform the following steps:
1. For a server that is connected to an HMC, use the HMC to restart the server or partition.
If the server is not connected to an HMC, stop the system, and then restart the server by
pressing the power button on the control panel.
2. For a partitioned server, watch the virtual terminal window on the HMC.
For a full server partition, watch the firmware console.
3. Look for the POST
5
indicators (memory, keyboard, network, SCSI, and speaker) that
appear across the bottom of the screen. Press the numeric 1 key after the word keyboard
appears and before the word speaker appears.
The SMS menus is useful to defining the operating system installation method, choosing the
installation boot device, or setting the boot device priority list for a full managed server or a
logical partition. In the case of a network boot, SMS menus are provided to set up the network
parameters and network adapter IP address.
HMC
The Hardware Management Console is a system that controls managed systems, including
IBM System p5 and p6 hardware, and logical partitions. To provide flexibility and availability,
there are different ways to implement HMCs.
Web-based System Manager Remote Client
The Web-based System Manager Remote Client is an application that is usually installed on
a PC and can be downloaded directly from an installed HMC. When an HMC is installed and
5
POST stands for power-on-self-test.
72IBM Power 570 Technical Overview and Introduction
Draft Document for Review May 28, 2009 1:59 pm4405ch02 Architecture and technical overview.fm
HMC Ethernet IP addresses have been assigned, it is possible to download the Web-based
System Manager Remote Client from a web browser, using the following URL:
http://HMC_IP_address/remote_client.html
You can then use the PC to access other HMCs remotely. Web-based System Manager
Remote Clients can be present in private and open networks. You can perform most
management tasks using the Web-based System Manager Remote Client. The remote HMC
and the Web-based System Manager Remote Client allow you the flexibility to access your
managed systems (including HMCs) from multiple locations using multiple HMCs.
For more detailed information about the use of the HMC, refer to the IBM Systems Hardware
Information Center.
Open Firmware
A System p6 server has one instance of Open Firmware both when in the partitioned
environment and when running as a full system partition. Open Firmware has access to all
devices and data in the server. Open Firmware is started when the server goes through a
power-on reset. Open Firmware, which runs in addition to the POWER Hypervisor in a
partitioned environment, runs in two modes: global and partition. Each mode of Open
Firmware shares the same firmware binary that is stored in the flash memory.
In a partitioned environment, Open Firmware runs on top of the global Open Firmware
instance. The partition Open Firmware is started when a partition is activated. Each partition
has its own instance of Open Firmware and has access to all the devices assigned to that
partition. However, each instance of Open Firmware has no access to devices outside of the
partition in which it runs. Partition firmware resides within the partition memory and is
replaced when AIX or Linux takes control. Partition firmware is needed only for the time that is
necessary to load AIX or Linux into the partition server memory.
The global Open Firmware environment includes the partition manager component. That
component is an application in the global Open Firmware that establishes partitions and their
corresponding resources (such as CPU, memory, and I/O slots), which are defined in partition
profiles. The partition manager manages the operational partitioning transactions. It responds
to commands from the service processor external command interface that originates in the
application running on the HMC. The ASMI can be accessed during boot time or using the
ASMI and selecting the boot to Open Firmware prompt.
For more information about Open Firmware, refer to Partitioning Implementations for IBM
eServer Partitioning Implementations for IBM Eserver p5 Servers, SG24-7039, which is
available at:
4405ch03 Virtualization.fmDraft Document for Review September 2, 2008 5:05 pm
3.1 POWER Hypervisor
Combined with features designed into the POWER6 processors, the POWER Hypervisor
delivers functions that enable other system technologies, including logical partitioning
technology, virtualized processors, IEEE VLAN compatible virtual switch, virtual SCSI
adapters, and virtual consoles. The POWER Hypervisor is a basic component of the system’s
firmware and offers the following functions:
Provides an abstraction between the physical hardware resources and the logical
partitions that use them.
Enforces partition integrity by providing a security layer between logical partitions.
Controls the dispatch of virtual processors to physical processors (see 3.2.3, “Processing
mode” on page 79).
Saves and restores all processor state information during a logical processor context
switch.
Controls hardware I/O interrupt management facilities for logical partitions.
Provides virtual LAN channels between logical partitions that help to reduce the need for
physical Ethernet adapters for inter-partition communication.
Monitors the Service Processor and will perform a reset/reload if it detects the loss of the
Service Processor, notifying the operating system if the problem is not corrected.
The POWER Hypervisor is always active, regardless of the system configuration and also
when not connected to the HMC. It requires memory to support the resource assignment to
the logical partitions on the server. The amount of memory required by the POWER
Hypervisor firmware varies according to several factors. Factors influencing the POWER
Hypervisor memory requirements include the following:
Number of logical partitions.
Number of physical and virtual I/O devices used by the logical partitions.
Maximum memory values given to the logical partitions.
The minimum amount of physical memory to create a partition is the size of the system’s
Logical Memory Block (LMB). The default LMB size varies according to the amount of
memory configured in the CEC as shown in Table 3-1.
But in most cases, the actual requirements and recommendations are between 256 MB and
512 MB for AIX, Red Hat, and Novell SUSE Linux. Physical memory is assigned to partitions
in increments of Logical Memory Block (LMB).
The POWER Hypervisor provides the following types of virtual I/O adapters:
Virtual SCSI
76IBM Power 570 Technical Overview and Introduction
Draft Document for Review September 2, 2008 5:05 pm4405ch03 Virtualization.fm
Virtual Ethernet
Virtual (TTY) console
Virtual SCSI
The POWER Hypervisor provides a virtual SCSI mechanism for virtualization of storage
devices (a special logical partition to install the Virtual I/O Server is required to use this
feature, as described in 3.3.2, “Virtual I/O Server” on page 81). The storage virtualization is
accomplished using two, paired, adapters: a virtual SCSI server adapter and a virtual SCSI
client adapter. Only the Virtual I/O Server partition can define virtual SCSI server adapters,
other partitions are
client partitions. The Virtual I/O Server is available with the optional
PowerVM Edition features.
Virtual Ethernet
The POWER Hypervisor provides a virtual Ethernet switch function that allows partitions on
the same server to use a fast and secure communication without any need for physical
interconnection. The virtual Ethernet allows a transmission speed in the range of 1 to 3 Gbps.
depending on the MTU
Version 5.3, or appropriate level of Linux supporting Virtual Ethernet devices (see chapter
3.3.7, “Operating System support for PowerVM” on page 90). The virtual Ethernet is part of
the base system configuration.
1
size and CPU entitlement. Virtual Ethernet support starts with AIX 5L
Virtual Ethernet has the following major features:
The virtual Ethernet adapters can be used for both IPv4 and IPv6 communication and can
transmit packets with a size up to 65408 bytes. Therefore, the maximum MTU for the
corresponding interface can be up to 65394 (65390 if VLAN tagging is used).
The POWER Hypervisor presents itself to partitions as a virtual 802.1Q compliant switch.
The maximum number of VLANs is 4096. Virtual Ethernet adapters can be configured as
either untagged or tagged (following the IEEE 802.1Q VLAN standard).
A partition supports 256 virtual Ethernet adapters. Besides a default port VLAN ID, the
number of additional VLAN ID values that can be assigned per Virtual Ethernet adapter is
20, which implies that each Virtual Ethernet adapter can be used to access 21 virtual
networks.
Each partition operating system detects the virtual local area network (VLAN) switch as an
Ethernet adapter without the physical link properties and asynchronous data transmit
operations.
Any virtual Ethernet can also have connectivity outside of the server if a layer-2 bridge to a
physical Ethernet adapter is set in one Virtual I/O server partition (see 3.3.2, “Virtual I/O
Server” on page 81 for more details about shared Ethernet). Also known as Shared Ethernet
Adapter.
Note: Virtual Ethernet is based on the IEEE 802.1Q VLAN standard. No physical I/O
adapter is required when creating a VLAN connection between partitions, and no access to
an outside network is required.
Virtual (TTY) console
Each partition needs to have access to a system console. Tasks such as operating system
installation, network setup, and some problem analysis activities require a dedicated system
console. The POWER Hypervisor provides the virtual console using a virtual TTY or serial
1
Maximum transmission unit
Chapter 3. Virtualization 77
4405ch03 Virtualization.fmDraft Document for Review September 2, 2008 5:05 pm
adapter and a set of Hypervisor calls to operate on them. Virtual TTY does not require the
purchase of any additional features or software such as the PowerVM Edition features.
Depending on the system configuration, the operating system console can be provided by the
Hardware Management Console virtual TTY, IVM virtual TTY, or from a terminal emulator that
is connected to a system port.
3.2 Logical partitioning
Logical partitions (LPARs) and virtualization increase utilization of system resources and add
a new level of configuration possibilities. This section provides details and configuration
specifications about this topic.
3.2.1 Dynamic logical partitioning
Logical partitioning (LPAR) was introduced with the POWER4™ processor-based product line
and the AIX 5L Version 5.1 operating system. This technology offered the capability to divide
a pSeries system into separate logical systems, allowing each LPAR to run an operating
environment on dedicated attached devices, such as processors, memory, and I/O
components.
Later, dynamic logical partitioning increased the flexibility, allowing selected system
resources, such as processors, memory, and I/O components, to be added and deleted from
logical partitions while they are executing. AIX 5L Version 5.2, with all the necessary
enhancements to enable dynamic LPAR, was introduced in 2002. The ability to reconfigure
dynamic LPARs encourages system administrators to dynamically redefine all available
system resources to reach the optimum capacity for each defined dynamic LPAR.
3.2.2 Micro-Partitioning
Micro-Partitioning technology allows you to allocate fractions of processors to a logical
partition. This technology was introduced with POWER5 processor-based systems. A logical
partition using fractions of processors is also known as a Shared Processor Partition or
Micro-Partition. Micro-Partitions run over a set of processors called Shared Processor Pool.
And virtual processors are used to let the operating system manage the fractions of
processing power assigned to the logical partition. From an operating system perspective, a
virtual processor cannot be distinguished from a physical processor, unless the operating
system has been enhanced to be made aware of the difference. Physical processors are
abstracted into virtual processors that are available to partitions. The meaning of the term
physical processor in this section is a processor core. For example, in a 2-core server there
are two physical processors.
When defining a shared processor partition, several options have to be defined:
The minimum, desired, and maximum processing units. Processing units are defined as
processing power, or the fraction of time the partition is dispatched on physical
processors. Processing units define the capacity entitlement of the partition.
The shared processor pool. Pick one from the list with the names of each configured
shared processor pool. This list also displays the pool ID of each configured shared
processor pool in parentheses. If the name of the desired shared processor pool is not
available here, you must first configure the desired shared processor pool using the
Shared Processor Pool Management window. Shared processor partitions use the default
shared processor pool called DefaultPool by default.
78IBM Power 570 Technical Overview and Introduction
Draft Document for Review September 2, 2008 5:05 pm4405ch03 Virtualization.fm
Select whether the partition will be able or not to access extra processing power to “fill up”
its virtual processors above its capacity entitlement. Selecting either to capp or uncapp
your partition. If there is spare processing power available in the shared processor pool or
other partitions are not using their entitlement, an uncapped partition can use additional
processing units if its entitlement is not enough to satisfy its application processing
demand.
The weight (preference) in the case of an uncapped partition.
The minimum, desired, and maximum number of virtual processors.
The POWER Hypervisor calculates partition’s processing power based on minimum, desired,
and maximum values, processing mode and also based on other active partitions’
requirements. The actual entitlement is never smaller than the processing units desired value
but can exceed that value in the case of an uncapped partition and up to the number of virtual
processors allocated.
A partition can be defined with a processor capacity as small as 0.10 processing units. This
represents 1/10th of a physical processor. Each physical processor can be shared by up to 10
shared processor partitions and the partition’s entitlement can be incremented fractionally by
as little as 1/100th of the processor. The shared processor partitions are dispatched and
time-sliced on the physical processors under control of the POWER Hypervisor. The shared
processor partitions are created and managed by the HMC or Integrated Virtualization
Management.
This system supports up to a 16-core configuration, therefore up to sixteen dedicated
partitions, or up to 160 micro-partitions, can be created. It is important to point out that the
maximums stated are supported by the hardware, but the practical limits depend on the
application workload demands.
Additional information on virtual processors:
A virtual processor can be either running (dispatched) on a physical processor or standby
waiting for a physical processor to became available.
Virtual processors do not introduce any additional abstraction level; they really are only a
dispatch entity. When running on a physical processor, virtual processors run at the same
speed as the physical processor.
Each partition’s profile defines CPU entitlement that determines how much processing
power any given partition should receive. The total sum of CPU entitlement of all partitions
cannot exceed the number of available physical processors in a shared processor pool.
The number of virtual processors can be changed dynamically through a dynamic LPAR
operation.
3.2.3 Processing mode
When you create a logical partition you can assign entire processors for dedicated use, or you
can assign partial processor units from a shared processor pool. This setting will define the
processing mode of the logical partition. Figure 3-1 on page 80 shows a diagram of the
concepts discussed in this section.
Chapter 3. Virtualization 79
4405ch03 Virtualization.fmDraft Document for Review September 2, 2008 5:05 pm
Figure 3-1 Logical partitioning concepts
Dedicated mode
In dedicated mode, physical processors are assigned as a whole to partitions. The
simultaneous multithreading feature in the POWER6 processor core allows the core to
execute instructions from two independent software threads simultaneously. To support this
feature we use the concept of
one physical processor as two logical processors if the simultaneous multithreading feature is
on. It can be turned off and on dynamically while the operating system is executing (for AIX,
use the smtctl command). If simultaneous multithreading is off, then each physical processor
is presented as one logical processor and thus only one thread at a time is executed on the
physical processor.
logical processors. The operating system (AIX or Linux) sees
Shared dedicated mode
On POWER6 servers, you can configure dedicated partitions to become processor donors for
idle processors they own. Allowing for the donation of spare CPU cycles from dedicated
processor partitions to a Shared Processor Pool. The dedicated partition maintains absolute
priority for dedicated CPU cycles. Enabling this feature may help to increase system
utilization, without compromising the computing power for critical workloads in a dedicated
processor.
Shared mode
In shared mode, logical partitions use virtual processors to access fractions of physical
processors. Shared partitions can define any number of virtual processors (maximum number
is 10 times the number of processing units assigned to the partition). From the POWER
Hypervisor point of view, virtual processors represent dispatching objects. The POWER
Hypervisor dispatches virtual processors to physical processors according to partition’s
processing units entitlement. One Processing Unit represents one physical processor’s
processing capacity. At the end of the POWER Hypervisor’s dispatch cycle (10 ms), all
partitions should receive total CPU time equal to their processing units entitlement. The
logical processors are defined on top of virtual processors. So, even with a virtual processor,
the concept of logical processor exists and the number of logical processor depends whether
the simultaneous multithreading is turned on or off.
80IBM Power 570 Technical Overview and Introduction
Draft Document for Review September 2, 2008 5:05 pm4405ch03 Virtualization.fm
3.3 PowerVM
The PowerVM platform is the family of technologies, capabilities and offerings that deliver
industry-leading virtualization on this 570. It is the new umbrella branding term for Power
Systems Virtualization (Logical Partitioning, Micro-Partitioning, Hypervisor, Virtual I/O Server,
Advanced Power Virtualization, Live Partition Mobility, Workload Partitions, etc.). As with
Advanced Power Virtualization in the past, PowerVM is a combination of hardware
enablement and value-added software. Section 3.3.1, “PowerVM editions” on page 81
discusses the licensed features of each of the 2 different editions of PowerVM.
3.3.1 PowerVM editions
This section provides information about the virtualization capabilities of the PowerVM
Standard Edition and Enterprise Edition which are available on this system. Upgrading from
the PowerVM Standard Edition to the PowerVM Express Edition is possible and is completely
undisruptive. The upgrade doesn’t even require the installation of additional software, the
customer just has to enter a key code in the hypervisor in order to unlock the next level of
function.
Table 3-2, outlines the functional elements of both of the PowerVM editions.
Table 3-2 PowerVM capabilities
PowerVM capabilityPowerVM Standard
Micro-partitionsYesYes
Virtual I/O ServerYesYes
Shared Dedicated CapacityYesYes
Multiple Shared-Processor PoolsYesYes
Lx86YesYes
Live Partition MobilityNoYes
Maximum # Logical PartitionsUp to 10 per coreUp to 10 per core
For more information about the different PowerVM editions please refer to PowerVM Virtualization on IBM System p Introduction and Configuration, SG24-7940.
Note The 570 has to be managed with the Hardware Management Console.
3.3.2 Virtual I/O Server
The Virtual I/O Server is part of all PowerVM Editions. It is a special purpose partition that
allows the sharing of physical resources between logical partitions to allow more efficient
utilization ( for example consolidation). In this case the Virtual I/O Server owns the physical
resources (SCSI, Fibre Channel, network adapters, and optical devices) and allows client
partitions to share access to them, thus minimizing the number of physical adapters in the
system. The Virtual I/O Server eliminates the requirement that every partition owns a
dedicated network adapter, disk adapter, and disk drive. The Virtual I/O Server supports
OpenSSH for secure remote logins. It also provides a firewall for limiting access by ports,
Edition (FC 7942)
PowerVM Enterprise
Edition (FC 7995)
Chapter 3. Virtualization 81
4405ch03 Virtualization.fmDraft Document for Review September 2, 2008 5:05 pm
network services and IP addresses. Figure 3-2 shows an overview of a Virtual I/O Server
configuration.
external network
physical disk
physical disk
Virtual I/O Server
Shared Ethernet
physical
Ethernet adapter
physical disk
adapter
Adapter
virtual Ethernet
adapter
virtual SCSI
adapter
Hypervisor
Virtual I/O client 1
virtual Ethernet
adapter
virtual SCSI
adapter
Virtual I/O client 2
virtual Ethernet
adapter
virtual SCSI
adapter
Figure 3-2 Architectural view of the Virtual I/O Server
Because the Virtual I/O server is an operating system-based appliance server, redundancy
for physical devices attached to the Virtual I/O Server can be provided by using capabilities
such as Multipath I/O and IEEE 802.3ad Link Aggregation.
Installation of the Virtual I/O Server partition is performed from a special system backup DVD
that is provided to clients that order any PowerVM edition. This dedicated software is only for
the Virtual I/O Server (and IVM in case it is used) and is only supported in special Virtual I/O
Server partitions. Two major functions are provided with the Virtual I/O Server: a Shared
Ethernet Adapter and Virtual SCSI.
Shared Ethernet Adapter
A Shared Ethernet Adapter (SEA) can be used to connect a physical Ethernet network to a
virtual Ethernet network. The Shared Ethernet Adapter provides this access by connecting
the internal Hypervisor VLANs with the VLANs on the external switches. Because the Shared
Ethernet Adapter processes packets at layer 2, the original MAC address and VLAN tags of
the packet are visible to other systems on the physical network. IEEE 802.1 VLAN tagging is
supported.
The Shared Ethernet Adapter also provides the ability for several client partitions to share one
physical adapter. Using an SEA, you can connect internal and external VLANs using a
physical adapter. The Shared Ethernet Adapter service can only be hosted in the Virtual I/O
Server, not in a general purpose AIX or Linux partition, and acts as a layer-2 network bridge
to securely transport network traffic between virtual Ethernet networks (internal) and one or
more (EtherChannel) physical network adapters (external). These virtual Ethernet network
adapters are defined by the POWER Hypervisor on the Virtual I/O Server
Tip: A Linux partition can provide bridging function as well, by using the brctl command.
Figure 3-3 on page 83 shows a configuration example of an SEA with one physical and two
virtual Ethernet adapters. An SEA can include up to 16 virtual Ethernet adapters on the
Virtual I/O Server that share the same physical access.
82IBM Power 570 Technical Overview and Introduction
Draft Document for Review September 2, 2008 5:05 pm4405ch03 Virtualization.fm
Figure 3-3 Architectural view of a Shared Ethernet Adapter
A single SEA setup can have up to 16 Virtual Ethernet trunk adapters and each virtual
Ethernet trunk adapter can support up to 20 VLAN networks. Therefore, it is possible for a
single physical Ethernet to be shared between 320 internal VLAN. The number of shared
Ethernet adapters that can be set up in a Virtual I/O server partition is limited only by the
resource availability as there are no configuration limits.
Unicast, broadcast, and multicast is supported, so protocols that rely on broadcast or
multicast, such as Address Resolution Protocol (ARP), Dynamic Host Configuration Protocol
(DHCP), Boot Protocol (BOOTP), and Neighbor Discovery Protocol (NDP) can work across
an SEA.
Note: A Shared Ethernet Adapter does not need to have an IP address configured to be
able to perform the Ethernet bridging functionality. It is very convenient to configure IP on
the Virtual I/O Server. This is because the Virtual I/O Server can then be reached by
TCP/IP, for example, to perform dynamic LPAR operations or to enable remote login. This
can be done either by configuring an IP address directly on the SEA device, or on an
additional virtual Ethernet adapter in the Virtual I/O Server. This leaves the SEA without
the IP address, allowing for maintenance on the SEA without losing IP connectivity in case
SEA failover is configured.
For a more detailed discussion about virtual networking, see:
4405ch03 Virtualization.fmDraft Document for Review September 2, 2008 5:05 pm
Virtual SCSI
Virtual SCSI is used to refer to a virtualized implementation of the SCSI protocol. Virtual SCSI
is based on a client/server relationship. The Virtual I/O Server logical partition owns the
physical resources and acts as server or, in SCSI terms, target device. The client logical
partitions access the virtual SCSI backing storage devices provided by the Virtual I/O Server
as clients.
The virtual I/O adapters (virtual SCSI server adapter and a virtual SCSI client adapter) are
configured using an HMC or through the Integrated Virtualization Manager on smaller
systems. The virtual SCSI server (target) adapter is responsible for executing any SCSI
commands it receives. It is owned by the Virtual I/O Server partition. The virtual SCSI client
adapter allows a client partition to access physical SCSI and SAN attached devices and
LUNs that are assigned to the client partition. The provisioning of virtual disk resources is
provided by the Virtual I/O Server.
Physical disks presented to the Virtual/O Server can be exported and assigned to a client
partition in a number of different ways:
The entire disk is presented to the client partition
The disk is divided into several logical volumes, these can be presented to a single client
or multiple different clients
As of Virtual I/O Server 1.5, files can be created on these disks and file backed storage
devices can be created
The Logical volumes or files can be assigned to different partitions. Therefore, virtual SCSI
enables sharing of adapters as well as disk devices.
Figure 3-4 shows an example where one physical disk is divided into two logical volumes by
the Virtual I/O Server. Each of the two client partitions is assigned one logical volume, which
is then accessed through a virtual I/O adapter (VSCSI Client Adapter). Inside the partition,
the disk is seen as a normal hdisk.
Figure 3-4 Architectural view of virtual SCSI
84IBM Power 570 Technical Overview and Introduction
Draft Document for Review September 2, 2008 5:05 pm4405ch03 Virtualization.fm
At the time of writing, virtual SCSI supports Fibre Channel, parallel SCSI, iSCSI, SAS, SCSI
RAID devices and optical devices, including DVD-RAM and DVD-ROM. Other protocols such
as SSA and tape devices are not supported.
For more information about the specific storage devices supported for Virtual I/O Server, see:
Virtual I/O Server includes a number of features, including monitoring solutions:
Support for Live Partition Mobility on POWER6 processor-based systems with the
PowerVM Enterprise Edition. More information about Live Partition Mobility can be found
on 3.3.4, “PowerVM Live Partition Mobility” on page 87.
Support for virtual SCSI devices backed by a file. These are then accessed as standard
SCSI-compliant LUNs.
Virtual I/O Server Expansion Pack with additional security functions like Kerberos
(Network Authentication Service for users and Client and Server Applications), SNMP v3
(Simple Network Management Protocol) and LDAP (Lightweight Directory Access
Protocol client functionality).
System Planning Tool (SPT) and Workload Estimator are designed to ease the
deployment of a virtualized infrastructure. More on the System Planning Tool in section
3.4, “System Planning Tool” on page 91.
IBM Systems Director and a number of preinstalled Tivoli® agents are included like Tivoli
Identity Manager (TIM) in order to allow easy integration into an existing Tivoli Systems
Management infrastructure, and Tivoli Application Dependency Discovery Manager
(ADDM) which creates and maintains automatically application infrastructure maps
including dependencies, change histories and deep configuration values.
vSCSI eRAS
Additional Command Line Interface (CLI) statistics in svmon, vmstat, fcstat and topas
Monitoring solutions to help manage and monitor the Virtual I/O Server and shared
resources. New commands and views provide additional metrics for memory, paging,
processes, Fibre Channel HBA statistics and virtualization.
For more information on the Virtual I/O Server and its implementation, refer to PowerVM virtualization on IBM System p, Introduction and Configuration, SG24-7940.
3.3.3 PowerVM Lx86
The IBM PowerVM Lx86 feature creates a virtual x86 Linux application environment on
POWER processor-based systems, so most 32-bit x86 Linux applications can run without
requiring clients or ISVs to recompile the code. This brings new benefits to organizations who
want the reliability and flexibility of consolidating (through virtualization) on Power Systems
and use applications that have not yet been ported to the platform.
PowerVM Lx86 dynamically translates x86 instructions to Power Architecture instructions,
operating much like the Just-in-time compiler (JIT) in a Java™ system. The technology
creates an environment in which the applications being translated run on the new target
platform, in this case Linux on POWER. This environment encapsulates the application and
runtime libraries and runs them on the Linux on POWER operating system kernel. These
applications can be run side by side with POWER native applications on a single system
image and do not require a separate partition.
Chapter 3. Virtualization 85
4405ch03 Virtualization.fmDraft Document for Review September 2, 2008 5:05 pm
Figure 3-5 shows the diagram of the Linux x86 application environment.
Figure 3-5 Diagram of the Linux x86 Application Environment
Supported Operating Systems
PowerVM Lx86 version 1.1 will support the following Linux on POWER operating systems:
Red Hat Enterprise Linux 4 (RHEL 4) for POWER version 4.4 and 4.5. Also, x86 Linux
applications running on RHEL 4.3 are supported.
SUSE Linux Enterprise Server 9 (SLES 9) for POWER Service Pack 3
SUSE Linux Enterprise Server 10 (SLES 10) for POWER Service Pack 1
Note:
PowerVM LX86 is supported under the VIOS Software Maintenance Agreement
(SWMA).
When using PowerVM Lx86 on an IBM System p POWER6 processor-based system
only SLES 10 with SP1 and RHEL 4.5 are supported.
Make sure the x86 version is the same as your Linux on POWER version. Do not try to
use any other version because it is unlikely to work. One exception is with Red Hat
Enterprise Linux, both the Advanced Server and Enterprise Server option at the correct
release will work.
As stated in a previous paragraph, PowerVM Lx86 runs most x86 Linux applications, but
PowerVM Lx86 cannot run applications that:
Directly access hardware devices (for example, graphics adapters)
Require nonstandard kernel module access or use kernel modules not provided by the
Linux for POWER operating system distribution
Do not use only the Intel® IA-32 instruction set architecture as defined by the 1997 Intel
Architecture Software Developer's Manual consisting of Basic Architecture (Order Number
243190), Instruction Set Reference Manual (Order Number 243191) and the System
Programming Guide (Order Number 243192) dated 1997
86IBM Power 570 Technical Overview and Introduction
Loading...
+ hidden pages
You need points to download manuals.
1 point = 1 manual.
You can buy points or you can get point for every manual you upload.