Designed for demanding virtualization and database workloads, the UCS B420 M4 combines a large memory
footprint with four-socket scalability, leveraging the Intel® Xeon® processor E5-4600 v3 and v4 product
family. The UCS B420 M4 supports 2133-MHz and 2400-MHz DDR4 memory and utilizes UCS Virtual Interface
Card (VIC) technology to achieve up to 160 Gb aggregate I/O bandwidth, all in a dense, full-width blade
form factor. The UCS B420 M4 maintains memory performance even as capacity grows, and the large power
envelope of the UCS 5108 Blade Server Chassis means the UCS B420 can handle up to 3 TB of memory
without compromise to CPU speed or core count. Up to four UCS B420 M4 Blade Servers can be installed in
the UCS 5108 Blade Server Chassis.
The UCS B420 M4 Blade Server is centrally managed using Cisco UCS Manager. Cisco UCSM provides a single
management interface for all hardware and software components of a UCS domain. Cisco UCSM manages up
to 160 servers, using Single Connect Technology to consolidate network, storage, and management traffic.
With Cisco UCS Central, management can be extended globally to thousands of servers in multiple domains.
Figure 1 Cisco UCS B420 M4 Blade Server
Cisco UCS B420 M4 Blade Server
3
Page 4
DETAILED VIEWS
Notes
305013
1
5781310 11
912
1414
6
234
DETAILED VIEWS
Blade Server Front View
Figure 2 is a detailed front view of the Cisco UCS B420 M4 Blade Server.
Figure 2 Blade Server Front View
1Drive bay 18Power button and LED
2Drive bay 29Network link status button
3Drive bay 310Blade health LED
4Drive bay 411
5Left blade ejector handle12Reset button access
6Asset pull handle
(a blank asset tag is provided on which
you can add your own label or sticker or
you can use a marker to write your asset
information on the tag)
7Right blade ejector handle14Ejector thumbscrew
1. For information about the KVM local I/O cable that plugs into the console connector (a cable is included with
every Cisco UCS 5100 Series blade server chassis accessory kit), see CHOOSE LOCAL KVM I/O CABLE*
(OPTIONAL) on page 37.
13Beaconing LED and button
Local console connector
1
4
Cisco UCS B420 M4 Blade Server
Page 5
BASE SERVER STANDARD CAPABILITIES and FEATURES
BASE SERVER STANDARD CAPABILITIES and FEATURES
Table 1 lists the capabilities and features of the base server. Details about how to configure the server for
a particular feature or capability (for example, number of processors, disk drives, or amount of memory)
are provided in
Table 1 Capabilities and Features
Capability/FeatureDescription
Blade server chassisThe full-width UCS B420 M4 Blade Server mounts in a Cisco UCS 5100 series
CPUUp to four Intel® E5-4600 v3 or v4 series processor family CPUs.
CONFIGURING the SERVER on page 7.
NOTE: The B420 M4 blade server requires UCS Manager (UCSM) to operate as part of
the UCS system.
■ The B420 M4 with E5-4600 v3 or v4 CPUs requires UCSM 2.2(5a) or later
blade server chassis
2- and 4-CPU configurations are supported on the UCS B420 M4.
ChipsetIntel® C610 chipset
Memory48 total slots for registered ECC DIMMs
I/OThree configurable adapter slots:
■ One slot supports the Cisco VIC 1340/1240 adapter only.
■ Two additional slots support Cisco adapters and Cisco UCS Storage
Accelerator adapters.
Storage controller ■ Diskless system (no drive bays or RAID controller), or
■ Optional Cisco FlexStorage 12G SAS RAID controller options, offering:
Storage devicesEither two or four hot-plug 2.5” SFF drive bays
One internal USB 3.0 port
Two Cisco Flexible Flash secure digital card slots
Interfaces Front panel
■ One console connector
(see CHOOSE LOCAL KVM I/O CABLE* (OPTIONAL) on page 37)
Cisco UCS B420 M4 Blade Server
5
Page 6
BASE SERVER STANDARD CAPABILITIES and FEATURES
Table 1 Capabilities and Features (continued)
Capability/FeatureDescription
VideoThe Cisco Integrated Management Controller (CIMC) provides video using the
Matrox G200e video/graphics controller:
■ Integrated 2D graphics core with hardware acceleration
■ DDR2/3 memory interface supports up to 512 MB of addressable memory
(8 MB is allocated by default to video memory)
■ Supports display resolutions up to 1920 x 1200 16bpp @ 60Hz
■ High-speed integrated 24-bit RAMDAC
■ Single lane PCI-Express host interface running at Gen 1 speed
Power subsystem Integrated in the Cisco UCS 5100 series blade server chassis
Fans Integrated in the Cisco UCS 5100 series blade server chassis
Integrated management
processor
Cisco UCS Diagnostics
for Cisco UCS B-Series
Blade Servers
The built-in Cisco Integrated Management Controller (CIMC) GUI or CLI
interface enables you to monitor the server inventory, health, and system
event logs.
The Cisco UCS Blade Server Diagnostics tool for Cisco UCS Blade Servers
enables you to verify the health of the hardware components on your
servers. The diagnostics tool provides a variety of tests to exercise and
stress the various hardware subsystems on the Cisco UCS Blade Servers, such
as memory and CPU. You can use the tool to run a sanity check on the state
of your Cisco UCS Blade Servers after you fix or replace a hardware
component. You can also use this tool to run comprehensive burn-in tests
before you deploy a new Cisco UCS Blade Server in your production
environment.
Follow these steps to configure the Cisco UCS B420 M4 Blade Server:
■STEP 1 VERIFY SERVER SKU, page 8
■STEP 2 CHOOSE CPU(S), page 9
■STEP 3 CHOOSE MEMORY, page 11
■STEP 5 CHOOSE HARD DISK DRIVES or SOLID-STATE DRIVES (OPTIONAL), page 19
■STEP 6 CHOOSE ADAPTERS, page 21
■STEP 7 ORDER A TRUSTED PLATFORM MODULE (OPTIONAL), page 25
■STEP 8 ORDER CISCO FLEXIBLE FLASH SECURE DIGITAL CARDS, page 26
■STEP 9 ORDER INTERNAL USB 3.0 DRIVE (OPTIONAL), page 27
■STEP 10 CHOOSE OPERATING SYSTEM AND VALUE-ADDED SOFTWARE, page 28
■STEP 11 CHOOSE OPERATING SYSTEM MEDIA KIT (OPTIONAL), page 31
■STEP 12 CHOOSE SERVICE and SUPPORT LEVEL, page 32
■STEP 13 CHOOSE LOCAL KVM I/O CABLE* (OPTIONAL), page 37
CONFIGURING the SERVER
Cisco UCS B420 M4 Blade Server
7
Page 8
CONFIGURING the SERVER
STEP 1VERIFY SERVER SKU
Verify the product ID (PID) of the server as shown in Table 2.
Table 2 PID of the Base UCS B420 M4 Blade Server
Product ID (PID)Description
UCSB-B420-M4UCS B420 M4 Blade Server with no CPU, memory, HDD, SSD, or adapter cards
The base Cisco UCS B420 M4 blade server does not include the following components. They must
be selected during product ordering:
■CPUs
■Memory
■Disk drives
■Cisco adapters (such as the VIC 1340, VIC 1240, VIC 1380, VIC 1280, and Port Expander)
■Cisco UCS Storage Accelerators
NOTE: Use the steps on the following pages to order servers with the
configurable components that you want configured in your servers.
8
Cisco UCS B420 M4 Blade Server
Page 9
STEP 2CHOOSE CPU(S)
Notes
The standard CPU features are:
■Intel Xeon processor E5-4600 v3 or v4 series processor family CPUs
■Core counts of up to 18 for v3 CPUs and 22 for v4 CPUs
■Cache sizes of up to 45 MB for v3 CPUs and 55 MB for v4 CPUs
Select CPUs
The supported Intel Xeon E5-4600 v3 and v4 series CPUs on the UCS B420 M4 are listed in
Table 3.
Table 3 Supported Intel CPUs: E5-4600 v3 and v4 Series Processor Family CPUs
CONFIGURING the SERVER
Clock
Freq
(GHz)
Power
(W)
Cache
Size (MB)
CoresQPI
Highest DDR4 DIMM
Clock Support (MHz)
Product ID (PID)
Intel
Number
Intel Xeon E5-4600 v4
UCS-CPU-E5-4669EE5-4669 v42.2013555229.62400
UCS-CPU-E5-4667EE5-4667 v42.2013545189.62400
UCS-CPU-E5-4660EE5-4660 V42.212040169.62400
UCS-CPU-E5-4650EE5-4650 V42.210535149.62400
UCS-CPU-E5-4655EE5-4655 V42.51353089.62400
UCS-CPU-E5-4640EE5-4640 V42.110530128.02133
1
UCS-CPU-E5-4627E
E5-4627 V42.613525108.02400
UCS-CPU-E5-4620EE5-4620 V42.110525108.02133
UCS-CPU-E5-4610EE5-4610 V41.810525106.41866
Intel Xeon E5-4600 v3
UCS-CPU-E5-4669DE5-4669 v32.1013545189.62133
UCS-CPU-E5-4667DE5-4667 v32.0013540169.62133
UCS-CPU-E5-4660DE5-4660 v32.1012035149.62133
UCS-CPU-E5-4655DE5-4655 v32.901353069.62133
UCS-CPU-E5-4650DE5-4650 v32.1010530129.62133
UCS-CPU-E5-4640DE5-4640 v31.9010530128.01866
2
UCS-CPU-E5-4627D
E5-4627 v32.6013525108.02133
UCS-CPU-E5-4620DE5-4620 v32.0010525108.01866
UCS-CPU-E5-4610DE5-4610 v31.7010525106.41600
1. The UCS-CPU-E5-4627E CPU does not support Hyper-Threading technology.
2. The UCS-CPU-E5-4627D CPU does not support Hyper-Threading technology.
Cisco UCS B420 M4 Blade Server
9
Page 10
CONFIGURING the SERVER
Supported Configurations
(1) Two-CPU Configuration
■Choose two identical CPUs from any one of the rows of Table 3. CPUs 1 and 2 will be
populated.
(2) Four-CPU Configuration
■Choose four identical CPUs from any one of the rows of Table 3.
Notes
■The system will run at the lowest CPU or DIMM clock speed.
■System speed is also dependent on how many DIMMs are populated per channel.
■For 2-CPU systems, only 24 DIMM slots are functional (channels A - H).
■See Table 12 on page 23 for supported adapter combinations in 2-CPU systems.
■For 4-CPU systems, all 48 DIMM slots are functional.
10
Cisco UCS B420 M4 Blade Server
Page 11
STEP 3CHOOSE MEMORY
CPU 1CPU 2
48 DIMMS
4 memory channels per CPU
3 DIMMs per channel
CPU 3CPU 4
Channel A
Channel B
Channel C
Channel D
Channel E
Channel F
Channel G
Channel H
Channel M
Channel N
Channel O
Channel P
Channel I
Channel J
Channel K
Channel L
A1A2A3
B3
C3
D3
I3
J3
K3
I3
M3
N3
O3
P3
E3
F3
G3
H3
B1B2
C1C2
D1D2
Bank 1
Bank 2
Bank 3
E1 E2
F1 F2
G1 G2
H1 H2
Bank 1
Bank 2
Bank 3
Bank 1
Bank 2
Bank 3
M1 M2
N1 N2
O1 O2
P1 P2
I1I2
J1J2
K1K2
I1I2
Bank 1
Bank 2
Bank 3
The standard memory features are:
■DIMMs
—DDR4 ECC registered DIMMs (RDIMMs), through silicon via RDIMMs (TSV-RDIMMs), or
load-reduced DIMMs (LRDIMMs)
—Clock speed: 2133 MHz or 2400 MHz
—Ranks per DIMM: up to 8
—Operational voltage: 1.2 V
■Memory is organized with four memory channels per CPU, with up to three DIMMs per
channel (DPC), as shown in
Figure 3 UCS B420 M4 Memory Organization
Figure 3.
CONFIGURING the SERVER
Cisco UCS B420 M4 Blade Server
11
Page 12
CONFIGURING the SERVER
Choose DIMMs and Memory Mirroring
Select the memory configuration and whether or not you want the memory mirroring option.
The supported memory DIMMs and the mirroring option are listed in
When memory mirroring is enabled, the memory subsystem simultaneously writes identical data
to two adjacent channels. If a memory read from one of the channels returns incorrect data due
to an uncorrectable memory error, the system automatically retrieves the data from the other
channel. A transient or soft error in one channel does not affect the mirrored data, and
operation continues unless there is a simultaneous error in exactly the same location on a DIMM
and its mirrored DIMM. Memory mirroring reduces the amount of memory available to the
operating system by 50% because only one of the two populated channels provides data.
Table 4 Supported DDR4 DIMMs and Memory Mirroring Option
—The order in which DIMMs must be installed in a channel is Bank 1 (blue socket),
then Bank 2 (black socket), then Bank 3 (white socket).
—You cannot mix DIMM capacities within a bank
—You can mix DIMM capacities within a channel
Cisco UCS B420 M4 Blade Server
Page 13
—64GB TSV-RDIMMs cannot be mixed with any other DIMMs
—You cannot mix RDIMMs with LRDIMMs
■To optimize memory performance:
—Configure DIMMs identically for each CPU
—Fill banks equally across the CPU. See DIMM population guidelines in Table 27 on
page 41, Table 28 on page 42, and Table 29 on page 42.
—Populate less than 3 DPC
■For systems shipping with E5-4600 v3 series processor family CPUs, DIMM speeds are as
shown in
Table 5, Table 6, and Table 7.
Table 5 DIMM Speeds for 2133-MHz E5-4600 v3 Series CPUs
DPCDIMM Type
64 GB (TSV)16/32 GB (DR)8 GB (SR)
1DPC213321332133
CONFIGURING the SERVER
2DPC
3 DPC
213321332133
160018661600
Table 6 DIMM Speeds for 1866-MHz E5-4600 v3 Series CPUs
DPCDIMM Type
64 GB (TSV)16/32 GB (DR)8 GB (SR)
1DPC186618661866
2DPC
3 DPC
186618661866
133316001333
Table 7 DIMM Speeds for 1600-MHz E5-4600 v3 Series CPUs
DPCDIMM Type
1DPC160016001600
2DPC
Cisco UCS B420 M4 Blade Server
64 GB (TSV)16/32 GB (DR)8 GB (SR)
160016001600
13
Page 14
CONFIGURING the SERVER
Table 7 DIMM Speeds for 1600-MHz E5-4600 v3 Series CPUs (continued)
3 DPC106613331066
For systems shipping with E5-4600 v4 series processor family CPUs, DIMM speeds are as shown in Table 8.
Table 8 DIMM Speeds for E5-4600 v4 Series CPUs
DIMM and CPU
Frequencies
DIMM FREQ - 2400
CPU FREQ - 2400
DIMM FREQ - 2133
CPU FREQ - 2400
DIMM FREQ - 2400
CPU FREQ - 2133
DIMM FREQ - 2133
CPU FREQ - 2133
DIMMs per
Channel
LRDIMM speed
(quad rank x4)
64 GB 1.2 V
RDIMM speed
(dual rank x4)
32 GB 1.2 V
LRDIMM speed
(quad rank x4)
32 GB 1.2 V
LRDIMM speed
(single rank x4)
16 GB 1.2 V
1 DPC2400240024002400
2 DPC2400240024002400
3 DPC2133186621332133
1 DPCNA213321332133
2 DPCNA213321332133
3 DPCNA186618661866
1 DPC2133213321332133
2 DPC2133213321332133
3 DPC1866186618661866
1 DPCNA213321332133
2 DPCNA213321332133
DIMM FREQ - 2400
CPU FREQ - 1866
14
3 DPCNA186618661866
1 DPC1866186618661866
2 DPC1866186618661866
3 DPC1866160018661600
Cisco UCS B420 M4 Blade Server
Page 15
Table 8 DIMM Speeds for E5-4600 v4 Series CPUs (continued)
CONFIGURING the SERVER
DIMM and CPU
Frequencies
DIMM FREQ - 2133
CPU FREQ - 1866
For more information regarding memory, see DIMM and CPU Layout on page 40.
Supported Configurations
(1) Without memory mirroring:
■Select from 1 to 12 DIMMs per CPU (note that there are 12 DIMM slots per CPU). There must
be at least 1 DIMM per CPU.
(2) With memory mirroring:
■Select 4 DIMMs per CPU (installed in bank 1), 8 DIMMs per CPU (installed in banks 1, 2), or 12
DIMMs per CPU (installed in banks 1, 2, 3).
DIMMs per
Channel
LRDIMM speed
(quad rank x4)
64 GB 1.2 V
RDIMM speed
(dual rank x4)
32 GB 1.2 V
LRDIMM speed
(quad rank x4)
32 GB 1.2 V
LRDIMM speed
(single rank x4)
16 GB 1.2 V
1 DPCNA186618661866
2 DPCNA186618661866
3 DPCNA160018661600
Cisco UCS B420 M4 Blade Server
15
Page 16
CONFIGURING the SERVER
Notes
STEP 4CHOOSE RAID CONTROLLER AND DISK DRIVE BAYS
The UCS B420 M4 can be ordered with or without the Cisco FlexStorage RAID controller with disk
drive bays (local storage subsystem).
There is no RAID or local storage controller or drive bays included (embedded) with the B420 M4
server. Therefore, the Cisco FlexStorage RAID controller with drive bays must be purchased in
order to support any (1, 2, 3, or 4) local hard disk drives (HDDs) or solid state drives (SSDs). The
Cisco FlexStorage RAID controller option includes drive bays one and two, and supports RAID
0,1,10. An optional passthrough module includes drive bays three and four.
For servers that do not need any local HDDs or SSDs (for example, for booting from SAN), you can
order the server without the Cisco FlexStorage RAID controller with disk drive bays; however be
sure to keep Cisco FlexStorage blanking panels installed to maintain proper cooling airflow.
The ordering information is listed in Table 11.
Table 9 Card Cage and RAID Ordering Options
Product ID (PID)PID Description
UCSB-MRAID12G
1
UCSB-MRAID12G-HE
2, 3, 4
Cisco FlexStorage 12G SAS RAID controller with drive bays
Cisco FlexStorage 12G SAS RAID controller with 2 GB flash-backed write
cache and drive bays
1. The UCSB-MRAID-12G RAID controller runs iMegaRAID, which has a lower-capability software stack, provides
RAID 0/1 levels, and has no support for caching. Note that you cannot upgrade from this RAID controller to the
to UCSB-MRAID12G-HE RAID controller.
2. The UCSB-MRAID12G-HE RAID controller runs a MegaRAID fully-featured software stack, has access to cache,
and is recommended for VSAN. Its queue depth is much higher than that of the UCSB-MRAID-12G. Supports RAID
0/1/5/6 levels. Note that you cannot upgrade to this RAID controller from the UCSB-MRAID-12G controller
3. The B420 M4 with the UCSB-MRAID12G-HE requires UCSM 2.2(6) or later
4. The flash-backed write cache provides RAID controller cache protection using NAND flash memory and a
supercapacitor. In the event of a power or server failure, cached data is automatically transferred from the
RAID controller DRAM write cache to flash. Once power is restored, the data in the NAND flash is copied back
into the DRAM write cache until it can be flushed to the disk drives.
See Figure 4 on page 17 for a top view of the B420 M4 server with and without the Cisco
FlexStorage SAS RAID controller and drive bays. See Figure 5 on page 17 for a front view of
these configurations. When no drive bays are installed, blanking panels must be installed.
16
Cisco UCS B420 M4 Blade Server
Page 17
Figure 4 UCS B420 M4 With and Without Drive Bays (Top View)
RAID
Controller
(integrated
in drive bay)
Disk Drive
Bays (4)
(a) Server with four drive bays
(b) Server with no drive bays
Blanking Panels
(a) Server with four drive bays
(b) Server with no drive bays
Blanking Panels
Blanking Panel
CONFIGURING the SERVER
Figure 5 UCS B420 M4 With and Without Drive Bays (Front View)
Cisco UCS B420 M4 Blade Server
17
Page 18
CONFIGURING the SERVER
Notes
Supported Configurations
1
(1) Four Drive Bays with RAID 0, 1, 10
■Choose the following:
—One UCSB-MRAID12G. This provides two drive bays installed on the left side of the
blade server (viewed from the front). The RAID controller is integrated in the drive
bays and provides RAID 0,1,10.
—One UCSB-LSTOR-PT. This provides two drive bays installed on the right side of the
blade server (viewed from the front) and includes a passthrough connector that
allows the drives to be managed from the UCSB-MRAID12G RAID controller (RAID
0,1,10).
(2) Four Drive Bays with RAID 0, 1, 10, 5, 6
■Choose the following:
—One UCSB-MRAID12G-HE. This provides two drive bays installed on the left side of
the blade server (viewed from the front). The RAID controller includes a 2 GB
flash-backed write cache (FBWC)
bays, and provides RAID 0, 1,10, 5, 6, 50, 60.
—One UCSB-LSTOR-PT. This provides two drive bays installed on the right side of the
blade server (viewed from the front) and includes a passthrough connector that
allows the drives to be managed from the UCSB-MRAID12G-HE RAID controller (RAID
0, 1,10, 5, 6).
2
for higher performance, is integrated in the drive
(3) Two Drive Bays with RAID 0, 1, 10
■Choose the following:
—One UCSB-MRAID12G. This provides two drive bays installed on the left side of the
blade server (viewed from the front). The RAID controller is integrated in the drive
bays and provides RAID 0,1,10. The two right-hand bays are fitted with blanking
panels to maintain proper cooling airflow.
(4) Two Drive Bays with RAID 0, 1, 10, 5, 6
■Choose the following:
—One UCSB-MRAID12G-HE. This provides two drive bays installed on the left side of
the blade server (viewed from the front). The RAID controller includes a 2 GB
flash-backed write cache for higher performance, is integrated in the drive bays and
provides RAID 0, 1,10, 5, 6. The two right-hand bays are fitted with blanking panels
to maintain proper cooling airflow.
(5) No Drive Bays
■System is installed with blanking panels.
1. Any empty drive bays must be fitted with blanking panels to maintain proper cooling airflow.
2. The flash-backed write cache provides RAID controller cache protection using NAND flash memory and a
supercapacitor. In the event of a power or server failure, cached data is automatically transferred from the
RAID controller DRAM write cache to flash. Once power is restored, the data in the NAND flash is copied back
into the DRAM write cache until it can be flushed to the disk drives.
18
Cisco UCS B420 M4 Blade Server
Page 19
CONFIGURING the SERVER
STEP 5CHOOSE HARD DISK DRIVES or SOLID-STATE DRIVES (OPTIONAL)
The UCS B420 M4 can be ordered with or without drives. If you ordered one of the RAID
controller with drive bay options in
16 (UCSB-MRAID12G or UCSB-MRAID12G-HE for drive bays one and two, and optionally
UCSB-LSTOR-PT for drive bays three and four), you can order drives listed in this section. The
B420 M4 provides up to four hot plug 2.5” SFF drive bays.
Choose Drives
The supported drives in the UCS B420 M4 are listed in Table 10.
NOTE: 4K format drives are supported and qualified as bootable with Cisco UCS
Manager Release 3.1(2b)and later versions.
Table 10 Supported Hot Plug Drives
CHOOSE RAID CONTROLLER AND DISK DRIVE BAYS on page
1.6 TB 2.5 inch Enterprise Performance 12G SAS SSD
(10X endurance)
800 GB 2.5 inch Enterprise Performance 12G SAS SSD
(10X endurance)
400 GB 2.5 inch Enterprise Performance 12G SAS SSD
(10X endurance)
Drive
Type
SAS1.6 TB
SAS800 GB
SAS400 GB
Capacity
Cisco UCS B420 M4 Blade Server
19
Page 20
CONFIGURING the SERVER
Table 10 Supported Hot Plug Drives (continued)
Product ID (PID)PID Description
Drive
Type
Capacity
6 Gbps Drives
UCS-SD16TBKS4-EV1.6 TB 2.5 inch Enterprise Value 6G SATA SSD (Intel 3510)SATA 1.6 TB
UCS-SD480GBKS4-EV480 GB 2.5 inch Enterprise Value 6G SATA SSD (Intel 3510)SATA480 GB
UCS-SD240GBKS4-EV240 GB 2.5 inch Enterprise Value 6G SATA SSDSATA 240 GB
UCS-SD120GBKS4-EV120 GB 2.5 inch Enterprise Value 6G SATA SSD (Intel 3510)SATA120 GB
Supported Configurations
■Select up to 4 of the drives listed in Table 10. This is dependent on the number of drive
bays selected in CHOOSE RAID CONTROLLER AND DISK DRIVE BAYS on page 16.
Notes
■When creating a RAID volume, mixing different capacity drives causes the system to use the
lowest-capacity drive.
■Mixing of drive types is supported, but performance may be impacted.
■Multiple RAID volumes are supported. RAID volumes should use the same media type.
20
Cisco UCS B420 M4 Blade Server
Page 21
STEP 6CHOOSE ADAPTERS
The adapter offerings are:
■Cisco Virtual Interface Cards (VICs)
Cisco developed 1300 Series and 1200 Series Virtual Interface Cards (VICs) to provide flexibility
to create multiple NIC and HBA devices. The VICs also support adapter Fabric Extender and
Virtual Machine Fabric Extender technologies. The VIC features are listed here:
—1200 Series VICs enable advanced networking features including Netflow for
network statistics, and DPDK, USNIC for low-latency computing applications.
—1300 Series VICs include all of the 1200 Series features plus additional
enhancements including network overlay offload support for NVGRE and VXLAN, and
RoCE services.
—In addition, 1300 Series VICs support PCIe Gen 3.0 for greater bandwidth than 1200
Series VICs
—Two Converged Network Adapter (CNA) ports, supporting both Ethernet and FCoE
—Delivers 160 Gbs total I/O throughput to the server, for example:
CONFIGURING the SERVER
• VIC 1240 supports 4 x 10 Gbs Unified I/O ports, Port Expander supports 4 x 10 Gbs
Unified I/O ports, and VIC 1280 supports 8 x 10 Gbs Unified ports, for a total of
160 Gbs (see
Figure 26 on page 56).
• VIC 1340 supports dual 4 x 10 Gbs Unified I/O ports, Port Expander enables
additional ports on the VIC 1340 to create either dual 4 x 10 Gbs or dual native
40G Unified I/O ports, and VIC 1380 supports 8 x 10 Gbs Unified ports, for a total
of 160 Gbs (see
Figure 26 on page 56).
—Creates up to 256 fully functional unique and independent PCIe adapters and
interfaces (NICs or HBAs) without requiring single-root I/O virtualization (SR-IOV)
support from operating systems or hypervisors
—Provides virtual machine visibility from the physical network and a consistent
network operations model for physical and virtual servers
—Supports customer requirements for a wide range of operating systems and
hypervisors
■Cisco UCS Storage Accelerator Adapters
Cisco UCS Storage Accelerator adapters are designed specifically for the Cisco UCS B-series M4
blade servers and integrate seamlessly to allow improvement in performance and relief of I/O
bottlenecks.
Table 11 shows the supported adapters.
To help ensure that your operating system is compatible with the cards you have selected,
please check the Hardware Compatibility List at this URL:
UCSB-MLOM-PT-01Cisco UCS Port Expander Card for VIC. This is a hardware option
Adapter 2
to enable an additional 4 ports of the VIC 1340 or VIC 1240,
bringing the total capability of the VIC 1340 or VIC 1240 to 8 x 10
GbE or a dual native 40G interface for the VIC 1340.
Cisco Storage Accelerators1,
2
UCSB-F-FIO-1600MSUCS 1600 GB Fusion ioMemory3 SX Scale line for B-SeriesAdapter 2 or
Adapter 3
UCSB-F-FIO-1300MPUCS 1300 GB Fusion ioMemory3 PX Performance line for B-SeriesAdapter 2 or
Adapter 3
1. Fusion io storage accelerators must be identical if you choose two of them.
2. As of UCSM version 2.2(4b), the Cisco UCS Storage Accelerators are capable of providing additional ports for
increased bandwidth if placed in adapter slot 2, similar to the Cisco UCS Port Expander Card for VIC.
Supported Configurations
Table 12 on page 23 and Table 13 on page 24 shows the supported adapter combinations. The
configuration rules are summarized as follows:
■You must select at least one VIC.
■Adapter slot 1 is dedicated for the VIC 1240 or VIC 1340 only. No other mezzanine card can
fit in Adapter Slot 1.
■The Port Expander Card can only be selected if the VIC 1240 or VIC 1340 is also selected for
the server.
■You can select up to two Storage Acceleration adapters.
22
Cisco UCS B420 M4 Blade Server
Page 23
Select an adapter combination according to Table 12 or Table 13 on page 24.
Notes
NOTE: CPU1 controls adapter slot 1, CPU 2 controls adapter slot 3, and CPU 4
controls adapter slot 2.
Table 12 Supported Adapter Combinations for 4 CPUs Installed
CONFIGURING the SERVER
Adapter Configurations
Adapter Slot 1Adapter Slot 2 Adapter Slot
3
Fabric Extenders Aggregate Bandwidth
(Gb/s)
1
2 x 2304
2 x
2208XP
2 x
2204XP
2 x
2104XP
Fabric
Interconnects
VIC 1340 Combinations
VIC 1340not populated not
populated
VIC 1340Port Expander
Card
not
populated
VIC 1340not populated VIC 138012012060no
VIC 1340Port Expander
VIC 1380
Card
not
not populated VIC 1380
populated
VIC 1340Cisco UCS
Storage
not
populated
404020no
support
80
2
80
2
40no
support
support
160
2
160
1
80no
support
80
3
80
1
40no
support
80
3
80
4
40no
support
2 x 62xx or
2 x 63xx
2 x 62xx or
2 x 63xx
2 x 62xx or
2 x 63xx
2 x 62xx or
2 x 63xx
2 x 62xx or
2 x 63xx
2 x 62xx or
2 x 63xx
Accelerator
(Fusion-io)
VIC 1340not populated Cisco UCS
Storage
404020no
support
2 x 62xx or
2 x 63xx
Accelerator
(Fusion-io)
Note: for the configuration in the following row, slots 2 and 3 must have identical types of storage
cards.
VIC 1340Cisco UCS
Storage
Accelerator
(Fusion-io)
Cisco UCS
Storage
Accelerator
(Fusion-io)
80
3
80
3
40402 x 62xx or
2 x 63xx
VIC 1240 Combinations
VIC 1240 not populated not
populated
404020202 x 61xx or
2 x 62xx or
2 x 63xx
VIC 1240Port Expander
Card
not
populated
808040no
support
2 x 62xx or
2 x 63xx
VIC 1240 not populated VIC 128012012060402 x 61xx or
2 x 62xx or
2 x 63xx
VIC 1240 Port Expander
Card
VIC 128016016080no
support
2 x 62xx or
2 x 63xx
1. 2304XP is only supported with B420 M4 v4 CPUs starting with UCSM 3.1(2)
Cisco UCS B420 M4 Blade Server
23
Page 24
CONFIGURING the SERVER
Notes
2. Dual 40G via 4x10 port-channel configuration.
3. Dual native 40G Interface
4. If running 2.2(4b) or greater and equipped with either storage accelerator SKU UCSB-F-FIO-1300MP or
UCSB-F-FIO-1600MS, then there is a dual 40G 4x10 port-channel interface for the 2208 or a dual native 40G
interface for the 2304. Otherwise, the bandwidth is 20G (2x10G port-channel) per FI.
Table 13 Supported Adapter Combinations for 2 CPUs Installed
Adapter Configurations
Adapter Slot 1Adapter Slot
1
2
VIC 1340not
populated
VIC 1340Port
Expander
Card
VIC 1340Port
Expander
Card
VIC 1340not
populated
VIC 1340not
populated
VIC 1240not
populated
VIC 1240not
populated
VIC 1240Port
Expander
Card
Fabric Extenders Aggregate
Bandwidth
Adapter Slot
2 x 2304
3
not
404020no
populated
not
80
populated
VIC 1380160
VIC 1380
Cisco UCS
120
40
Storage
2
2 x
2208XP
2 x
2204XP
2 x
2104XP
support
3
80
3
40no
support
2
160
80no
support
4
12060no
support
4
40
5
20no
support
Fabric
Interconnects
2 x 62xx or
2 x 63xx
2 x 62xx or
2 x 63xx
2 x 62xx or
2 x 63xx
2 x 62xx or
2 x 63xx
2 x 62xx or
2 x 63xx
Accelerator
(Fusion-io)
not
populated
404020202 x 61xx or
2 x 62xx or
2 x 63xx
VIC 128012012060402 x 61xx or
2 x 62xx or
2 x 63xx
not
populated
808040no
support
2 x 62xx or
2 x 63xx
1. Adapter slot 2 is not available except for the Port Expander card.
2. 2304XP is only supported with B420 M4 v4 CPUs starting with UCSM 3.1(2)
3. Dual 40G via 4x10 port-channel configuration
4. Dual native 40G interface
5. If running 2.2(4b) or greater and equipped with either storage accelerator SKU UCSB-F-FIO-1300MP or
UCSB-F-FIO-1600MS, then there is a dual 40G 4x10 port-channel interface for the 2208 or a dual native 40G
interface for the 2304. Otherwise, the bandwidth is 20G (2x10G port-channel) per FI.
24
Cisco UCS B420 M4 Blade Server
Page 25
CONFIGURING the SERVER
STEP 7ORDER A TRUSTED PLATFORM MODULE (OPTIONAL)
Trusted Platform Module (TPM) is a computer chip (microcontroller) that can securely store
artifacts used to authenticate the platform (server). These artifacts can include passwords,
certificates, or encryption keys. A TPM can also be used to store platform measurements that
help ensure that the platform remains trustworthy. Authentication (ensuring that the platform
can prove that it is what it claims to be) and attestation (a process helping to prove that a
platform is trustworthy and has not been breached) are necessary steps to ensure safer
computing in all environments.
The TPM ordering information is listed in Table 14.
Table 14 Trusted Platform Module
Product ID (PID)PID Description
UCSX-TPM2-001Tru s te d Pl a tf o rm M od u l e for UCS (SPI-based)
UCSX-TPM2-002Trusted Platform Module 2.0 for UCS servers
NOTE: TPM installation is supported after-factory. However, a TPM installs with a
one-way screw and cannot be replaced, upgraded, or moved to another server. If a
server with a TPM is returned, the replacement server must be ordered with a new
TPM. If there is no existing TPM in the server, you can install TPM 2.0. You must first
upgrade to UCS firmware that supports Intel E5-4600 v4 CPUs, which is Cisco UCS
Manager Release 2.2(8) and later or Release 3.1(2) and later (because Cisco aligned
support for TPM 2.0 with these CPUs).
.
Table 15 Trusted Platform Module \ Support Matrix by Intel CPU Version
Intel CPUTPM VersionTPM Version
Intel E5-4600 v3TPM 1.2Release 2.2(5)
TPM 2.0Release 2.2(8) or Release 3.1(2)
Intel E5-4600 v4TPM 1.2Release 2.2(8) or Release 3.1(2)
TPM 2.0Release 2.2(8) or Release 3.1(2)
Cisco UCS B420 M4 Blade Server
25
Page 26
CONFIGURING the SERVER
STEP 8ORDER CISCO FLEXIBLE FLASH SECURE DIGITAL CARDS
Dual SDHC flash card sockets are provided on the front left side of the server. Mirroring of two
SDHC cards is supported.
The SDHC card ordering information is listed in Table 16.
Table 16 PIDs for Secure Digital Card(s)
Product ID (PID)PID Description
UCS-SD-64G-S64 GB SD Card for UCS servers
UCS-SD-32G-S32 GB SD Card for UCS servers
Supported Configurations
(1) Select one or two Cisco Flexible Flash secure digital cards
■Select up to two 32 GB SD cards or two 64 GB SD cards
(2) Do not mix 64 GB cards and 32 GB cards
26
Cisco UCS B420 M4 Blade Server
Page 27
STEP 9ORDER INTERNAL USB 3.0 DRIVE (OPTIONAL)
You can order one optional internal USB 3.0 drive. The USB drive ordering information is listed in
Table 17.
Table 17 USB 3.0 Drive
Product ID (PID)PID Description
UCS-USBFLSHB-16GBUCS Servers 16 GB Flash USB Drive
CONFIGURING the SERVER
Cisco UCS B420 M4 Blade Server
27
Page 28
CONFIGURING the SERVER
STEP 10 CHOOSE OPERATING SYSTEM AND VALUE-ADDED SOFTWARE
Several software programs are available. Select as desired from Table 18.
Table 18 OSs and Value-Added Software
PID DescriptionProduct ID (PID)
Cisco One
C1F2PUCSK9Cisco ONE Foundation Perpetual UCS
C1A1PUCSK9Cisco ONE Enterprise Cloud Perpetual UCS
C1UCS-OPT-OUTCisco One Data Center Compute Opt Out Option
Energy Management (JouleX)
CEM-DC-PERPerpetual License Key for Cisco Energy Management for DC
UCS Director
CUIC-PHY-SERV-BM-UCisco UCS Director Resource Lic - 1 Phy Sevr node bare metal
CUIC-PHY-SERV-UCisco UCS Director Resource Lic - One physical Server node
CUIC-TERMAcceptance of Cisco UCS Director License Terms
UCS Performance Manager
UCS-PM-IEUCS Performance Manager
UCS-PM-EEUCS Performance Manager Express
EVAL-UCS-PM-IEUCS Performance Manager - 90 days evaluation
EVAL-UCS-PM-EEUCS Performance Manager Express - 90 days evaluation
Nexus 1000V for Hyper-V and vSphere
N1K-VSG-UCS-BUNNexus 1000V Adv Edition for vSphere Paper License Qty 1
IMC Supervisor
CIMC-SUP-B10IMC Supervisor- Branch Mgt SW for C & E-Series up to 1K Svrs
CIMC-SUP-B02IMC Supervisor- Branch Mgt SW for C & E-Series up to 250 Svrs
UCS Multi-Domain Manager
UCS-MDMGR-100SUCS Multi-Domain Manager (Central) Per Server License (100+)
UCS-MDMGR-50SUCS Multi-Domain Manager (Central) Per Server License (50+)
UCS-MDMGR-1SUCS Multi-Domain Manager (Central) Per Server License
UCS-MDMGR-10SUCS Multi-Domain Manager (Central) Per Server License (10+)
UCS-MDMGR-1DMNUCS Multi-Domain Manager (Central) Per Domain License
VMware vCenter
VMW-VCS-STD-1AVMware vCenter 6 Server Standard, 1 yr support required
VMW-VCS-STD-3AVMware vCenter 6 Server Standard, 3 yr support required
VMW-VCS-STD-5AVMware vCenter 6 Server Standard, 5 yr support required
VMW-VCS-FND-1AVMware vCenter 6 Server Foundation (3 Host), 1 yr supp reqd
VMW-VCS-FND-3AVMware vCenter 6 Server Foundation (3 Host), 3 yr supp reqd
VMW-VCS-FND-5AVMware vCenter 6 Server Foundation (3 Host), 5 yr supp reqd
28
Cisco UCS B420 M4 Blade Server
Page 29
CONFIGURING the SERVER
Table 18 OSs and Value-Added Software (continued)
PID DescriptionProduct ID (PID)
Microsoft Windows Server
MSWS-12-ST2SWindows Server 2012 Standard (2 CPU/2 VMs)
MSWS-12-DC2SWindows Server 2012 Datacenter (2 CPU/Unlimited VMs)
MSWS-12-ST2S-NSWindows Server 2012 Standard (2 CPU/2 VMs) No Cisco Svc
MSWS-12-DC2S-NSWindows Server 2012 Datacenter (2 CPU/Unlim VM) No Cisco Svc
MSWS-12R2-ST2SWindows Server 2012 R2 Standard (2 CPU/2 VMs)
MSWS-12R2-DC2SWindows Server 2012 R2 Datacenter (2 CPU/Unlimited VMs)
MSWS-12R2-ST2S-NSWindows Server 2012 R2 Standard (2 CPU/2 VMs) No Cisco SVC
MSWS-12R2-DC2S-NSWindows Server 2012 R2 Datacen (2 CPU/Unlim VM) No Cisco Svc
Red Hat
RHEL-2S2V-3ARed Hat Enterprise Linux (1-2 CPU,1-2 VN); 3-Yr Support Req
RHEL-2S2V-1ARed Hat Enterprise Linux (1-2 CPU,1-2 VN); 1-Yr Support Req
VMware
VMW-VSP-EPL-5AVMware vSphere 6 Ent Plus (1 CPU), 5-yr, Support Required
VMW-VSP-STD-3AVMware vSphere 6 Standard (1 CPU), 3-yr, Support Required
VMW-VSP-STD-1AVMware vSphere 6 Standard (1 CPU), 1-yr, Support Required
VMW-VSP-EPL-3AVMware vSphere 6 Ent Plus (1 CPU), 3-yr, Support Required
VMW-VSP-EPL-1AVMware vSphere 6 Ent Plus (1 CPU), 1-yr, Support Required
VMW-VSP-STD-5AVMware vSphere 6 Standard (1 CPU), 5-yr, Support Required
SLES SAP
SLES-SAP-2S2V-1SSLES for SAP Apps (1-2 CPU, 1-2 VM); Priority 1-Yr SnS
SLES-SAP-2SUV-1SSLES for SAP Apps (1-2 CPU, Unl VM); Priority 1-Yr SnS
SLES-SAP-2S2V-3SSLES for SAP Apps (1-2 CPU, 1-2 VM); Priority 3-Yr SnS
SLES-SAP-2SUV-3SSLES for SAP Apps (1-2 CPU, Unl VM); Priority 3-Yr SnS
SLES-SAP-2S2V-5SSLES for SAP Apps (1-2 CPU, 1-2 VM); Priority 5-Yr SnS
SLES-SAP-2SUV-5SSLES for SAP Apps (1-2 CPU, Unl VM); Priority 5-Yr SnS
SLES-SAP-2S2V-5ASLES for SAP Apps (1-2 CPU, 1-2 VM); 5-Yr Support Reqd
SLES-SAP-2SUV-3ASLES for SAP Apps (1-2 CPU, Unl VM); 3-Yr Support Reqd
SLES-SAP-2S2V-3ASLES for SAP Apps (1-2 CPU, 1-2 VM); 3-Yr Support Reqd
SLES-SAP-2SUV-5ASLES for SAP Apps (1-2 CPU, Unl VM); 5-Yr Support Reqd
SLES-SAP-2S2V-1ASLES for SAP Apps (1-2 CPU, 1-2 VM); 1-Yr Support Reqd
SLES-SAP-2SUV-1ASLES for SAP Apps (1-2 CPU, Unl VM); 1-Yr Support Reqd
SUSE
SLES-2S2V-1ASUSE Linux Enterprise Svr (1-2 CPU,1-2 VM); 1-Yr Support Req
SLES-2SUV-1A SUSE Linux Enterprise Svr (1-2 CPU,Unl VM); 1-Yr Support Req
SLES-2S2V-3A SUSE Linux Enterprise Svr (1-2 CPU,1-2 VM); 3-Yr Support Req
SLES-2SUV-3A SUSE Linux Enterprise Svr (1-2 CPU,Unl VM); 3-Yr Support Req
Cisco UCS B420 M4 Blade Server
29
Page 30
CONFIGURING the SERVER
Table 18 OSs and Value-Added Software (continued)
PID DescriptionProduct ID (PID)
SLES-2S2V-5A SUSE Linux Enterprise Svr (1-2 CPU,1-2 VM); 5-Yr Support Req
SLES-2SUV-5A SUSE Linux Enterprise Svr (1-2 CPU,Unl VM); 5-Yr Support Req
SLES-2S-HA-1S SUSE Linux High Availability Ext (1-2 CPU); 1yr SnS
SLES-2S-HA-3S SUSE Linux High Availability Ext (1-2 CPU); 3yr SnS
SLES-2S-HA-5S SUSE Linux High Availability Ext (1-2 CPU); 5yr SnS
SLES-2S-GC-1S SUSE Linux GEO Clustering for HA (1-2 CPU); 1yr Sns
SLES-2S-GC-3S SUSE Linux GEO Clustering for HA (1-2 CPU); 3yr SnS
SLES-2S-GC-5S SUSE Linux GEO Clustering for HA (1-2 CPU); 5yr SnS
30
Cisco UCS B420 M4 Blade Server
Page 31
CONFIGURING the SERVER
STEP 11 CHOOSE OPERATING SYSTEM MEDIA KIT (OPTIONAL)
Choose the optional operating system media listed in Table 19.
Table 19 OS Media
Product ID (PID)PID Description
RHEL-6RHEL 6 Recovery Media Only (Multilingual)
SLES-11SLES 11 media only (multilingual)
MSWS-12-ST2S-RMWindows Server 2012 Standard (2 CPU/2 VMs) Recovery Media
MSWS-12-DC2S-RMWindows Server 2012 Datacenter (2 CPU/Unlimited VM) Rec Media
MSWS-12R2-ST2S-RMWindows Server 2012 R2 Standard (2 CPU/2 VMs) Recovery Media
MSWS-12R2-DC2S-RMWindows Server 2012 R2 Datacen(2 CPU/Unlimited VM) Rec Media
Cisco UCS B420 M4 Blade Server
31
Page 32
CONFIGURING the SERVER
STEP 12 CHOOSE SERVICE and SUPPORT LEVEL
A variety of service options are available, as described in this section.
Unified Computing Warranty, No Contract
If you have noncritical implementations and choose to have no service contract, the following
coverage is supplied:
■Three-year parts coverage.
■Next business day (NBD) onsite parts replacement eight hours a day, five days a week.
■90-day software warranty on media.
■Ongoing downloads of BIOS, drivers, and firmware updates.
■UCSM updates for systems with Unified Computing System Manager. These updates include
minor enhancements and bug fixes that are designed to maintain the compliance of UCSM
with published specifications, release notes, and industry standards.
SMARTnet for UCS
For support of the entire Unified Computing System, Cisco offers the Cisco SMARTnet for UCS
Service. This service provides expert software and hardware support to help sustain
performance and high availability of the unified computing environment. Access to Cisco
Technical Assistance Center (TAC) is provided around the clock, from anywhere in the world.
For UCS blade servers, there is Smart Call Home, which provides proactive, embedded
diagnostics and real-time alerts. For systems that include Unified Computing System Manager,
the support service includes downloads of UCSM upgrades. The Cisco SMARTnet for UCS Service
includes flexible hardware replacement options, including replacement in as little as two hours.
There is also access to Cisco's extensive online technical resources to help maintain optimal
efficiency and uptime of the unified computing environment. You can choose a desired service
listed in
Table 20.
Table 20 Cisco SMARTnet for UCS Service
Product ID (PID)On Site?Description
CON-PREM-B420M4YesONSITE 24X7X2 UCS B420 M4 Blade Server
CON-OSP-B420M4YesONSITE 24X7X4 UCS B420 M4 Blade Server
CON-OSE-B420M4Ye sONSITE 8X5X4 UCS B420 M4 Blade Server
CON-OS-B420M4Ye sONSITE 8X5XNBD UCS B420 M4 Blade Server
CON-S2P-B420M4NoSMARTNET 24X7X2 UCS B420 M4 Blade Server
CON-SNTP-B420M4NoSMARTNET 24X7X4 UCS B420 M4 Blade Server
CON-SNTE-B420M4NoSMARTNET 8X5X4 UCS B420 M4 Blade Server
CON-SNT-B420M4NoSMARTNET 8X5XNBD UCS B420 M4 Blade Server
32
Cisco UCS B420 M4 Blade Server
Page 33
CONFIGURING the SERVER
SMARTnet for UCS Hardware Only Service
For faster parts replacement than is provided with the standard Cisco Unified Computing System
warranty, Cisco offers the Cisco SMARTnet for UCS Hardware Only Service. You can choose from
two levels of advanced onsite parts replacement coverage in as little as four hours. SMARTnet
for UCS Hardware Only Service provides remote access any time to Cisco support professionals
who can determine if a return materials authorization (RMA) is required. You can choose a
service listed in
Table 21.
Table 21 SMARTnet for UCS Hardware Only Service
Product ID (PID)
Service
Level GSP
On
Site?
Description
CON-UCW7-B420M4UCW7Ye sUC PLUS 24X7X4OS UCS B420 M4 Blade Server
CON-UCW5-B420M4UCW5Ye sUC PLUS 8X5XNBDOS UCS B420 M4 Blade
Server
Unified Computing Partner Support Service
Cisco Partner Support Service (PSS) is a Cisco Collaborative Services service offering that is
designed for partners to deliver their own branded support and managed services to enterprise
customers. Cisco PSS provides partners with access to Cisco's support infrastructure and assets
to help them:
■Expand their service portfolios to support the most complex network environments
■Lower delivery costs
■Deliver services that increase customer loyalty
Partner Unified Computing Support Options enable eligible Cisco partners to develop and
consistently deliver high-value technical support that capitalizes on Cisco intellectual assets.
This helps partners to realize higher margins and expand their practice.
PSS is available to all Cisco PSS partners, but requires additional specializations and
requirements. For additional information, see the following URL:
www.cisco.com/go/partnerucssupport
The two Partner Unified Computing Support Options include:
■Partner Support Service for UCS
■Partner Support Service for UCS Hardware Only
Partner Support Service for UCS provides hardware and software support, including triage
support for third party software, backed by Cisco technical resources and level three support.
CON-PSJ2-B420M4 PSJ2NoUCS SUPP PSS 8X5X4 UCS B420 M4 Blade Server
CON-PSJ3-B420M4 PSJ3NoUCS SUPP PSS 24X7X4 UCS B420 M4 Blade Server
CON-PSJ4-B420M4 PSJ4NoUCS SUPP PSS 24X7X2 UCS B420 M4 Blade Server
Partner Support Service for UCS Hardware Only provides customers with replacement parts in as
little as two hours. See
Table 23.
Table 23 Partner Support Service for UCS (Hardware Only)
Product ID (PID)
Service
Level
GSP
On
Site?
Description
CON-PSW2-B420M4 PSW2NoUCS W PL PSS 8X5X4 UCS B420 M4 Blade Server
CON-PSW3-B420M4PSW3NoUCS W PL PSS 24X7X4 UCS B420 M4 Blade Server
CON-PSW4-B420M4 PSW4NoUCS W PL PSS 24X7X2 UCS B420 M4 Blade Server
Unified Computing Combined Support Service
Combined Services makes it easier to purchase and manage required services under one
contract. SMARTnet services for UCS help increase the availability of your vital data center
infrastructure and realize the most value from your unified computing investment. The more
benefits you realize from the Cisco Unified Computing System (Cisco UCS), the more important
the technology becomes to your business. These services allow you to:
■Optimize the uptime, performance, and efficiency of your UCS
■Protect your vital business applications by rapidly identifying and addressing issues
■Strengthen in-house expertise through knowledge transfer and mentoring
■Improve operational efficiency by allowing UCS experts to augment your internal staff
resources
■Enhance business agility by diagnosing potential issues before they affect your operations
With the Cisco Unified Computing Drive Retention (UCDR) Service, you can obtain a new disk
drive in exchange for a faulty drive without returning the faulty drive. In exchange for a Cisco
replacement drive, you provide a signed Certificate of Destruction (CoD) confirming that the
drive has been removed from the system listed, is no longer in service, and has been destroyed.
On
Site?
No
Yes
Yes
Yes
Yes
No
No
No
Description
CMB SPT SVC 24X7X2 UCS B420 M4 Blade Server
CMB SPT SVC 24X7X2OS UCS B420 M4 Blade Server
CMB SPT SVC 24X7X4OS UCS B420 M4 Blade Server
CMB SPT SVC 8X5X4OS UCS B420 M4 Blade Server
CMB SPT SVC 8X5XNBDOS UCS B420 M4 Blade Server
CMB SPT SVC 8X5X4 UCS B420 M4 Blade Server
CMB SPT SVC 24X7X4 UCS B420 M4 Blade Server
CMB SPT SVC 8X5XNBD UCS B420 M4 Blade Server
Sophisticated data recovery techniques have made classified, proprietary, and confidential
information vulnerable, even on malfunctioning disk drives. The UCDR service enables you to
retain your drives and ensures that the sensitive data on those drives is not compromised, which
reduces the risk of any potential liabilities. This service also enables you to comply with
regulatory, local, and federal requirements.
If your company has a need to control confidential, classified, sensitive, or proprietary data, you
might want to consider one of the Drive Retention Services listed in
NOTE: Cisco does not offer a certified drive destruction service as part of this
service.
Cisco UCS B420 M4 Blade Server
Table 25.
35
Page 36
CONFIGURING the SERVER
Table 25 Drive Retention Service Options
Service Description
SMARTnet for UCS
Service with Drive
Retention
SMARTnet for UCS
HW ONLY+Drive
Retention
For more service and support information, see the following URL:
The local KVM I/O cable ships with every UCS 5100 Series blade chassis accessory kit. The cable provides a
connection into the server, providing a DB9 serial connector, a VGA connector for a monitor, and dual USB
ports for a keyboard and mouse. With this cable, you can create a direct connection to the operating
system and the BIOS running on the server.
The local KVM I/O cable ordering information is listed in Table 26.
Table 26 Local KVM I/O Cable
Product ID (PID)PID Description
N20-BKVM=Local KVM I/O cable for UCS servers console port
Figure 6 Local KVM I/O Cable
1Connector (to server front panel)3VGA connector (for a monitor)
2DB-9 serial connector4Two-port USB connector (for a mouse and
keyboard)
NOTE: *The blade chassis ships with the local KVM I/O cable.
Cisco UCS B420 M4 Blade Server
37
Page 38
SUPPLEMENTAL MATERIAL
SUPPLEMENTAL MATERIAL
System Board
A top view of the UCS B420 M4 system board is shown in Figure 7.
Figure 7 UCS B420 M4 System Board
2
3
4
1
3
1
5
6610
7
66
66
77
66
8
9
8
8
8
7
11
12
11
305015
1Ejector thumbscrew7Heat sink and CPU (underneath)
2SD card slots8CPU heat sink install guide pins
3Modular storage subsystem connector9Trusted Platform Module (TPM)
4USB memory10Adapter 1 slot supports the Cisco VIC
1340/1240 adapter only
38
Cisco UCS B420 M4 Blade Server
Page 39
5CMOS battery11Adapter cards:
■ Adapter 2 is the slot on the left (when
facing the server) and partially covers
Adapter 1
■ Adapter 3 is the slot on the right
(facing the server)
6DIMM slots12Diagnostic button
SUPPLEMENTAL MATERIAL
Cisco UCS B420 M4 Blade Server
39
Page 40
SUPPLEMENTAL MATERIAL
CPU 1CPU 2
48 DIMMS
4 memory channels per CPU
3 DIMMs per channel
CPU 3CPU 4
Channel A
Channel B
Channel C
Channel D
Channel E
Channel F
Channel G
Channel H
Channel M
Channel N
Channel O
Channel P
Channel I
Channel J
Channel K
Channel L
A1A2A3
B3
C3
D3
I3
J3
K3
I3
M3
N3
O3
P3
E3
F3
G3
H3
B1B2
C1C2
D1D2
Bank 1
Bank 2
Bank 3
E1 E2
F1 F2
G1 G2
H1 H2
Bank 1
Bank 2
Bank 3
Bank 1
Bank 2
Bank 3
M1 M2
N1 N2
O1 O2
P1 P2
I1I2
J1J2
K1K2
I1I2
Bank 1
Bank 2
Bank 3
DIMM and CPU Layout
Memory is organized as shown in Figure 8.
Figure 8 UCS B420 M4 Memory Organization
40
Cisco UCS B420 M4 Blade Server
Page 41
Each CPU controls four memory channels and 12 DIMM slots, as follows:
■CPU1: Channels A, B, C, and D
—Bank 1 - A1, B1, C1, and D1 (blue DIMM slots)
—Bank 2 - A2, B2, C2, and D2 (black DIMM slots)
—Bank 3 - A3, B3, C3, and D3 (white DIMM slots)
■CPU2: Channels E, F, G, and H
—Bank 1 - E1, F1, G1, and H1 (blue DIMM slots)
—Bank 2 - E2, F2, G2, and H2 (black DIMM slots)
—Bank 3 - E3, F3, G3, and H3 (white DIMM slots)
■CPU3: Channels I, J, K, and L
—Bank 1 - I1, J1, K1, and L1 (blue DIMM slots)
—Bank 2 - I2, J2, K2, and L2 (black DIMM slots)
—Bank 3 - I3, J3, K3, and L3 (white DIMM slots)
■CPU4: Channels M, N, O, and P
SUPPLEMENTAL MATERIAL
—Bank 1 - M1, N1, O1, and P1 (blue DIMM slots)
—Bank 2 - M2, N2, O2, and P2 (black DIMM slots)
—Bank 3 - M3, N3, O3, and P3 (white DIMM slots)
DIMM Population Order
Populate the DIMMs for a CPU according to Table 27.
Table 27 DIMM Population Order per CPU (continued)
DIMMs
per CPU
8A1, B1, C1, D1,
9A1, B1, C1,
Populate
CPU 1 Slots
A2, B2, C2, D2
A2, B2, C2,
A3, B3, C3
Populate
CPU 2 Slots
E1, F1, G1, H1,
E2, F2, G2, H2
E1, F1, G1,
E2, F2, G2,
E3, F3, G3
Populate
CPU 3 Slots
I1, J1, K1, L1,
I2, J2, K2, L2
I1, J1, K1,
I2, J2, K2,
I3, J3, K3
Populate
CPU 4 Slots
M1, N1, O1, P1,
M2, N2, O2, P2
M1, N1, O1,
M2, N2, O2,
M3, N3, O3
10Not recommended for performance reasons
11Not recommended for performance reasons
12A1, B1, C1, D1,
A2, B2, C2, D2,
A3, B3, C3, D3
E1, F1, G1, H1,
E2, F2, G2, H2,
E3, F3, G3, H3
I1, J1, K1, L1,
I2, J2, K2, L2,
I3, J3, K3, L3
M1, N1, O1, P1,
M2, N2, O2, P2,
M3, N3, O3, P3
A quick reference for the recommended DIMM population order is shown in Table 28 and Table 29. CPU 1
DIMM population order is shown as an example. Pick the number of DIMMs in the table you wish to populate,
then, starting with A1, populate down Bank 1, then down Bank 2, then down Bank 3 until you reach the
number you picked. If you picked 8 DIMMs, for example, use
Table 28 and populate DIMMs in A1, B1, C1,
D1, A2, B2, C2, and D2. If you picked 6 DIMMs, for example, use Table 29 and populate DIMMs in A1, B1, C1,
A2, B2, and C2, leaving D1 and D2 empty.
Table 28 DIMM Population Quick Reference for CPU 1 (for 1, 2, 4, 7, 8, 10, 11, or 12 DIMMS)
CPU ChannelBank 1Bank 2Bank 3
A159
B26
C3
1
7
10
11
1
1
D4812
1. Not recommended for performance reasons
Table 29 DIMM Population Quick Reference for CPU 1 (for 3, 5, 6, or 9 DIMMS)
CPU ChannelBank 1Bank 2Bank 3
A147
B2
1
5
8
C369
Demptyemptyempty
1. Not recommended for performance reasons
42
Cisco UCS B420 M4 Blade Server
Page 43
DIMM Physical Layout
305020
CPU 1
CPU 2
CPU 4
CPU 3
14
23
The overall DIMM and CPU physical layout is shown in Figure 9.
Figure 9 DIMM and CPU Layout
SUPPLEMENTAL MATERIAL
1Channel A - D DIMMs for CPU 13Channel I - L DIMMs for CPU 3
2Channel E - H DIMMs for CPU 24Channels M - P DIMMs for CPU 4
Cisco UCS B420 M4 Blade Server
43
Page 44
SUPPLEMENTAL MATERIAL
B3
B2
B1
A3
A2
A1
G1
G2
G3
H1
H2
H3
C1
C2
C3
D1
D2
D3
F3
F2
F1
E3
E2
E1
L3
L2
L1
K3
K2
K1
P3
P2
P1
O3
O2
O1
I1
I2
I3
J1
J2
J3
M1
M2
M3
N1
N2
N3
CPU 4
Front
Rear
CPU 1
CPU 3CPU 2
305018
Figure 10 shows how channels are physically laid out on the blade server. The DIMM slots are contiguous to
their associated CPU.
Figure 10 Physical Representation of DIMMs and CPUs
44
Cisco UCS B420 M4 Blade Server
Page 45
SUPPLEMENTAL MATERIAL
Network Connectivity
This section shows how the supported adapter card configurations for the B420 M4 connect to the Fabric
Extender modules in the 5108 blade server chassis.
There are three configurable adapter slots on the B420 M4. One slot supports only the VIC 1340/1240
adapter, and two additional slots accommodate Cisco adapters, as well as Cisco UCS Storage Accelerator
adapters.
install at least one VIC adapter in one of the three adapter slots.
Maximum bandwidth is a function of the Fabric Extender, the adapter, and the adapter slot, as shown in
Table 30.
Table 12 on page 23 and Table 13 on page 24 show supported adapter configurations. You must
Table 30 Maximum Bandwidth Using Dual Fabric Extenders (FEXs)
Maximum
FEX Model
Bandwidth
Using Dual
Example Configuration to Achieve Maximum Bandwidth
Figure 11 shows the configuration for maximum bandwidth, where the following ports are routed to 8-port
Fabric Extender Modules A and B inside the 5108 blade server chassis:
■4 x 10 Gb KR ports from the VIC 1340/1240 adapter
■4 x 10 Gb KR ports from the Port Expander
■8 x 10 Gb KR ports from the VIC 1380/1280 adapter
■4 x 10 Gb from a Cisco Storage Accelerator when placed in adapter slot 2, similar to the
Cisco UCS Port Expander Card for VIC. Requires UCSM 2.2(4b) or later.
The resulting aggregate bandwidth is 160 Gb (80 Gb to each Fabric Extender).
NOTE: In Figure 11, note that CPU4 controls adapter slot 2. In the case where a
Port Expander is installed in this slot, CPU4 is not needed. However, in the case
where a Storage Accelerator is installed in the slot, CPU4 must be installed.
Cisco UCS B420 M4 Blade Server
45
Page 46
SUPPLEMENTAL MATERIAL
Figure 11 UCS B420 M4 Connections to the Fabric Extenders
To Fabric InterconnectsTo Fabric Interconnects
UCS 5108 Blade Server Chassis
Fabric
Extender A
4 x 10G KR
4 x 10G KR
2 x 10G KR
Port
VIC 1380/1280
Adapter
Expander or
Storage
Accelerator
adapter slot #2adapter slot #3
CPU 4CPU 2
B420 M4 Blade Server
2 x 10G KR
4 x 10G KR
Fabric
Extender B
2 x 10G KR
VIC 1340/1240
Adapter Card
adapter slot #1 for VIC 1240
PCIe x16PCIe x16PCIe x16
CPU 1
2 x 10G KR
46
Cisco UCS B420 M4 Blade Server
Page 47
SUPPLEMENTAL MATERIAL
VIC 1340/1240 and Port Expander
Adapter slot 1 is dedicated to the VIC 1340 or VIC 1240 adapter, and no other adapter card can be installed
in this slot. There are two groups of four ports on the VIC 1340/1240:
■Two ports of the first group and two ports of the second group are wired through the UCS
5108 Blade Server chassis to Fabric Extender A and Fabric Extender B.
■The other two ports of each group are wired to adapter slot 2. The VIC 1340 or 1240 adapter
senses the type of adapter installed in adapter slot 2. If a Port Expander is installed in
adapter slot 2, the four 10G KR ports between the adapters are used for port expansion;
otherwise they are unused.
With the Port Expander installed, there are up to eight 10 Gb network interfaces to each Fabric Extender
(depending on the Fabric Extender installed), as represented in
Figure 12 VIC 1340/1240 Connectivity
Figure 12.
Fabric
Extender A
Cisco-certied
Adapter
adapter slot #3
CPU 2
Variable # of ports (depends on Fabric Extender type)
Cisco-certied
Adapter
adapter slot #2adapter slot #1 for VIC 1240
PCIe x16
PCIe x16
CPU 4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
Port Group APort Group B
VIC 1340/1240 Adapter
PCIe x16
10G KR
CPU 1
Fabric
Extender B
10G KR
Cisco UCS B420 M4 Blade Server
47
Page 48
SUPPLEMENTAL MATERIAL
Notes
Connectivity Using the Cisco UCS 2304 Fabric Extender
The connectivity options shown in Figure 13 on page 49 through Figure 22 on page 53 are summarized in
Table 31.
Table 31 Supported Adapter Combinations Using the 2304 FEX1
Adapter Slot 1 Adapter Slot 2Adapter Slot 3
Total Available
Bandwidth
2 x 2304
VIC 1340/1240Not populatedNot populated40 Gb
Figure 13 on page 49
VIC 1340
Port Expander Card2
Not populated80 Gb
Figure 14 on page 49
VIC 1340Not populatedVIC 1380120 Gb
Figure 15 on page 50
VIC 1240
Port Expander Card
VIC 1280160 Gb
Figure 16 on page 50
VIC 1340
Port Expander Card
2
VIC 1380160 Gb
Figure 17 on page 51
Not populatedNot populatedVIC 128080 Gb
Figure 18 on page 51
Not populatedNot populatedVIC 138080 Gb
Figure 19 on page 52
VIC 1340
Cisco UCS Storage Accelerator
3
Not populated80 Gb
Figure 20 on page 52
VIC 1340 not populatedCisco UCS Storage Accelerator40 Gb
Figure 21 on page 53
VIC 1340
4
Cisco UCS Storage Accelerator
3
Cisco UCS Storage Accelerator80 Gb
Figure 22 on page 53
1. Not all combinations are supported in 2-CPU configurations (see Table 12 on page 23). In addition, 2304XP is
only supported with B420 M4 v4 CPUs starting with UCSM 3.1(2).
2. Dual 40G native interface (for VIC 1340).
3. If running 2.2(4b) or greater and equipped with either storage accelerator SKU UCSB-F-FIO-1300MP or
UCSB-F-FIO-1600MS, then dual native 40G interface. Otherwise, the bandwidth is 20G (2x10G port-channel) per
FI.
4. This combination requires that both Storage Accelerators be identical.
48
Cisco UCS B420 M4 Blade Server
Page 49
SUPPLEMENTAL MATERIAL
VIC 1340/1240 Adapter
adapter slot #2adapter slot #1 for VIC 1340/1240
Port Group APort Group B
CPU 1
PCIe x16
No Adapter
Installed
No Adapter
Installed
adapter slot #3
CPU2
CPU4
2304 Fabric
Extender B
2304 Fabric
Extender A
EMPTYEMPTY
10G KR
10G KR
10G KR
10G KR
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
2304 Fabric
Extender B
2304 Fabric
Extender A
VIC 1340 Adapter
Port Expander Adapter
Port Group APort Group B
CPU 1
PCIe x16
10G KR
Native 40G interface
Native 40G interface
No
adapter
installed
adapter slot #3
adapter slot #2adapter slot #1 for VIC 1340
PCIe x16
CPU 2
EMPTY
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
In Figure 13, two ports from the VIC 1340/1240 adapter are channeled to 2304 Fabric Extender A and two
are channeled to 2304 Fabric Extender B. The result is 20 Gb of bandwidth to each Fabric Extender.
Figure 13 VIC 1340/1240 (adapter slots 2 and 3 empty)
In Figure 14, two ports from the VIC 1340 are connected to the 2304 Fabric Extender A and two are
connected to 2304 Fabric Extender B. The Port Expander Card installed in adapter slot 2 acts as a
pass-through device to provide two additional ports to each of the port groups and thus create a native 40G
interface to each of the Fabric Extenders. Adapter slot 3 is empty. The result is 40 Gb of bandwidth to each
Fabric Extender.
Figure 14 VIC 1340 and Port Expander in Adapter Slot 2 (adapter slot 3 empty)
Cisco UCS B420 M4 Blade Server
49
Page 50
SUPPLEMENTAL MATERIAL
2304 Fabric
Extender B
2304 Fabric
Extender A
VIC 1340 AdapterVIC 1380 Adapter
No Adapter
Installed
adapter slot #3
adapter slot #2
adapter slot #1 for VIC 1340
Port Group APort Group B
CPU 1
PCIe x16 PCIe x16
CPU 2
EMPTY
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
Port Group APort Group B
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
2304XP Fabric
Extender A
VIC 1240 Adapter
Port Expander Adapter
Port Group APort Group B
CPU 1
PCIe x16
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
adapter slot #2adapter slot #1 for VIC 1240
VIC 1280 Adapter
adapter slot #3
PCIe x16
CPU 2
Port Group APort Group B
2304XP Fabric
Extender B
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
In Figure 15, two ports from the VIC 1340 are channeled to 2304 Fabric Extender A and two are channeled
to 2304 Fabric Extender B. Adapter slot 2 is empty. The VIC 1380 installed in adapter slot 3 connects four
ports from a native 40G interface to each of the Fabric Extenders. The result is 60 Gb of bandwidth to each
Fabric Extender.
In Figure 16, two ports from the VIC 1240 are channeled to 2304 Fabric Extender A and two are channeled
to 2304 Fabric Extender B. The Port Expander Card installed in adapter slot 2 acts as a pass-through device,
channeling two ports to each of the Fabric Extenders. In addition, the VIC 1280 channels four ports to each
Fabric Extender. The result is 80 Gb of bandwidth to each Fabric Extender.
Figure 16 VIC 1240, Port Expander in Adapter Slot 2, and VIC 1280 in Adapter Slot 3
50
Cisco UCS B420 M4 Blade Server
Page 51
SUPPLEMENTAL MATERIAL
2304 Fabric
Extender B
2304 Fabric
Extender A
No VIC 1240 adapter installedNo adapter installedVIC 1280 Adapter
adapter slot #3adapter slot #2
adapter slot #1 for VIC 1240
CPU 1
PCIe x16 PCIe x16
CPU 2
EMPTYEMPTY
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
Port Group APort Group B
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
In Figure 17, two ports from the VIC 1340 are connected to 2304 Fabric Extender A and two are connected
to 2304 Fabric Extender B. The Port Expander Card installed in adapter slot 2 acts as a pass-through device
to provide additional ports to each of the port groups and thus create a native 40G interface to each of the
Fabric Extenders. In addition, the VIC 1380 connects four ports to each Fabric Extender. The result is 80 Gb
of bandwidth to each Fabric Extender.
Figure 17 VIC 1340, Port Expander in Adapter Slot 2, and VIC 1380 in Adapter Slot 3
2304XP Fabric
Extender A
Native 40G interface
Native 40G interface
Port Group APort Group B
VIC 1380 Adapter
adapter slot #3
PCIe x16
CPU 2
Port Group APort Group B
Port Expander Adapter
adapter slot #2adapter slot #1 for VIC 1240
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
CPU4
is installed in slot #2, CPU4 is required.
VIC 1340 Adapter
CPU 1
PCIe x16
2304XP Fabric
Extender B
Native 40G interface
In Figure 18, four ports from the VIC 1280 are channeled to 2304 Fabric Extender A and four are channeled
to 2304 Fabric Extender B. The VIC 1240 slot is empty and adapter slot 2 is empty. The result is 40 Gb of
bandwidth to each Fabric Extender.
Figure 18 VIC 1280 (VIC 1340 slot em pty and adapter slot 2 empty)
Cisco UCS B420 M4 Blade Server
51
Page 52
SUPPLEMENTAL MATERIAL
2304 Fabric
Extender B
2304 Fabric
Extender A
No VIC 1340 adapter installedNo adapter installedVIC 1380 Adapter
Native 40G interface
Native 40G interface
adapter slot #3adapter slot #2
adapter slot #1 for VIC 1340
CPU 1
PCIe x16 PCIe x16
CPU 2
EMPTYEMPTY
Port Group APort Group B
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
VIC 1340 Adapter
adapter slot #1 for VIC 1340
Port Group APort Group B
CPU1
PCIe x16
CPU 2
No Adapter
or Storage Accelerator
(optional) Installed
adapter slot #3
2304 Fabric
Extender B
EMPTY
Cisco UCS Storage Accelerator
Native 40G interface
2304 Fabric
Extender A
Native 40G interface
adapter slot #2
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
In Figure 19, four ports from the VIC 1380 are connected to 2304 Fabric Extender A and four are connected
to 2304 Fabric Extender B. The VIC 1340 slot is empty and adapter slot 2 is empty. A native 40G interface is
connected to each of the Fabric Extenders. The result is 40 Gb of bandwidth to each Fabric Extender.
Figure 19 VIC 1380 (VIC 1340 slot em pty and adapter slot 2 empty)
In Figure 20 (supported on 4-CPU configurations), the Storage Accelerator installed in adapter slot 2 also
has the port expander capability to increase the overall network I/O to the server. It acts as a pass-through
device to provide two additional ports to each of the port groups and thus create a native 40G interface to
each of the Fabric Extenders. The result is 40 Gb of bandwidth to each Fabric Extender. This configuration
requires 4 CPUs.
Figure 20 VIC 1340 (Cisco UCS Storage Accelerator in slot 2 and adapter slot 3 empty)
52
Cisco UCS B420 M4 Blade Server
Page 53
SUPPLEMENTAL MATERIAL
VIC 1340 Adapter
adapter slot #1 for VIC 1340
Port Group APort Group B
CPU1
PCIe x16
CPU 2
adapter slot #3
2304 Fabric
Extender B
2304 Fabric
Extender A
Cisco UCS Storage Accelerator
Native 40G interface
Native 40G interface
Cisco UCS Storage Accelerator
adapter slot #2
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
In Figure 21 (supported on 2-CPU configurations), two ports from the VIC 1340 adapter are channeled to
2304 Fabric Extender A and two are channeled to 2304 Fabric Extender B. The result is 20 Gb of bandwidth
to each Fabric Extender. The storage accelerator in slot 3 provides no network connectivity.
Figure 21 VIC 1340 (Cisco UCS Storage Accelerator in slot 3 and adapter slot 2 empty)
2304 Fabric
Extender A
10G KR
10G KR
Port Group APort Group B
10G KR
2304 Fabric
Extender B
10G KR
EMPTY
No Adapter
Cisco UCS Storage Accelerator
adapter slot #3
CPU 2
Installed
adapter slot #2
CPU4
VIC 1340 Adapter
adapter slot #1 for VIC 1340
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
PCIe x16
CPU1
In Figure 22, the Storage Accelerator installed in adapter slot 2 also has the port expander capability to
increase the overall network I/O to the server. It acts as a pass-through device to provide two additional
ports to each of the port groups and thus create a native 40G interface to each of the Fabric Extenders. The
result is 40 Gb of bandwidth to each Fabric Extender. Two identical Cisco UCS Storage Accelerator adapters
are installed in slots 2 and 3, but the one in slot 3 provides no network connectivity. This configuration
requires 4 CPUs.
Figure 22 VIC 1340 (Cisco UCS Storage Accelerator in slots 2 and 3)
Cisco UCS B420 M4 Blade Server
53
Page 54
SUPPLEMENTAL MATERIAL
Notes
Connectivity Using the Cisco UCS 2208XP Fabric Extender
The connectivity options shown in Figure 23 on page 55 through Figure 30 on page 58 are summarized in
Table 32.
Table 32 Supported Adapter Combinations Using the 2208XP FEX1
VIC 1340 not populatedCisco UCS Storage Accelerator40 Gb
Figure 29 on page 58
VIC 1340
4
Cisco UCS Storage Accelerator
3
Cisco UCS Storage Accelerator80 Gb
Figure 30 on page 58
1. Not all combinations are supported in 2-CPU configurations (see Table 12 on page 23)
2. Dual 40G via 4x10 port-channel
3. f running 2.2(4b) or greater and equipped with either storage accelerator SKU UCSB-F-FIO-1300MP or
UCSB-F-FIO-1600MS, then dual 40G via 4x10 port-channel. Otherwise, the bandwidth is 20G (2x10G
port-channel) per FI.
4. This combination requires that both Storage Accelerators be identical.
54
Cisco UCS B420 M4 Blade Server
Page 55
SUPPLEMENTAL MATERIAL
2208 Fabric
Extender B
2208 Fabric
Extender A
VIC 1340/1240 Adapter
adapter slot #2adapter slot #1 for VIC 1340/1240
Port Group APort Group B
CPU 1
PCIe x16
No Adapter
Installed
No Adapter
Installed
adapter slot #3
CPU2
CPU4
EMPTYEMPTY
10G KR
10G KR
10G KR
10G KR
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
2208 Fabric
Extender B
2208 Fabric
Extender A
VIC 1340 Adapter
Port Expander Adapter
Port Group APort Group B
CPU 1
PCIe x16
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
No
adapter
installed
adapter slot #3
adapter slot #2adapter slot #1 for VIC 1340
PCIe x16
CPU 2
EMPTY
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
10G KR
10G KR
10G KR
10G KR
In Figure 23, two ports from the VIC 1340/1240 adapter are channeled to 2208XP Fabric Extender A and
two are channeled to 2208XP Fabric Extender B. The result is 20 Gb of bandwidth to each Fabric Extender.
Figure 23 VIC 1340/1240 (adapter slots 2 and 3 empty)
In Figure 24, two ports from the VIC 1340 are channeled to 2208XP Fabric Extender A and two are
channeled to 2208XP Fabric Extender B. The Port Expander Card installed in adapter slot 2 acts as a
pass-through device, channeling two ports to each of the Fabric Extenders. Adapter slot 3 is empty. The
result is 40 Gb of bandwidth to each Fabric Extender.
Figure 24 VIC 1340 and Port Expander in Adapter Slot 2 (adapter slot 3 empty)
Cisco UCS B420 M4 Blade Server
55
Page 56
SUPPLEMENTAL MATERIAL
2208 Fabric
Extender B
2208 Fabric
Extender A
VIC 1380/1240 AdapterVIC 1380/1280 Adapter
No Adapter
Installed
adapter slot #3
adapter slot #2
adapter slot #1 for VIC 1340/1240
Port Group APort Group B
CPU 1
PCIe x16 PCIe x16
CPU 2
EMPTY
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
Port Group APort Group B
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
2208 Fabric
Extender B
2208 Fabric
Extender A
VIC 1340/1240 Adapter
Port Expander Adapter
Port Group APort Group B
CPU 1
PCIe x16
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
adapter slot #2adapter slot #1 for VIC 1340/1240
VIC 1380/1280 Adapter
adapter slot #3
PCIe x16
CPU 2
Port Group APort Group B
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
In Figure 25, two ports from the VIC 1340/1240 are channeled to 2208XP Fabric Extender A and two are
channeled to 2208XP Fabric Extender B. Adapter slot 2 is empty. The VIC 1380/1280 installed in adapter slot
3 channels four ports to each of the Fabric Extenders. The result is 60 Gb of bandwidth to each Fabric
Extender.
In Figure 26, two ports from the VIC 1340/1240 are channeled to 2208XP Fabric Extender A and two are
channeled to 2208XP Fabric Extender B. The Port Expander Card installed in adapter slot 2 acts as a
pass-through device, channeling two ports to each of the Fabric Extenders. In addition, the VIC 1380/1280
channels four ports to each Fabric Extender. The result is 80 Gb of bandwidth to each Fabric Extender.
Figure 26 VIC 1340/1240, Port Expander in Adapter Slot 2, and VIC 1380/1280 in Adapter Slot 3
56
Cisco UCS B420 M4 Blade Server
Page 57
SUPPLEMENTAL MATERIAL
2208 Fabric
Extender B
2208 Fabric
Extender A
VIC 1340 Adapter
adapter slot #1 for VIC 1340
Port Group APort Group B
CPU1
PCIe x16
CPU 2
No Adapter
Installed
adapter slot #3
EMPTY
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
Cisco UCS Storage Accelerator
10G KR
10G KR
10G KR
10G KR
adapter slot #2
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
In Figure 27, four ports from the VIC 1380 are channeled to 2208XP Fabric Extender A and four are
channeled to 2208XP Fabric Extender B. The VIC 1340 slot is empty and adapter slot 2 is empty. The result
is 40 Gb of bandwidth to each Fabric Extender.
Figure 27 VIC 1380 (VIC 1340 slot em pty and adapter slot 2 empty)
2208 Fabric
Extender A
10G KR
Port Group APort Group B
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
2208 Fabric
Extender B
EMPTYEMPTY
No VIC 1340 adapter installedNo adapter installedVIC 1380 Adapter
adapter slot #3adapter slot #2
If a Port Expander is installed in slot #2,
CPU 2
CPU4 is optional. If a Storage Accelerator
CPU4
is installed in slot #2, CPU4 is required.
adapter slot #1 for VIC 1340
PCIe x16 PCIe x16
CPU 1
In Figure 28 (supported on 4-CPU configurations), the Storage Accelerator installed in adapter slot 2 acts as
a pass-through device, channeling two ports to each of the Fabric Extenders. In addition, two ports from
the VIC 1340 adapter are channeled to 2208XP Fabric Extender A and two are channeled to 2208XP Fabric
Extender B. The result is 40 Gb of bandwidth to each Fabric Extender. This configuration requires 4 CPUs.
Figure 28 VIC 1340 (Cisco UCS Storage Accelerator in slot 2 and adapter slot 3 empty)
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
In Figure 29 (supported on 2-CPU configurations), two ports from the VIC 1340 adapter are channeled to
2208XP Fabric Extender A and two are channeled to 2208XP Fabric Extender B. The result is 20 Gb of
bandwidth to each Fabric Extender. The storage accelerator in slot 3 provides no network connectivity.
Figure 29 VIC 1340 (Cisco UCS Storage Accelerator in slot 3 and adapter slot 2 empty)
2208 Fabric
Extender A
10G KR
10G KR
Port Group APort Group B
10G KR
2208 Fabric
Extender B
10G KR
EMPTY
No Adapter
Cisco UCS Storage Accelerator
adapter slot #3
CPU 2
Installed
adapter slot #2
CPU4
VIC 1340 Adapter
adapter slot #1 for VIC 1340
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
PCIe x16
CPU1
In Figure 30, the Storage Accelerator installed in adapter slot 2 acts as a pass-through device, channeling
two ports to each of the Fabric Extenders. In addition, two ports from the VIC 1340 adapter are channeled
to 2208XP Fabric Extender A and two are channeled to 2208XP Fabric Extender B. The result is 40 Gb of
bandwidth to each Fabric Extender. Two identical Cisco UCS Storage Accelerator adapters are installed in
slots 2 and 3, but the one in slot 3 provides no network connectivity. This configuration requires 4 CPUs.
Figure 30 VIC 1340 (Cisco UCS Storage Accelerator in slots 2 and 3)
58
Cisco UCS B420 M4 Blade Server
Page 59
SUPPLEMENTAL MATERIAL
Notes
Connectivity using the Cisco UCS 2204XP Fabric Extender
The connectivity options shown in Figure 31 on page 60 through Figure 38 on page 63 are shown in
Table 33.
Table 33 Supported Adapter Combinations Using the 2204XP
VIC 1340 not populatedCisco UCS Storage Accelerator20 Gb
Figure 37 on page 63
VIC 1340
Cisco UCS Storage Accelerator
2
Cisco UCS Storage Accelerator40 Gb
Figure 38 on page 63
1. Not all combinations are supported in 2-CPU configurations (see Table 12 on page 23)
2. This combination requires that both Storage Accelerators be identical.
Cisco UCS B420 M4 Blade Server
59
Page 60
SUPPLEMENTAL MATERIAL
2204XP Fabric
Extender B
2204XP Fabric
Extender A
VIC 1340 Adapter
Port Expander Adapter
Port Group APort Group B
CPU 1
PCIe x16
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
No adapter
installed
adapter slot #3
adapter slot #2adapter slot #1 for VIC 1340
PCIe x16
CPU 2
EMPTY
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
In Figure 31, one port from the VIC 1340/1240 is channeled to 2204XP Fabric Extender A and one is
channeled to 2204XP Fabric Extender B. The result is 10 Gb of bandwidth to each Fabric Extender.
Figure 31 VIC 1340/1240 (adapter slots 2 and 3 empty)
2204XP Fabric
Extender A
10G KR
Port Group APort Group B
2204XP Fabric
Extender B
10G KR
EMPTYEMPTY
No Adapter Installed
adapter slot #3
CPU 2
adapter slot #2adapter slot #1 for VIC 1340/1240
If a Port Expander is installed in slot #2,
CPU4
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
VIC 1340/1240 AdapterNo Adapter Installed
PCIe x16
CPU 1
In Figure 32, one port from the VIC 1340 is channeled to 2204XP Fabric Extender A and one is channeled to
2204XP Fabric Extender B. The Port Expander Card installed in adapter slot 2 acts as a pass-through device,
channeling one port to each of the Fabric Extenders. Adapter slot 3 is empty. The result is 20 Gb of
bandwidth to each Fabric Extender.
Figure 32 VIC 1340 and Port Expander in Adapter Slot 2 (adapter slot 3 empty)
60
Cisco UCS B420 M4 Blade Server
Page 61
SUPPLEMENTAL MATERIAL
2204XP Fabric
Extender B
2204XP Fabric
Extender A
VIC 1340/1240 AdapterVIC 1380/1280 Adapter
No
adapter
installed
adapter slot #3adapter slot #2
adapter slot #1 for VIC 1340/1240
Port Group APort Group B
CPU 1
PCIe x16 PCIe x16
CPU 2
EMPTY
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
Port Group APort Group B
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
2204XP Fabric
Extender A
VIC 1340/1240 Adapter
Port Expander Adapter
Port Group APort Group B
CPU 1
PCIe x16
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
10G KR
adapter slot #2adapter slot #1 for VIC 1340/1240
VIC 1380/1280 Adapter
adapter slot #3
PCIe x16
CPU 2
Port Group APort Group B
2204XP Fabric
Extender B
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
In Figure 33, one port from the VIC 1340/1240 is channeled to 2204XP Fabric Extender A and one is
channeled to 2204XP Fabric Extender B. Adapter slot 2 is empty. The VIC 1380/1280 installed in adapter slot
3 channels two ports to each of the Fabric Extenders. The result is 30 Gb of bandwidth to each Fabric
Extender.
In Figure 34, one port from the VIC 1340/1240 is channeled to 2204XP Fabric Extender A and one is
channeled to 2204XP Fabric Extender B. The Port Expander Card installed in adapter slot 2 acts as a
pass-through device, channeling one port to each of the Fabric Extenders. In addition, the VIC 1380/1280
channels two ports to each Fabric Extender. The result is 40 Gb of bandwidth to each Fabric Extender.
Figure 34 VIC 1340/1240, Port Expander and VIC 1380/1280
Cisco UCS B420 M4 Blade Server
61
Page 62
SUPPLEMENTAL MATERIAL
2204XP Fabric
Extender B
2204XP Fabric
Extender A
No VIC 1340 installedVIC 1380 Adapter
No adapter
installed
adapter slot #3
adapter slot #2
adapter slot #1 for VIC 1340
CPU 1
PCIe x16 PCIe x16
CPU 2
EMPTYEMPTY
10G KR
10G KR
10G KR
10G KR
Port Group APort Group B
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
VIC 1340 Adapter
adapter slot #1 for VIC 1340
Port Group APort Group B
CPU 1
PCIe x16
CPU 2
adapter slot #3
2204XP Fabric
Extender B
2204XP Fabric
Extender A
EMPTY
10G KR
10G KR
No Adapter Installed
Cisco UCS Storage Accelerator
10G KR
10G KR
10G KR
10G KR
adapter slot #2
CPU4
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
In Figure 35, two ports from the VIC 1380 are channeled to 2204XP Fabric Extender A and two are
channeled to 2204XP Fabric Extender B. The VIC 1340 slot is empty and adapter slot 2 is empty. The result
is 20 Gb of bandwidth to each Fabric Extender.
Figure 35 VIC 1380 in Adapter Slot 3 (VIC 1340 slot empty and adapter slot 2 empty)
In Figure 36, the Storage Accelerator installed in adapter slot 2 acts as a pass-through device, channeling
one port to each of the Fabric Extenders. In addition, one port from the VIC 1340 is channeled to 2204XP
Fabric Extender A and one is channeled to 2204XP Fabric Extender B. The result is 20 Gb of bandwidth to
each Fabric Extender. This configuration requires 4 CPUs.
In Figure 37, one port from the VIC 1340 adapter is connected to 2204XP Fabric Extender A and one is
connected to 2204XP Fabric Extender B. The result is 10 Gb of bandwidth to each Fabric Extender. The
storage accelerator in slot 3 provides no network connectivity.
Figure 37 VIC 1340 (Cisco UCS Storage Accelerator in slot 3 and adapter slot 2 empty)
2204XP Fabric
Extender A
10G KR
Port Group APort Group B
10G KR
2204XP Fabric
Extender B
EMPTY
No Adapter
Cisco UCS Storage Accelerator
adapter slot #3
CPU 2
Installed
adapter slot #2
CPU4
VIC 1340 Adapter
adapter slot #1 for VIC 1340
If a Port Expander is installed in slot #2,
CPU4 is optional. If a Storage Accelerator
is installed in slot #2, CPU4 is required.
PCIe x16
CPU1
In Figure 38, the Storage Accelerator installed in adapter slot 2 acts as a pass-through device, channeling
one port to each of the Fabric Extenders. In addition, one port from the VIC 1340 is channeled to 2204XP
Fabric Extender A and one is channeled to 2204XP Fabric Extender B. The result is 20 Gb of bandwidth to
each Fabric Extender. Two identical Cisco UCS Storage Accelerator adapters are installed in slots 2 and 3,
but the one in slot 3 provides no network connectivity. This configuration requires 4 CPUs.
Figure 38 VIC 1340/1240 (Cisco UCS Storage Accelerator in slots 2 and 3)