QLogic HCA and InfiniPath® Software Install Guide
Version 2.2
Information furnished in this manual is believed to be accurate and reliable. However, QLogic Corporation assumes no
responsibility for its use, nor for any infringements of patents or other rights of third parties which may result from its
use. QLogic Corporation reserves the right to change product specifications at any time without notice. Applications
described in this document for any of these products are for illustrative purposes only. QLogic Corporation makes no
representation nor warranty that such applications are suitable for the specified use without further testing or
modification. QLogic Corporation assumes no responsibility for any errors that may appear in this document.
No part of this document may be copied nor reproduced by any means, nor translated nor transmitted to any magnetic
medium without the express written consent of QLogic Corporation. In accordance with the terms of their valid QLogic
agreements, customers are permitted to make electronic and paper copies of this document for their own exclusive
use.
The QHT7040, QHT7140, QLE7140, QLE7240, and QLE7280 QLogic Host Channel Adapters are covered by the
following patent: 7308535.
S
Document Revision History
Rev. 1.0, 8/20/2005
Rev. 1.1, 11/15/05
Rev. 1.2, 02/15/06
Rev. 1.3, Beta 1, 4/15/06
Rev 1.3, 6/15/06
Rev. 2.0 Beta, 8/15/06, QLogic Version Number IB0056101-00 A
Rev. 2.0 Beta 2 10/15/06, QLogic Version Number IB0056101-00 B
Rev. 2.0 11/30/06, QLogic Version Number IB0056101-00 C
Rev. 2.0 3/23/07, QLogic Version Number IB0056101-00 D
Rev. 2.1 8/24/07, QLogic Version Number IB0056101-00 E
Rev. 2.2 5/27/08, QLogic Version Number IB0056101-00 F
Rev. 2.2 9/5/08, QLogic Version Number IB0056101-00 G
ChangesSections Affected
Re-format to new template (no change to content)All
QLogic HCA and InfiniPath® Software Install Guide
Version 2.2
Notes
S
Page viiiIB0056101-00 G
1Introduction
This chapter describes the contents, intended audience, and organization of the
QLogic HCA and InfiniPath Software Install Guide.
The QLogic HCA and InfiniPath Software Install Guide contains instructions for
installing the QLogic Host Channel Adapters (HCAs) and the InfiniPath and
OpenFabrics software. The following adapters are covered in this guide:
This installation guide is intended for cluster administrators responsible for
installing the QLogic QLE7140, QLE7240, QLE7280 or QHT7040/QHT7140
adapter and InfiniPath software on their Linux
installation information and instructions for administering the QLogic cluster can
be found in the QLogic HCA and InfiniPath Software User Guide.
®
(PCIe)
®
cluster. Additional detailed
The QLogic HCA and InfiniPath Software Install Guide assumes that you are
familiar with both cluster networking and the specific hardware that you plan to
use. Before installing the HCA, you should have basic knowledge of your host and
target operating systems, and working knowledge of message passing concepts.
This document does not contain all the information you need to use basic Linux
commands or to perform all system administration tasks. For this information, see
the software documentation you received with your system.
How this Guide is Organized
The QLogic HCA and InfiniPath Software Install Guide is organized into these
sections:
Section 1, Introduction, contains an overview of the HCAs and software,
describes interoperability with other products, lists all related documentation,
and provides QLogic contact information.
Section 2, Feature Overview, contains features for this release, the
supported QLogic adapter models, supported distributions and kernels, and
a list of the software components.
IB0056101-00 G1-1
1 – Introduction
Overview
S
Section 3, Step-by-Step Installation Checklist, provides a high-level
overview of the hardware and software installation procedures.
Section 4, Hardware Installation, includes instructions for installing the
QLogic QLE7140, QLE7240, QLE7280, QHT7040, and QHT7140 HCAs.
Section 5, Software Installation, includes instructions for installing the
QLogic InfiniPath and OpenFabrics software.
Appendix A, Installation Troubleshooting, contains information about issues
that may occur during installation.
Appendix B, Configuration Files, contains descriptions of the configuration
and configuration template files used by the InfiniPath and OpenFabrics
software.
Appendix C, RPM Descriptions
Index, lists major subjects and concepts with page numbers for easy
reference.
Overview
The material in this documentation pertains to an InfiniPath cluster. A cluster is
defined as a collection of nodes, each attached to an InfiniBand™-based fabric
through the QLogic interconnect. The nodes are Linux-based computers, each
having up to 16 processors.
The QLogic HCAs are InfiniBand 4X. The Double Data Rate (DDR) QLE7240 and
QLE7280 adapters have a raw data rate of 20Gbps (data rate of 16Gbps). For the
Single Data Rate (SDR) adapters, the QLE7140 and QHT7140, the raw data rate
is 10Gbps (data rate of 8Gbps). The QLE7240 and QLE7280 can also run in SDR
mode.
The QLogic adapters utilize standard, off-the-shelf InfiniBand 4X switches and
cabling. The QLogic interconnect is designed to work with all InfiniBand-compliant
switches.
NOTE:
If you are using the QLE7240 or QLE7280, and want to use DDR mode,
then DDR-capable switches must be used.
1-2IB0056101-00 G
A
InfiniPath OpenFabrics software is interoperable with other vendors’ InfiniBand
Host Channel Adapters (HCAs) running compatible OpenFabrics releases. There
are several options for subnet management in your cluster:
Use the embedded Subnet Manager (SM) in one or more managed switches
supplied by your InfiniBand switch vendor.
Use the Open source Subnet Manager (OpenSM) component of
OpenFabrics.
Use a host-based Subnet Manager.
Interoperability
QLogic InfiniPath participates in the standard InfiniBand subnet management
protocols for configuration and monitoring. Note that:
InfiniPath OpenFabrics (including Internet Protocol over InfiniBand (IPoIB))
is interoperable with other vendors’ InfiniBand HCAs running compatible
OpenFabrics releases.
1 – Introduction
Interoperability
The QLogic MPI and Ethernet emulation stacks (
interoperable with other InfiniBand HCAs and Target Channel Adapters
(TCAs). Instead, InfiniPath uses an InfiniBand-compliant, vendor-specific
protocol that is highly optimized for MPI and Transmission Control Protocol
(TCP) between InfiniPath-equipped hosts.
NOTE:
See the OpenFabrics web site at www.openfabrics.org
on the OpenFabrics Alliance.
Conventions Used in this Guide
This guide uses the typographical conventions listed in Tab le 1- 1.
Table 1-1. Typographical Conventions
ConventionMeaning
commandFixed-space font is used for literal items such as commands, func-
tions, programs, files and pathnames, and program output.
variableItalic fixed-space font is used for variable names in programs and
command lines.
ipath_ether) are not
for more information
conceptItalic font is used for emphasis and concepts, as well as for docu-
mentation names/titles.
IB0056101-00 G1-3
1 – Introduction
Documentation
S
Table 1-1. Typographical Conventions (Continued)
ConventionMeaning
user inputBold fixed-space font is used for literal items in commands or con-
structs that you type.
$Indicates a command line prompt.
#Indicates a command line prompt as root when using bash or sh.
[ ]Brackets enclose optional elements of a command or program con-
struct.
...Ellipses indicate that a preceding element can be repeated.
>A right caret identifies the cascading path of menu commands used
in a procedure.
2.2The current version number of the software is included in the RPM
names and within this documentation.
NOTE:Indicates important information.
Documentation
The product documentation includes:
The QLogic HCA and InfiniPath Software Install Guide
The QLogic HCA and InfiniPath Software User Guide
The QLogic FastFabric Users Guide (for information on QLogic InfiniServ Tools)
The OFED+ Users Guide (for information on QLogic VNIC and QLogic SRP)
Release Notes
Quick Start Guide
Readme file
For more information on system administration, using the QLogic
Message-Passing Interface (MPI), and troubleshooting adapter hardware and
software, see the QLogic HCA and InfiniPath Software User Guide.
1-4IB0056101-00 G
A
Contact Information
Support HeadquartersQLogic Corporation
QLogic Web Sitewww.qlogic.com
Technical Support Web Sitesupport.qlogic.com
Technical Support Emailsupport@qlogic.com
Technical Training Emailtech.training@qlogic.com
North American Region
Emailsupport@qlogic.com
Phone+1-952-932-4040
1 – Introduction
Contact Information
4601 Dean Lakes Blvd
Shakopee, MN 55379
USA
Fax+1 952-974-4910
All other regions of the world
QLogic Web Sitewww.qlogic.com
IB0056101-00 G1-5
1 – Introduction
Contact Information
Notes
S
1-6IB0056101-00 G
2Feature Overview
This section contains the features for this release, the supported QLogic adapter
models, supported distributions and kernels, and a list of the software
components.
What’s New in this Release
This release adds support for the QLE7240 and QLE7280 InfiniBand DDR Host
Channel Adapters (HCAs), which offer twice the link bandwidth of SDR HCAs.
The extra bandwidth improves performance for both latency-sensitive and
bandwidth-intensive applications.
This version of the InfiniPath software provides support for all of the QLogic HCAs
in Ta bl e 2- 1.
Table 2-1. QLogic Adapter Model Numbers
QLogic Model
Number
QHT7040Single port 10Gbps SDR 4X InfiniBand to HTX adapter. For
systems with HTX expansion slots.
QHT7140
QLE7140Single port 10Gbps SDR 4X InfiniBand to PCI Express x8
QLE7240Single port 20Gbps DDR 4X InfiniBand to PCI Express x8
a
Single port 10Gbps SDR 4X InfiniBand to HTX adapter. For
systems with HTX expansion slots.
adapter. Supported on systems with PCI Express (PCIe) x8 or
x16 slots.
adapter. Supported on systems with PCI Express x8 or
x16 slots.
Description
IB0056101-00 G2-1
2 – Feature Overview
New Features
S
Table 2-1. QLogic Adapter Model Numbers (Continued)
QLogic Model
Number
QLE7280Single port 20Gbps DDR 4X InfiniBand to PCI Express x16
Table Notes
PCIe is Gen 1
a
The QHT7140 has a smaller form factor than the QHT7040, but is otherwise the same. Throughout
this document, the QHT7040 and QHT7140 will be collectively referred to as the QHT7140 unless
otherwise noted.
New Features
The following features are new to the 2.2 release:
Expanded MPI scalability enhancements for PCI Express have been added.
On the QLE7240 and QLE7280, up to 16 dedicated hardware contexts per
node are available. The QHT7140 has eight per node. The QLE7140 has
four per node.
A total of 64 processes on the QLE7240 and QLE7280 are supported when
context sharing is enabled. The QHT7040 and QHT7140 support a total of
32 processes per adapter. The QLE7140 supports a total of 16 processes
per adapter.
Description
adapter. Supported on systems with PCI Express x16 slots.
The QLE7280 is backward compatible; it can also be used with
PCIe adapters that connect to x8 slots.
This release continues support for multiple high-performance native PSM
Message Passing Interface (MPI) implementations added in the 2.1 release.
(PSM is QLogic’s accelerated library for high performance MPIs). In addition
to QLogic MPI, the currently supported MPI implementations are HP-MPI,
Open MPI, MVAPICH, and Scali. Open MPI provides MPI-2 functionality,
including one-sided operations and dynamic processes. These all offer the
same high performance as QLogic MPI.
Dual PCIe QLogic adapters per node are supported.
Driver support for the QLogic Virtual Network Interface Controller (VNIC) is
provided in this release. The VNIC Upper Layer Protocol (ULP) works in
concert with firmware running on Virtual Input/Output (VIO) hardware such
as the SilverStorm™ Ethernet Virtual I/O Controller (EVIC), providing virtual
Ethernet connectivity for Linux operating systems.
The QLogic InfiniBand Fabric Suite CD is available separately for purchase.
The CD includes FastFabric, the QLogic Subnet Manager (SM), and the
Fabric Viewer GUI for the QLogic SM.
2-2IB0056101-00 G
A
2 – Feature Overview
Supported Distributions and Kernels
A subset of the QLogic InfiniBand Fabric Suite, the enablement tools, are
offered with this release.
Two separate SCSI RDMA Protocol (SRP) modules are provided: the
standard OpenFabrics (OFED) SRP, and QLogic SRP.
QLogic MPI supports running exclusively on a single node without the
installation of the HCA hardware.
OpenMPI and MVAPICH libraries built with the GNU, PGI, PathScale™, and
4K Maximum Transfer Unit (MTU) is supported and is on by default. To take
The Lustre
Additional up-to-date information can be found on the QLogic web site,
specifically:
The high performance computing page at
The InfiniBand HCA page at
®
Intel
compilers are available, with corresponding mpitests RPMs. You
can use mpi-selector to choose which MPI you want. These all run over
PSM.
advantage of 4KB MTU, use a switch that supports 4KB MTU. QLogic also
supports 2KB switches, and 4KB MTU switches configured for 2KB MTU.
QLogic switches with firmware version 4.1 or later are recommended.
The QLogic interconnect runs on AMD™ Opteron™ and Intel EM64T systems
running Linux
versions for InfiniPath and OpenFabrics are listed in Table 2-2. The kernels are
the ones that shipped with the distributions.
®
. The currently supported distributions and associated Linux kernel
Scientific Linux 5.0, 5.12.6.18, 2.6.18-53 (x86_64)
®
SUSE
(SLES 10 GM, SP 1)
NOTE:
Linux Enterprise Server
Fedora Core 4 and Fedora Core 5 are not supported in the InfiniPath 2.2
release.
Compiler Support
QLogic MPI supports a number of compilers. These include:
PathScale Compiler Suite 3.0 and 3.1
PGI 5.2 , 6.0, and 7.1
Intel 9.x and 10.1
GNU gcc 3.3.x, 3.4.x, 4.0 and 4.1 compiler suites
gfortran
InfiniPath/OpenFabrics Supported
Kernels
2.6.18-8, 2.6.18-53 (x86_64)
2.6.16.21, 2.6.16.46 (x86_64)
The PathScale Compiler Suite Version 3.x is now supported on systems that have
the GNU 4.0 and 4.1 compilers and compiler environment (header files and
libraries).
Please check the QLogic web site for updated information on supported
compilers.
Software Components
The software includes the InfiniPath HCA driver, QLogic MPI, standard networking
over Ethernet emulation, InfiniPath Subnet Management Agent and associated
utilities, and OFED for InfiniPath.
This release includes a full set of OFED 1.3 usermode RPMs, with some
enhancements, including a new version of the VNIC tools and driver, and support
for the QLE7240 and QLE7280 adapters.
2-4IB0056101-00 G
A
2 – Feature Overview
Software Components
Included components are:
InfiniPath driver
InfiniPath Ethernet emulation (ipath_ether)
InfiniPath libraries
InfiniPath utilities, configuration, and support tools, including
ipath_checkout, ipath_control, ipath_pkt_test, and
ipathstats
QLogic MPI
QLogic MPI benchmarks and utilities
OpenMPI and MVAPICH libraries built with the GNU, PGI, PathScale, and
Intel compilers, with corresponding mpitests RPMs and mpi-selector
OpenFabrics protocols, including Subnet Management Agent
OpenFabrics libraries and utilities
QLogic VNIC module
Enablement tools
This release provides support for the following protocols and transport services:
IPoIB (TCP/IP networking in either Connected or Datagram mode)
Sockets Direct Protocol (SDP)
Open source Subnet Manager (OpenSM)
Unreliable Datagram (UD)
Reliable Connection (RC)
Unreliable Connection (UC)
Shared Receive Queue (SRQ)
Reliable Datagram Sockets (RDS)
iSCSI Extensions for RDMA (iSER)
This release supports two versions of SCSI RDMA Protocol (SRP):
OFED SRP
QLogic SRP
No support is provided for Reliable Datagram (RD).
NOTE:
OpenFabrics programs (32-bit) using the verbs interfaces are not supported
in this InfiniPath release, but may be supported in a future release.
IB0056101-00 G2-5
2 – Feature Overview
Software Components
More details about the hardware and software can be found in Section 4 and
Section 5.
S
2-6IB0056101-00 G
3Step-by-Step
Installation Checklist
This section provides an overview of the hardware and software installation
procedures. Detailed steps are found in Section 4 “Hardware Installation” and
Section 5 “Software Installation”.
Hardware Installation
The following steps summarize the basic hardware installation procedure:
1.Check that the adapter hardware is appropriate for your platform. See
Table 4-1.
2.Check to see that you have the appropriate cables and switches, as
described in “Cabling and Switches” on page 4-3.
3.Check to see that you are running a supported Linux distribution/kernel. See
Table 2-2 or Table 5-1.
4.Verify that the BIOS for your system is configured for use with the QLogic
adapter. See “Configuring the BIOS” on page 4-4.
5.Following the safety instructions in “Safety with Electricity” on page 4-5.
Unpack the adapter (“Unpacking Information” on page 4-5) and verify the
package contents.
6.Install the adapter by following the instructions in “Hardware Installation” on
page 4-9.
7.Cable the adapter to the switch, as described in “Cabling the Adapter to the
InfiniBand Switch” on page 4-17. Check that all InfiniBand switches are
configured.
8.Follow the steps in “Completing the Installation” on page 4-18 to finish the
installation.
The following steps summarize the basic InfiniPath and OpenFabrics software
installation and startup. These steps must be performed on each node in the
cluster:
1.Make sure that the HCA hardware installation has been completed
according to the instructions in “Hardware Installation” on page 4-1.
2.Verify that the Linux kernel software is installed on each node in the cluster.
The required kernels and supported Linux distributions for both InfiniPath
and OpenFabrics are defined in Table 5-1.
3.Make sure that your environment has been set up as described in “Setting
Up Your Environment” on page 5-2.
4.Download your version of the InfiniPath/OpenFabrics software from the
QLogic web site to a local server directory. See “Choosing the Appropriate
Download Files” on page 5-3.
5.Unpack the tar file and check for any missing files or RPMs. See
“Unpacking the InfiniPath tar File” on page 5-4 and “Check for Missing Files
or RPMs” on page 5-5.
S
6.Install the appropriate packages on each cluster node as described in
“Installing the InfiniPath and OpenFabrics RPMs” on page 5-6. Also see
Table 5-4 for a list of RPMs to install.
NOTE:
Rocks may be used as a cluster install method. See “Managing and
Installing Software Using Rocks” on page 5-36.
7.Under the following circumstances, the system needs to be rebooted:
If this is the first InfiniPath installation OR
If you have installed VNIC with the OpenFabrics RPM set
The system can be rebooted after all the software has been installed. If you
are only upgrading from a prior InfiniPath software installation, drivers can
be restarted manually. See “Starting the InfiniPath Service” on page 5-10.
8.If you want to use the optional InfiniPath (ipath_ether) and OpenFabrics
drivers (ipoib) and services (opensm, srp), configure them as described in
“Configuring the InfiniPath Drivers” on page 5-11 and “OpenFabrics Drivers
and Services Configuration and Startup” on page 5-16.
9.Check the system state by observing the LEDs. See “LED Link and Data
Indicators” on page 5-29.
3-2IB0056101-00 G
A
3 – Step-by-Step Installation Checklist
Software Installation
10.Optimize your adapter for the best performance. See “Adapter Settings” on
page 5-30. Also see the Performance and Management Tips section in the
QLogic HCA and InfiniPath Software User Guide.
11.Perform the recommended health checks. See “Customer Acceptance
Utility” on page 5-31.
12.After installing the InfiniPath and OpenFabrics software, refer to the QLogic HCA and InfiniPath Software User Guide for more information about using
InfiniPath, QLogic MPI, and OpenFabrics products.
This section lists the requirements and provides instructions for installing the
QLogic InfiniPath Interconnect adapters. Instructions are included for the QLogic
DDR PCI Express adapters, the QLE7240 and QLE7280; the QLogic InfiniPath
PCIe adapter and PCIe riser card, QLE7140; and the QHT7040 or QHT7140
adapter hardware and HTX riser card. These components are collectively referred
to as the adapter and the riser card in the remainder of this document.
The adapter is a low-latency, high-bandwidth, high message rate cluster
interconnect for InfiniBand. The QLogic interconnect is InfiniBand 4X, with a raw
data rate of 20Gbps (data rate of 16Gbps) for the QLE7240 and QLE7280; and
10Gbps (data rate of 8Gbps) for the QLE7140, QHT7040, and QHT7140.
OpenFabrics is interoperable with other vendors’ InfiniBand Host Channel
Adapters (HCAs) running compatible OpenFabrics releases.
Hardware Installation Requirements
This section lists hardware and software environment requirements for installing
the QLogic QLE7240, QLE7280, QLE7140, QHT7040, or QHT7140.
Hardware
QLogic interconnect adapters are for use with UL listed computers. The following
statement is true for all the adapters:
This device complies with part 15 of the FCC Rules. Operation is subject to
the following two conditions: (1) This device may not cause harmful
interference, and (2) this device must accept any interference received,
including interference that may cause undesired operations.
Different adapter cards work on different platforms. Tab le 4 -1 shows the
relationship between the adapter model and different types of motherboards.
Table 4-1. Adapter Models and Related Platforms
QLogic
Model
Number
QLE7240PCI Express systemsStandard PCI Express x8 or x16 slot
Table 4-1. Adapter Models and Related Platforms (Continued)
QLogic
Model
Number
QLE7140PCI Express systemsStandard PCI Express x8 or x16 slot
S
PlatformPlugs Into
Form Factors
QHT7040
QHT7140
Installation of the QLE7240, QLE7280, QLE7140, QHT7040, or QHT7140 in a 1U
or 2U chassis requires the use of a riser card. See Figure 4-4 for an illustration of
a PCI Express (PCIe) slot in a typical motherboard. See Figure 4-7 for an
illustration of an HTX slot for a typical Opteron motherboard.
The motherboard vendor is the optimal source for information on the layout and
use of HyperTransport and PCI Express-enabled expansion slots on supported
motherboards.
The QLE7240, QLE7280, and QLE7140 are the model numbers for the adapters
that ship in the standard PCI Express half-height, short-form factor. These
adapters can be used with either full-height or low-profile face plates.
The QHT7040 is the model number for the adapter that shipped in the HTX
full-height factor. The HTX low-profile form factor is referred to as the QHT7140. It
is the same as the QHT7040, except for its more compact size. In either case, the
adapter is backward and forward compatible for the motherboards in which it is
supported. The QHT7040 and QHT7140 HTX adapters are collectively referred to
as the QHT7140 unless otherwise stated.
Motherboards with HTX connectors
Motherboards with HTX connectors
HyperTransport HTX slot
HyperTransport HTX slot
When the QHT7040 or QHT7140 adapter is installed with the riser card, it may
prevent some or all of the other PCI expansion slots from being used, depending
on the form factor of the adapter and motherboard.
Run i
path_control -i to see information on which form adapter is installed.
The file
same information. For more information, see the Troubleshooting appendix in the
QLogic HCA and InfiniPath Software User Guide.
4-2IB0056101-00 G
/sys/bus/pci/drivers/ib_ipath/00/boardversion contains the
A
Cabling and Switches
The cable installation uses a standard InfiniBand (IB) 4X cable. Any InfiniBand
cable that has been qualified by the vendor should work. For SDR, the longest
passive copper IB cable that QLogic has currently qualified is 20 meters. For
DDR-capable adapters and switches, the DDR-capable passive copper cables
cannot be longer than 10 meters. Active cables can eliminate some of the cable
length restrictions.
InfiniBand switches are available through QLogic.
NOTE:
If you are using the QLE7240 or QLE7280 and want to use DDR mode, then
DDR-capable switches must be used.
The copper cables listed in Table 4-2 are available from QLogic:
4 – Hardware Installation
Hardware Installation Requirements
Table 4-2. QLogic InfiniBand Cables
Product Number Description
7104-1M-Cable4x-4x cable— 1 meter
7104-2M-Cable4x-4x cable— 2 meters
7104-3M-Cable4x-4x cable— 3 meters
7104-4M-Cable4x-4x cable— 4 meters
7104-5M-Cable4x-4x cable— 5 meters
7104-6M-Cable4x-4x cable— 6 meters
7104-7M-Cable4x-4x cable— 7 meters
7104-8M-Cable4x-4x cable— 8 meters
7104-9M-Cable4x-4x Cable— 9 meters
7104-10M-Cable4x-4x cable—10 meters
7104-12M-Cable4x-4x cable—12 meters (SDR only)
7104-14M-Cable4x-4x cable—14 meters (SDR only)
7104-16M-Cable4x-4x cable—16 meters (SDR only)
7104-18M-Cable4x-4x cable—18 meters (SDR only)
For cabling instructions, see “Cabling the Adapter to the InfiniBand Switch” on
The QLogic adapter also supports connection to the switch by means of optical
fibres through optical media converters such as the EMCORE™ QT2400. Not all
switches support these types of convertors. For more information on the
EMCORE convertor, see www.emcore.com
®
Intel
and Zarlink™ also offer optical cable solutions. See www.intel.com and
www.zarlink.com
for more information.
Configuring the BIOS
To achieve the best performance with QLogic adapters, you need to configure
your BIOS with specific settings. The BIOS settings, which are stored in
non-volatile memory, contain certain parameters characterizing the system. These
parameters may include date and time, configuration settings, and information
about the installed hardware.
There are two issues concerning BIOS settings of which you need to be aware:
Advanced Configuration and Power Interface (ACPI) needs to be enabled.
S
.
Memory Type Range Registers (MTRR) mapping needs to be set to
“Discrete”.
MTRR is used by the InfiniPath driver to enable write combining to the on-chip
transmit buffers. This option improves write bandwidth to the QLogic chip by
writing multiple words in a single bus transaction (typically 64 bytes). This option
applies only to x86_64 systems.
However, some BIOS’ do not have the MTRR mapping option. It may have a
different name, depending on the chipset, vendor, BIOS, or other factors. For
example, it is sometimes referred to as 32 bit memory hole, which must be
enabled.
You can check and adjust these BIOS settings using the BIOS Setup utility. For
specific instructions, follow the hardware documentation that came with your
system.
QLogic also provides a script, ipath_mtrr, which sets the MTRR registers.
These registers enable maximum performance from the InfiniPath driver. This
script can be run after the InfiniPath software has been installed. It needs to be
run after each system reboot.
For more details, see “BIOS Settings” on page A-1.
4-4IB0056101-00 G
A
Safety with Electricity
Observe these guidelines and safety precautions when working around computer
hardware and electrical equipment:
Locate the power source shutoff for the computer room or lab where you are
working. This is where you will turn OFF the power in the event of an
emergency or accident. Never assume that power has been disconnected
for a circuit; always check first.
Do not wear loose clothing. Fasten your tie or scarf, remove jewelry, and roll
up your sleeves. Wear safety glasses when working under any conditions
that might be hazardous to your eyes.
Shut down and disconnect the system’s power supply from AC service
before you begin work, to insure that standby power is not active. Power off
all attached devices such as monitors, printers, and external components.
Note that many motherboards and power supplies maintain standby power
at all times. Inserting or removing components while standby is active can
damage them.
4 – Hardware Installation
Safety with Electricity
Use normal precautions to prevent electrostatic discharge, which can
damage integrated circuits.
Unpacking Information
This section provides instructions for safely unpacking and handling the QLogic
adapter. To avoid damaging the adapter, always take normal precautions to avoid
electrostatic discharge.
Verify the Package Contents
The QLogic adapter system should arrive in good condition. Before unpacking,
check for any obvious damage to the packaging. If you find any obvious damage
to the packaging or to the contents, please notify your reseller immediately.
List of the Package Contents
The package contents for the QLE7240 adapter are:
QLogic QLE7240
Additional short bracket
Quick Start Guide
Standard PCIe risers can be used, typically supplied by your system or
motherboard vendor.
IB0056101-00 G4-5
4 – Hardware Installation
Unpacking Information
The package contents for the QLE7280 adapter are:
QLogic QLE7280
Additional short bracket
Quick Start Guide
Standard PCIe risers can be used, typically supplied by your system or
motherboard vendor.
The package contents for the QLE7140 adapter are:
QLogic QLE7140
Quick Start Guide
Standard PCIe risers can be used, typically supplied by your system or
motherboard vendor. The contents are illustrated in Figure 4-2.
The package contents for the QHT7140 adapter are:
QLogic QHT7140
HTX riser card for use in 1U or 2U chassis
Quick Start Guide
S
The contents are illustrated in Figure 4-3.
The IBA6120, IBA6110, and IBA7220 are the QLogic ASICs, which are the central
components of the interconnect. The location of the IBA7220 ASIC on the adapter
is shown in Figure 4-1. The location of the IBA6120 ASIC on the adapter is shown
in Figure 4-2. The location of the IBA6110 ASIC on the adapter is shown in
Figure 4-3.
4-6IB0056101-00 G
A
InfiniBand
Connector
Face Plate
IBA7220 ASIC
PCI Express
Edge Connectors
PCI Express
Edge Connectors
InfiniBand
Connector
Face Plate
IBA6120 ASIC
PCI Express Riser Card. Not
supplied; shown for reference.
4 – Hardware Installation
Unpacking Information
Figure 4-1. QLogic QLE7280 with IBA7220 ASIC
Figure 4-2. QLogic QLE7140 Card with Riser, Top View
IB0056101-00 G4-7
4 – Hardware Installation
HTX Riser Card
InfiniBand
Connector
QHT7040 Full Height
Short Card
Face Plate
IBA6110 ASIC
HTX Edge
Connectors
InfiniBand
Connector
Face Plate
QHT7140 Low
Profile Card
Unpacking Information
S
PathScale
PathScale
Figure 4-3. QLogic QHT7040/QHT7140 Full and Low Profile Cards with Riser, Top
View
Unpacking the QLogic Adapter
Follow these steps when unpacking the QLogic adapter:
1.When unpacking, ground yourself before removing the QLogic adapter from
the anti-static bag.
2.Grasping the QLogic adapter by its face plate, pull the adapter out of the
anti-static bag. Handle the adapter only by its edges or the face plate. Do not
allow the adapter or any of its components to touch any metal parts.
3.After checking for visual damage, store the adapter and the riser card in their
anti-static bags until you are ready to install them.
4-8IB0056101-00 G
4 – Hardware Installation
A
Hardware Installation
Hardware Installation
This section contains hardware installation instructions for the QLE7240,
QLE7280, QLE7140, QHT7040, and QHT7140.
Hardware Installation for QLE7240, QLE7280, or QLE7140
with PCI Express Riser
Installation for the QLE7240, QLE7280, and QLE7140 is similar. The following
instructions are for the QLE7140, but can be used for any of these three adapters.
Most installations will be in 1U and 2U chassis, using a PCIe right angle riser card.
This results in an installation of the adapter that is parallel to the motherboard.
This type of installation is described first. Installation in a 3U chassis is described
in “Hardware Installation for the QHT7140 Without an HTX Riser” on page 4-16.
Installing the QLogic QLE7140 in a 1U or 2U chassis requires a PCIe right angle
riser card.
A taller riser card can be used if necessary. The QLE7140 can connect to any of
the standard compatible PCI Express riser cards.
Dual Adapter Installation
If you have a motherboard with dual PCIe slots, dual adapters can be installed.
The adapters must match. For example, on a motherboard with two x16 slots,
dual QLE7280 adapters can be installed, but not a QLE7240 adapter and a
QLE7280 adapter. Check the design of your motherboard to see how riser cards
can be used.
Follow the instructions in “Installation Steps” on page 4-9.
See the Using MPI section in the QLogic HCA and InfiniPath Software User Guide
for information on using the IPATH_UNIT environment variable to control which
HCA to use.
Installation Steps
To install the QLogic adapter with a PCIe riser card:
1.The BIOS should already be configured properly by the motherboard
manufacturer. However, if any additional BIOS configuration is required, it
will usually need to be done before installing the QLogic adapter. See
“Configuring the BIOS” on page 4-4.
2.Shut down the power supply to the system into which you will install the
QLogic adapter.
3.Take precautions to avoid electrostatic damage (ESD) to the cards by
properly grounding yourself or touching the metal chassis to discharge static
electricity before handling the cards.
IB0056101-00 G4-9
4 – Hardware Installation
PCIe slot in
typical motherboard
Hardware Installation
4.Remove the cover screws and cover plate to expose the system’s
motherboard. For specific instructions on how to do this, follow the hardware
documentation that came with your system.
5.Locate the PCIe slot on your motherboard. Note that the PCIe slot has two
separate sections, with the smaller slot opening located towards the front
(see Figure 4-4). These two sections correspond to the shorter and longer
connector edges of the adapter and riser.
S
Figure 4-4. PCIe Slot in a Typical Motherboard
6.Determine if a blanking panel is installed in your chassis. If it is, remove it so
that the InfiniBand connector will be accessible. Refer to your system vendor
instructions for how to remove the blanking panel.
7.Remove the QLogic adapter from the anti-static bag.
8.Locate the face plate on the connector edge of the card.
4-10IB0056101-00 G
A
PCIe Riser Card
QLogic Adapter
LEDs
Face Plate
InfiniBand connector
4 – Hardware Installation
Hardware Installation
9.Connect the QLogic adapter and PCIe riser card together, forming the
assembly that you will insert into your motherboard. First, visually line up the
adapter slot connector edge with the edge connector of the PCIe riser card
(see Figure 4-5).
.
Figure 4-5. QLogic PCIe HCA Assembly with Riser Card
10.Holding the QLogic adapter by its edges, carefully insert the card slot
connector into the PCIe riser card edge connector, as show in Figure 4-5.
The result is a combined L-shaped assembly of the PCIe riser card and
QLogic adapter. This assembly is what you will insert into the PCIe slot on
the motherboard in the next step.
11.Turn the assembly so that the riser card connector edge is facing the PCIe
slot on the motherboard, and the face plate is toward the front of the chassis.
12.Holding this assembly above the motherboard at about a 45 degree angle,
slowly lower it so that the connector on the face plate clears the blanking
panel opening of the chassis from the inside. Slowly align the connector
edge of the riser card with the motherboard’s PCIe slot. The short section of
the connector must align with the short section of the slot.
IB0056101-00 G4-11
4 – Hardware Installation
Hardware Installation
13.Insert the riser assembly into the motherboard’s PCIe slot, ensuring good
contact. The QLogic adapter should now be parallel to the motherboard and
about one inch above it (see Figure 4-6).
S
Figure 4-6. Assembled PCIe HCA with Riser
14.Secure the face plate to the chassis. The QLogic adapter has a screw hole
on the side of the face plate that can be attached to the chassis with a
retention screw. The securing method may vary depending on the chassis
manufacturer. Refer to the system documentation for information about
mounting details such as mounting holes, screws to secure the card, or
other brackets.
The QLogic PCIe HCA with PCIe riser card is now installed. Next, install the
cables as described in “Cabling the Adapter to the InfiniBand Switch” on
page 4-17. Then test your installation by powering up and verifying link status (see
“Completing the Installation” on page 4-18).
Hardware Installation for QHT7140 with HTX Riser
Most installations will be in 1U and 2U chassis, using the HTX riser card. This
results in a horizontal installation of the QHT7140. This type of installation is
described in this section. Installation in a 3U chassis is described in “Hardware
Installation for the QHT7140 Without an HTX Riser” on page 4-16.
Installation of QLogic QHT7140 in a 1U or 2U chassis requires an HTX riser card.
NOTE:
The illustrations in this section are shown for the full height short form factor.
Installation of the HTX low profile form factor follows the same steps.
4-12IB0056101-00 G
A
HTX Slot in a Typical Opteron Motherboard
4 – Hardware Installation
Hardware Installation
To install the QLogic adapter with an HTX riser card:
1.The BIOS should be already be configured properly by the motherboard
manufacturer. However, if any additional BIOS configuration is required, it
will usually need to be done before installing the QLogic adapter. See
“Configuring the BIOS” on page 4-4.
2.Shut down the power supply to the system into which you will install the
QLogic adapter.
3.Take precautions to avoid electrostatic discharge (ESD) damage to the
cards by properly grounding yourself or touching the metal chassis to
discharge static electricity before handling the cards.
4.Remove the cover screws and cover plate to expose the system’s
motherboard. For specific instructions on how to do this, follow the hardware
documentation that came with your system.
5.Locate the HTX slot on your motherboard. Note that the HTX slot has two
separate connectors, corresponding to the connector edges of the adapter.
See Figure 4-7.
Figure 4-7. HTX Slot
6.Determine if a blanking panel is installed in your chassis. If it is, remove it so
that the InfiniBand connector will be accessible. Refer to your system vendor
instructions for how to remove the blanking panel.
IB0056101-00 G4-13
4 – Hardware Installation
HTX Riser Card
InfiniBand Connector
LEDs
Face Plate
QLogic Adapter
Hardware Installation
7.Remove the QLogic QHT7140 from the anti-static bag.
NOTE:
8.Locate the face plate on the connector edge of the card.
9.Connect the QLogic adapter and HTX riser card together, forming the
assembly that you will insert into your motherboard. First, visually line up the
adapter slot connector edge with the edge connector of the HTX riser card
(see Figure 4-8).
S
Be careful not to touch any of the components on the printed circuit
board during these steps. You can hold the adapter by its face plate or
edges.
Figure 4-8. QLogic QHT7140 Adapter with Riser Card
10.Holding the QLogic adapter by its edges, carefully insert the card slot
connector into the HTX riser card edge connector, as show in Figure 4-8.
The result is a combined L-shaped assembly of the HTX riser card and
QLogic adapter. This assembly is what you will insert into the HTX slot on
the motherboard in the next step.
11.Turn the assembly so that the riser card connector edge is facing the HTX
slot on the motherboard, and the face plate is toward the front of the chassis.
12.Holding this assembly above the motherboard at about a 45 degree angle,
slowly lower it so that the connector on the face plate clears the blanking
panel opening of the chassis from the inside. Slowly align the connector
edge of the HTX riser card with the motherboard’s HTX slot. The HTX riser
and HTX slot must line up perfectly.
4-14IB0056101-00 G
A
4 – Hardware Installation
Hardware Installation
13.Insert the HT riser assembly into the motherboard’s HTX slot, ensuring good
contact. The QLogic adapter should now be parallel to the motherboard and
about one inch above it, as shown in Figure 4-9.
Figure 4-9. Assembled QHT7140 with Riser
14.Secure the face plate to the chassis. The QLogic adapter has a screw hole
on the side of the face plate that can be attached to the chassis with a
retention screw. The securing method may vary depending on the chassis
manufacturer. Refer to the system documentation for information about
mounting details such as mounting holes, screws to secure the card, or
other brackets.
The QLogic QHT7140 with HTX riser card is now installed. Next, install the cables
as described in “Cabling the Adapter to the InfiniBand Switch” on page 4-17. Then
test your installation by powering up and verifying link status (see “Completing the
Installation” on page 4-18).
Hardware Installation for QLE7240, QLE7280, and QLE7140
Without a PCI Express Riser
Installing the QLogic QLE7240, QLE7280, or QLE7140 without a PCI Express
riser card requires a 3U or larger chassis.
Installation is similar to the QHT7140 HTX adapter, except that the card slot
connectors on these adapters fit into the PCIe slot rather than the HTX slot.
Follow the instructions in “Hardware Installation for the QHT7140 Without an HTX
Riser” on page 4-16, substituting the PCIe slot for the HTX slot.
IB0056101-00 G4-15
4 – Hardware Installation
Hardware Installation
S
Hardware Installation for the QHT7140 Without an HTX Riser
Installing the QLogic QHT7140 without an HTX riser card requires a 3U or larger
chassis. The card slot connectors on the QHT7140 fit into the HTX slot in a
vertical installation.
To install the QLogic adapter without the HTX riser card:
1.The BIOS should already be configured properly by the motherboard
manufacturer. However, if any additional BIOS configuration is required, it
will usually need to be done before installing the QLogic adapter. See
“Configuring the BIOS” on page 4-4.
2.Shut down the power supply to the system into which you will install the
QLogic adapter.
3.Take precautions to avoid electrostatic discharge (ESD) damage to the
cards by properly grounding yourself or touching the metal chassis to
discharge static electricity before handling the cards.
4.If you are installing the QLogic adapter into a covered system, remove the
cover screws and cover plate to expose the system’s motherboard. For
specific instructions on how to do this, follow the hardware documentation
that came with your system.
5.Locate the HTX slot on your motherboard (see Figure 4-7).
6.Remove the QLogic adapter from the anti-static bag. Hold the card by the
top horizontal section of the bracket, and the top rear corner of the card. Be
careful not to touch any of the components on the printed circuit card.
7.Without fully inserting, gently align and rest the HTX card’s gold fingers on
top of the motherboard’s HTX slot.
4-16IB0056101-00 G
A
4 – Hardware Installation
Switch Configuration and Monitoring
8.Insert the card by pressing firmly and evenly on the top of the horizontal
bracket and the top rear corner of the card simultaneously. The card should
insert evenly into the slot. Be careful not to push, grab, or put pressure on
any other part of the card, and avoid touching any of the components. See
Figure 4-10.
Figure 4-10. QHT7140 Without Riser Installed in a 3U Chassis
9.Secure the face plate to the chassis. The QLogic adapter has a screw hole
on the side of the face plate that can be attached to the chassis with a
retention screw. The securing method may vary depending on the chassis
manufacturer. Refer to the system documentation for information about
mounting details such as mounting holes, and screws to secure the card, or
other brackets.
Next, install the cables, as described in “Cabling the Adapter to the InfiniBand
Switch” on page 4-17. Then test your installation by powering up the system (see
“Completing the Installation” on page 4-18).
Switch Configuration and Monitoring
The QLogic interconnect is designed to work with all InfiniBand-compliant
switches. Follow the vendor documentation for installing and configuring your
switches.
Cabling the Adapter to the InfiniBand Switch
Follow the recommendations of your cable vendor for cable management and
proper bend radius.
IB0056101-00 G4-17
4 – Hardware Installation
Completing the Installation
The QLE7240, QLE7280, QLE7140, QHT7040, and QHT7140 adapters are all
cabled the same way.
To install the InfiniBand cables:
1.Check that you have removed the protector plugs from the cable connector
ends.
2.Different vendor cables might have different latch mechanisms. Determine if
your cable has a spring-loaded latch mechanism.
If your cable is spring-loaded, grasp the metal shell and pull on the
If your cable latch mechanism is not spring-loaded, push on the metal
3.The InfiniBand cables are symmetric; either end can be plugged into the
switch. Connect the InfiniBand cable to the connector on the QLogic
QLE7240, QLE7280, QLE7140 or QHT7140. Depress the side latches of the
cable when connecting. (On some cables, this latch is located at the top of
the cable connector.) Make sure the lanyard handle on the cable connector
is slid forward toward the card connector until fully engaged.
S
plastic latch to release the cable. To insert, push and the cable snaps
into place. You will hear a short “click” sound from the cable connector
when it snaps in.
case, then push the plastic latch to lock the cable in place.
4.Connect the other end of the cable to the InfiniBand switch.
Completing the Installation
To complete the hardware installation:
1.Complete any other installation steps for other components.
2.Replace the cover plate and back panel.
3.Verify that the power cable is properly connected.
4.Turn on the power supply and boot the system normally.
5.Watch the LED indicators. The LEDs will flash only once, briefly, at
power-up. The LEDs are functional only after the InfiniPath software has
been installed, the driver has been loaded, and the system is connected to
an InfiniBand switch. To use the LEDs to check the state of the adapter, see
“LED Link and Data Indicators” on page 5-29.
4-18IB0056101-00 G
5Software Installation
This section provides instructions for installing the InfiniPath and OpenFabrics
software. The InfiniPath software includes drivers, protocol libraries, QLogic’s
implementation of the MPI message passing standard, and example programs,
including benchmarks. A complete list of the provided software is in “Software
Components” on page 2-4.
Cluster Setup
Information on clusters, supported distributions and kernels, and environment
setup is provided in “Types of Nodes in a Cluster Environment” on page 5-1,
“Supported Linux Distributions” on page 5-2, and “Setting Up Your Environment”
on page 5-2.
Types of Nodes in a Cluster Environment
In a cluster environment, different nodes can be used for different functions, such
as launching jobs, developing software, or running jobs. The nodes are defined as
follows:
Front end node. This node launches jobs.
Compute node. This node runs jobs.
Developmentorbuild node. These are the machines on which examples
or benchmarks can be compiled.
Any machine can serve any combination of these three purposes, but a typical
cluster has many compute nodes and just a few (or only one) front end nodes.
The number of nodes used for development will vary. These node names are
used throughout this guide.
IB0056101-00 G5-1
5 – Software Installation
Cluster Setup
Supported Linux Distributions
The currently supported distributions and associated Linux kernel versions for
InfiniPath and OpenFabrics are listed in Table 5-1. The kernels are the ones that
shipped with the distributions.
Table 5-1. InfiniPath/OpenFabrics Supported Distributions and Kernels
Scientific Linux 5.0, 5.12.6.18, 2.6.18-53 (x86_64)
SUSE
(SLES 10 GM, SP 1)
NOTE:
®
Enterprise Linux® 4.4, 4.5, 4.6
®
Linux Enterprise Server
Fedora Core 4 and Fedora Core 5 are not supported in the InfiniPath 2.2
release.
InfiniPath/OpenFabrics Supported
Kernels
2.6.9-42 (U4), 2.6.9-55 (U5)
2.6.9-67 (U6) (x86_64)
2.6.18-8, 2.6.18-53 (x86_64)
2.6.16.21, 2.6.16.46 (x86_64)
Setting Up Your Environment
Keep the following in mind when setting up the environment:
The kernel-devel (for Red Hat and Red Hat-derived kernels) RPM or
kernel-source (for SLES) RPM for your distribution/kernel must be
installed.
The runtime and build environments must be the same. Compatibility
between executables built on different Linux distributions cannot be
guaranteed.
Install sysfsutils for your distribution before installing the OpenFabrics
RPMs, as there are dependencies. Check your distribution’s documentation
for information about sysfsutils. This package is called udev on
SLES 10.
5-2IB0056101-00 G
A
5 – Software Installation
Downloading and Unpacking the InfiniPath and OpenFabrics Software
Make sure that all previously existing (stock) OpenFabrics RPMs are
uninstalled. See “Removing Software Packages” on page 5-33 for more
information on uninstalling. If you are using RHEL5, make sure that
opensm-* is manually uninstalled. See “Version Number Conflict with
opensm-* on RHEL5 Systems” on page A-4 for more information.
Among the many optional packages that each distribution offers, the
InfiniPath software requires openssh and openssh-server and, if the
Multi-Purpose Daemon (MPD) job launcher or the ipath_mtrr script is to
be used, python. These packages must be on every node. Note that in the
SLES 10 distribution, openssh-server is a part of the openssh package.
It is possible to have a cluster running with different kernel versions.
However, QLogic recommends and supports clusters where all nodes run
equivalent software.
Different distributions require different versions of the InfiniPath software
distribution, as described in “Installing the InfiniPath and OpenFabrics
RPMs” on page 5-6.
Downloading and Unpacking the InfiniPath and
OpenFabrics Software
This section assumes that the correct Linux kernel, a supported distribution, and
suggested packages (see “Setting Up Your Environment” on page 5-2) have been
installed on every node.
Choosing the Appropriate Download Files
Several components are available as separate downloads, as noted in Tab le 5 -2 .
All files are available from the QLogic web site: http://www.qlogic.com
Table 5-2. Available Packages
PackageDescriptionComments
InfiniPath 2.2
software
QLogic SRP QLogic’s version of SRP.
Multiple interdependent RPM
packages make up the InfiniPath
and OpenFabrics software.
Includes QLogic VNIC.
Available as a separate download.
Associated documentation
includes Readme and Release
Notes for InfiniPath, and the
QLogic HCA and InfiniPath software installation and user’s
guides.
Has associated Readme,
Release Notes, and documentation.
.
IB0056101-00 G5-3
5 – Software Installation
Downloading and Unpacking the InfiniPath and OpenFabrics Software
Table 5-2. Available Packages (Continued)
PackageDescriptionComments
S
Enablement
tools
QLogic Fabric
Suite
OFED 1.3
source
Subset of QLogic InfiniBand Fabric Suite.
Available as a separate download.
The QLogic Fabric Suite CD is
available separately for purchase. This CD provides management tools and the QLogic
host-based SM.
Optional source for the Open Fabrics Enterprise Distribution
(OFED 1.3) libraries and utilities
as built and shipped with the
InfiniPath 2.2 release. Packaged
as a single tar file.
Has associated Readme,
Release Notes, and documentation.
CD may be purchased separately. Follow the links on the
QLogic download page. Documentation is included.
Documentation is included in
the tar file.
Follow the links for your distribution to the desired download, then follow the
instructions on the web page for downloading the files to a convenient directory.
Instructions for installing the InfiniPath and OpenFabrics software are provided in
the following sections.
QLogic SRP and the enablement tools have separate instructions that are
available from the download page on the QLogic web site.
NOTE:
Installation information for the QLogic Fabric Suite and the OFED 1.3 source
packages is included with the packages.
Unpacking the InfiniPath tar File
After downloading the InfiniPath 2.2 tar file, type:
$ tar zxvf InfiniPath2.2-xxx-yyy.tgz
In this command, xxx is the distribution identifier (RHEL4 or SLES10), and yyy is
the platform architecture,
on the
5-4IB0056101-00 G
tar file name and places the RPMs and other files in this directory.
x86_64. The tar command creates a directory based
A
5 – Software Installation
Downloading and Unpacking the InfiniPath and OpenFabrics Software
NOTE:
The files can be downloaded to any directory. The install process will create
and install the files in the correct directories. The locations of the directories
after installation are listed in “Installed Layout” on page 5-9.
A complete list of RPMs is in “RPM Descriptions” on page C-1.
Check for Missing Files or RPMs
Run the rpm command with the --verifyoption to check if there are files
missing from the RPMs. For example:
$ rpm -a --verify ’InfiniPath-MPI/mpi*’ ’InfiniPath/infinipath*’
In the case of OpenFabrics RPMs, the verification command is slightly different,
since these RPMs have many different prefixes:
$ rpm --verify rpm_name_pre
rpm_name_pre is the descriptive name that precedes the version and repository
identifiers in an RPM. For example:
$ rpm --verify libibverbs
This command verifies the contents of:
libibverbs-2.2-xxx.1_1.yyy.x86_64.rpm
The rpm command cannot check for missing RPMs. Use ipath_checkout after
installation to flag missing RPMs. See “Adapter Settings” on page 5-30 for more
information.
IB0056101-00 G5-5
5 – Software Installation
Installing the InfiniPath and OpenFabrics RPMs
S
Installing the InfiniPath and OpenFabrics RPMs
Linux distributions of InfiniPath and OpenFabrics software are installed from
binary RPMs. RPM is a Linux packaging and installation tool used by Red Hat,
SUSE, and CentOS. There are multiple interdependent RPM packages that make
up the InfiniPath and OpenFabrics software.
The OpenFabrics kernel module support is now part of the InfiniPath RPMs. This
kernel module includes support for kernel-based protocols such as IPoIB, SRP,
the multicast communications manager, and the subnet administration agent.
Programs that incorporate the user IB verbs interfaces, such as diagnostics,
benchmarks, verbs-based MPIs (for example, Intel MPI Version 3.0), and SDP
sockets must have RPMs installed in the OpenFabrics directories. See
Appendix C “RPM Descriptions” for a complete description of the RPMs’ contents.
Each set of RPMs uses a build identifier xxx and a distribution identifier yyy.
The distribution identifiers for this release are
distribution identifiers are listed in Table 5-3 with the associated operating
systems.
rhel4 and sles10. The RPM
Table 5-3. InfiniPath and OpenFabrics RPMs to Use for Each Node in a
Cluster
Distribution
Identifier
rhel4Fedora 6, Red Hat Enterprise Linux 4.4 (RHEL4.4), RHEL4.5,
RHEL 4.6, CentOS 4.4-4.6 (Rocks 4.4-4.6), Scientific Linux 4.4-4.6,
RHEL5.0-5.1, CentOS 5.0-5.1, Scientific Linux 5.0-5.1, for x86_64
systems
sles10SLES 10 and SLES 10 SP1 for x86_64 systems
NOTE:
RPMs contain config files. Your current config files will not be
overwritten when the new RPMs are installed. New config files contain the
suffix .rpmnew and can be found in the /etc/sysconfig directory.
Check the new files to see if there is anything you want to add to your
standard config files.
Used On
5-6IB0056101-00 G
A
Choosing the RPMs to Install
Although QLogic recommends that all RPMs are installed on all nodes, some are
optional depending on which type of node is being used. To see which RPMs are
required or optional for each type of node, according to its function as a compute
node, front end node, development machine, or Subnet Manager (SM), see
Appendix C “RPM Descriptions”.
Installing all the RPMs (except for OpenSM) on all nodes is easiest.
If you do not plan to run all of the available programs, you can install a
subset of the RPMs. See Tab le 5 -4 .
5 – Software Installation
Installing the InfiniPath and OpenFabrics RPMs
Install the
host-based SM. The
OpenSMRPM only if you do not plan to use a switch-based or
OpenSM RPM is normally installed on the node on which
it will be used. If installed, it is on by default. This behavior can be modified.
See “OpenSM” on page 5-23 for more information.
Install the infinipath RPM on the all nodes where you install the
mpi-frontend RPM.
Table 5-4. RPMS to Install
If you want to:Then install:
Use InfiniPath, QLogic MPI, and basic
OpenFabrics protocols such as IPoIB
and SRP
Run OpenFabrics programs that use the
IB verbs interfaces, such as diagnostics,
benchmarks, verbs-based MPI, VNIC,
and SDP sockets
Use all of the above programsThe RPM sets in the InfiniPath*,
Add OpenSM as a your subnet managerThe RPM sets in the OpenFabrics and
The RPM sets in the InfiniPath* and
InfiniPath-MPI directories
The RPM sets in the InfiniPath* and the
OpenFabrics* directories
InfiniPath-MPI, and the OpenFabrics* directories
OpenSM directories. (OpenSM has depen-
dencies on OpenFabrics.)
Use other MPIs (MVAPICH and Open
MPI) that have been compiled with PathScale, GNU, PGI, and Intel compilers.
Includes test files.
Use other HCAs with OpenFabricsThe RPM sets in the OpenFabrics* and
IB0056101-00 G5-7
The RPM sets in the InfiniPath*, InfiniPath-MPI, and OtherMPIs directories
OtherHCAs* directories
5 – Software Installation
Installing the InfiniPath and OpenFabrics RPMs
Using rpm to Install InfiniPath and OpenFabrics
The RPMs need to be available on each node on which they will be used. One
way to do this is to copy the RPMs to a directory on each node that will need
them. Another way is to put the RPMs in a directory that is accessible (e.g., via
Network File System (NFS)) to every node. After making sure the RPMs are
available on each node, login as root and, for the
Install the OpenSM RPM only if you do not plan to use a switch-based or
host-based SM. The
be used. If installed, it is on by default. This behavior can be modified. See
“OpenSM” on page 5-23 for more information.
# rpm -Uvh OpenSM/*.rpm
The opensm-devel RPM is located with the other OpenFabrics-Devel RPMs.
The InfiniPath kernel modules in this release are installed in:
/lib/modules/‘uname -r‘/updates
Putting the modules in this directory avoids replacing kernel modules that may be
provided by your Linux distribution; you may want to use these modules if the
InfiniPath software is removed. Modules are renamed if they can cause conflicts.
For example, the module
ib_ipath.ko, but conflicts can arise if ipath_core.ko is still present. If it is found
during installation of the infinipath-kernel RPM,
ipath_core.ko.bak.
ipath_core.ko was previously renamed to
ipath_core.ko is renamed
IB0056101-00 G5-9
5 – Software Installation
Starting the InfiniPath Service
Other OFED-installed modules may also be in this directory; these are also
renamed if found during the install process.
Starting the InfiniPath Service
If this is the initial installation of InfiniPath (see “Lockable Memory Error on Initial
Installation of InfiniPath” on page A-7), or if you have installed VNIC with the
OpenFabrics RPM set, reboot after installing.
If this is an upgrade, you can restart the InfiniPath service without rebooting. To
enable the driver, run the command (as root):
# chkconfig infinipath on 2345
Then start the InfiniPath support (as root) by typing:
# /etc/init.d/infinipath start
Complete information about starting, stopping, and restarting the services are in
“Starting and Stopping the InfiniPath Software” on page 5-26.
S
When all of the InfiniPath and OpenFabrics software has been installed correctly,
the default settings at startup are as follows:
InfiniPath
InfiniPath
the ipath_ether Network Interface” on page 5-12 for configuration
instructions.
OpenFabrics IPoIB is not running until it is configured. Once enabled, IPoIB
is running in connected mode. See “Configuring the IPoIB Network
Interface” on page 5-17 for configuration instructions.
OFED SRP is not running until the module is loaded and the SRP devices
on the fabric have been discovered.
VNIC is not running until it is configured.
OpenSM is enabled on startup. Either install it on only one node, or disable it
on all nodes except where it will be used as an SM.
Other optional drivers can now be configured and enabled, as described in
“InfiniPath and OpenFabrics Driver Overview” on page 5-11, “Configuring the
InfiniPath Drivers” on page 5-11, and “OpenFabrics Drivers and Services
Configuration and Startup” on page 5-16.
ib_ipath is enabled.
ipath_ether is not running until it is configured. See “Configuring
5-10IB0056101-00 G
5 – Software Installation
InfiniPath
components
TCP/IP
ipath_ether
OpenS
IPoIB
OpenFabrics
components
ib_ipath
A
InfiniPath and OpenFabrics Driver Overview
InfiniPath and OpenFabrics Driver Overview
The InfiniPath ib_ipath module provides low level QLogic hardware support, and
is the base driver for both the InfiniPath and OpenFabrics software components.
The
ib_ipath module does hardware initialization, handles InfiniPath-specific
memory management, and provides services to other InfiniPath and OpenFabrics
modules.
It provides the hardware and hardware management functions for MPI/PSM
programs, the
protocols such as IPoIB and SDP. The module also contains a subnet
management agent.
Figure 5-1 shows the relationship between the InfiniPath and OpenFabrics
software. Not all components are shown.
ipath_ether Ethernet emulation, and general OpenFabrics
Figure 5-1. Relationship Between InfiniPath and OpenFabrics Software
If you want to enable Transmission Control Protocol-Internet Protocol (TCP/IP)
networking for running Ethernet traffic over the InfiniPath link, you can configure
the optional
NOTE:
It is not necessary to configure the
Optional configurable OpenFabrics components are:
IPoIB network interface
VNIC
OpenSM
SRP (OFED and QLogic modules)
MPI over uDAPL (can be used by Intel MPI or HP-MPI)
ipath_ether network interface files.
ipath_ether driver to run MPI jobs.
Configuring the InfiniPath Drivers
This section provides information on InfiniPath driver configuration.
IB0056101-00 G5-11
5 – Software Installation
Configuring the InfiniPath Drivers
Configuring the ib_ipath Driver
The ib_ipath module provides both low-level InfiniPath support and
management functions for OpenFabrics protocols. The ib_ipath driver has
several configuration variables that set reserved buffers for the software, define
events to create trace records, and set the debug level.The startup script for
ib_ipath is installed automatically as part of the software installation, and
normally does not need to be changed.
The primary configuration file for the InfiniPath drivers ib_ipath and
ipath_ether, and other modules and associated daemons, is:
/etc/sysconfig/infinipath
Normally, this configuration file is set up correctly at installation and the drivers are
loaded automatically during system boot once the RPMs have been installed. If
you are upgrading, your existing configuration files will not be overwritten.
The device files are:
/dev/ipath
/dev/ipath0, /dev/ipath1, ...
S
The numbered device files allow access to a specific InfiniPath unit.
See the
ib_ipath man page for more details.
Configuring the ipath_ether Network Interface
These instructions are for enabling TCP/IP networking over the InfiniPath link. To
enable IPoIB networking, see “Configuring the IPoIB Network Interface” on
page 5-17.
You must create a network device configuration file for the layered Ethernet
device on the QLogic adapter. This configuration file will resemble the
configuration files for the other Ethernet devices on the nodes.
Two slightly different procedures are given in the following sections for the ipath
configuration; one for Fedora/RHEL (see “ipath_ether Configuration on Red Hat”
on page 5-12) and one for SLES 10 (see “ipath_ether Configuration on SLES” on
page 5-14).
Many of the entries that are used in the configuration directions are explained in
the file
/usr/share/doc/initscripts-*/sysconfig.txt
ipath_ether Configuration on Red Hat
The following procedure will cause the ipath_ether network interfaces to be
automatically configured the next time you reboot the system. These instructions
are for the Fedora 6, Red Hat Enterprise Linux 4 (RHEL4), and RHEL5
distributions.
sysconfig.txt, located in the following directory:
5-12IB0056101-00 G
A
5 – Software Installation
Configuring the InfiniPath Drivers
Servers typically have two Ethernet devices present, numbered as 0 (eth0) and
1(
eth1). This example creates a third device, eth2.
NOTE:
When multiple QLogic HCAs are present, the configuration for
and so on follow the same format as for adding
eth2 in the following
eth3, eth4,
example.
1.Check for the number of Ethernet drivers you currently have by typing either
one of the following two commands:
$ ifconfig -a
$ ls /sys/class/net
It is assumed that two Ethernet devices (numbered 0 and 1) are already
present.
2.Edit the file
alias eth2 ipath_ether
/etc/modprobe.conf (as root); add the following line:
3.Create or edit the following file (as root):
/etc/sysconfig/network-scripts/ifcfg-eth2
If you are using Dynamic Host Configuration Protocol (DHCP), add the
following lines to
If you are using static IP addresses, use the following lines instead,
substituting your own IP address for the one in the example. The normal
matching netmask is shown.
# QLogic Interconnect Ethernet
DEVICE=eth2
BOOTPROTO=static
ONBOOT=YES
IPADDR=192.168.5.101 #Substitute your IP address here
NETMASK="255.255.255.0"#Normal matching netmask
TYPE=Ethernet
This change causes the ipath_ether Ethernet driver to be loaded and
configured during system startup. To check your configuration, and make the
ipath_ether Ethernet driver available immediately, type the following
command (as root):
# /sbin/ifup eth2
IB0056101-00 G5-13
5 – Software Installation
Configuring the InfiniPath Drivers
4.Check whether the Ethernet driver has been loaded with:
$ lsmod | grep ipath_ether
5.Verify that the driver is up with:
$ ifconfig -a
ipath_ether Configuration on SLES
The following procedure causes the ipath_ether network interfaces to be
automatically configured the next time you reboot the system. These instructions
are for the
SLES 10 distribution.
S
Servers typically have two Ethernet devices present, numbered as 0 (
and1(
NOTE:
As the root user, perform the following steps:
1.Be sure that the ipath_ether module is loaded:
2.Determine the MAC address that will be used:
eth1). This example creates a third device, eth2.
When multiple QLogic HCAs are present, the configuration for
eth4, and so on follow the same format as for adding eth2 in the
following example. Similarly, in Step 2, add one to the unit number
(replace
and so on.
The Media Access Control (MAC) address is a unique identifier attached
to most forms of networking equipment. Step 2 determines the MAC
address to use, and will be referred to as $MAC in the subsequent steps.
$MAC must be replaced in each case with the string printed in Step 2.
Make sure that you substitute your own IP address for the IPADDR in the
example. The
your network.
6.To verify that the configuration files are correct, run the commands:
# ifup eth2
# ifconfig eth2
You may have to reboot the system for the configuration changes to take
effect.
BROADCAST, NETMASK, and NETWORK lines need to match for
OpenFabrics Drivers and Services Configuration
and Startup
The IPoIB network interface, VNIC interface, and OpenSM components of
OpenFabrics can be configured to be on or off. IPoIB is off by default; VNIC is off
by default; OpenSM is on by default.
IPoIB, VNIC, OpenSM, SRP, and MPI over uDAPL configuration and startup is
explained in more detail in the following sections.
NOTE:
The following instructions work for all supported distributions.
5-16IB0056101-00 G
A
OpenFabrics Drivers and Services Configuration and Startup
Configuring the IPoIB Network Interface
The following instructions show you how to manually configure your OpenFabrics
IPoIB network interface. This example assumes that you are using
your shell, all required InfiniPath and OpenFabrics RPMs are installed, and your
startup scripts have been run (either manually or at system boot).
For this example, the IPoIB network is 10.1.17.0 (one of the networks reserved for
private use, and thus not routable on the Internet), with a /8 host portion, and
therefore requires that the netmask be specified.
This example assumes that no hosts files exist, the host being configured has the
IP address 10.1.17.3, and DHCP is not used.
NOTE:
Instructions are only for this static IP address case. Configuration methods
for using DHCP will be supplied in a later release.
1.Type the following commands (as root):
5 – Software Installation
sh or bash as
# ifconfig ib0 10.1.17.3 netmask 0xffffff00
2.To verify the configuration, type:
# ifconfig ib0
The output from this command will be similar to this:
ib0 Link encap:InfiniBand HWaddr
00:00:00:02:FE:80:00:00:00:00:00:00:00:00:00:00:00:00:00:00
inet addr:10.1.17.3 Bcast:10.1.17.255 Mask:255.255.255.0
UP BROADCAST RUNNING MULTICAST MTU:4096 Metric:1
RX packets:0 errors:0 dropped:0 overruns:0 frame:0
TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:128
RX bytes:0 (0.0 b) TX bytes:0 (0.0 b)
3.Type:
# ping -c 2 -b 10.1.17.255
The output of the ping command will be similar to the following, with a line
for each host already configured and connected:
WARNING: pinging broadcast address
PING 10.1.17.255 (10.1.17.255) 517(84) bytes of data.
174 bytes from 10.1.17.3: icmp_seq=0 ttl=174 time=0.022 ms
64 bytes from 10.1.17.1: icmp_seq=0 ttl=64 time=0.070 ms (DUP!)
64 bytes from 10.1.17.7: icmp_seq=0 ttl=64 time=0.073 ms (DUP!)
The IPoIB network interface is now configured.
IB0056101-00 G5-17
5 – Software Installation
OpenFabrics Drivers and Services Configuration and Startup
NOTE:
The configuration must be repeated each time the system is rebooted.
IPoIB-CM (Connected Mode) is enabled by default. If you want to
change this to datagram mode, edit the file
/etc/sysconfig/infinipath. Change this line to:
IPOIB_MODE="datagram"
S
Restart
# /etc/init.d/infinipath restart
The default IPOIB_MODE setting is "
infinipath (as root) by typing:
CM" for Connected Mode.
Configuring and Administering the VNIC Interface
The VirtualNIC (VNIC) Upper Layer Protocol (ULP) works in conjunction with
firmware running on Virtual Input/Output (VIO) hardware such as the SilverStorm
Ethernet Virtual I/O Controller (EVIC™) or the InfiniBand/Ethernet Bridge Module
for IBM
The VNIC interface must be configured before it can be used. To do so, perform
the following steps:
1.Discover EVIC/VEx Input/Output Controllers (IOCs) present on the fabric
OpenFabrics Drivers and Services Configuration and Startup
service[ 1]: 1000066a00000101
InfiniNIC.InfiniConSys.Data:01
.
.
.
NOTE:
A VIO hardware card can contain up to six IOCs (and therefore up to six
IOCGUIDs); one for each Ethernet port on the VIO hardware card. Each VIO
hardware card contains a unique set of IOCGUIDs: (e.g., IOC 1 maps to
Ethernet Port 1, IOC 2 maps to Ethernet Port 2, IOC 3 maps to Ethernet
Port 3, etc.).
2.Create the VNIC interfaces using the configuration file
/etc/infiniband/qlgc_vnic.cfg.
Look at the qlgcvnictools/qlgc_vnic.cfg sample to see how VNIC
configuration files are written. You can use this configuration file as the basis
for creating a configuration file by replacing the Destination Global Identifier
(DGID), IOCGUID, and IOCSTRING values with those of the EVIC/VEx
IOCs present on your fabric.
QLogic recommends using the DGID of the EVIC/VEx IOC, as it ensures the
quickest startup of the VNIC service. When DGID is specified, the IOCGUID
must also be specified. For more details, see the qlgc_vnic.cfg sample
file.
3.Edit the VirtualNIC configuration file,
/etc/infiniband/qlogic_vnic.cfg. For each IOC connection, add a
CREATE block to the file using the following format:
Do not create EIOC names with similar character strings (e.g.,
eioc3 and eioc30). There is a limitation with certain Linux
operating systems that cannot recognize the subtle differences.
The result is that the user will be unable to ping across the
network.
S
b.Format 2: Defining an IOC using the IOCSTRING. Defining the IOC
using the IOCSTRING allows VIO hardware to be hot-swapped in and
out of a specific slot. The host attempts to connect to the specified IOC
(1, 2, or 3) on the VIO hardware that currently resides in the specified
slot of the specified chassis. Use the following format to allow the host
to connect to a VIO hardware that resides in a specific slot of a specific
chassis:
The IOCSTRING field is a literal, case-sensitive string. Its syntax
must be exactly in the format shown in the previous example,
including the placement of commas. to reduce the likelihood to
syntax error, use the command ib_qlgc_vnic_query -es.
Note that the chassis serial number must match the chassis Ox
(Hex) value. The slot serial number is specific to the line card as
well.
5-20IB0056101-00 G
A
5 – Software Installation
OpenFabrics Drivers and Services Configuration and Startup
Each CREATE block must specify a unique NAME. The NAME
represents the Ethernet interface name that will be registered with the
Linux operating system.
c.Format 3: Starting VNIC using DGID. Following is an example of a
DGID and IOCGUID VNIC configuration. This configuration allows for
the quickest start up of VNIC service:
4.Create VirtualNIC interface configuration files. For each Ethernet interface
defined in the /etc/sysconfig/ics_inic.cfg file, create an interface
configuration file, /etc/sysconfig/network-scripts/ifcfg-<NAME>
(or /etc/sysconfig/network/ifcfg-<NAME> on Linux 2.6 kernels),
where <NAME> is the value of the NAME field specified in the CREATE
block.
Following is an example of an ifcfg-eiocx setup for Red Hat systems:
5 – Software Installation
OpenFabrics Drivers and Services Configuration and Startup
5.Start the QLogic VNIC driver and the QLogic VNIC interfaces. Once you
have created a configuration file, you can start the VNIC driver and create
the VNIC interfaces specified in the configuration file by running the
following command:
# /etc/init.d/qlgc_vnic start
You can stop the VNIC driver and bring down the VNIC interfaces by running
the following command:
# /etc/init.d/qlgc_vnic stop
To restart the QLogic VNIC driver, run the following command:
# /etc/init.d/qlgc_vnic restart
If you have not started the InfiniBand network stack (infinipath or
OFED), then running the
starts the InfiniBand network stack, since the QLogic VNIC service requires
the InfiniBand stack.
If you start the InfiniBand network stack separately, then the correct starting
order is:
/etc/init.d/qlgc_vnic start command also
S
Start the InfiniBand stack.
Start QLogic VNIC service.
For example, if you use infinipath, the correct starting order is:
If you try to stop the InfiniBand stack when the QLogic VNIC service is
running, an error message displays, indicating that some of the modules of
the InfiniBand stack are in use by the QLogic VNIC service. Also, any
QLogic VNIC interfaces that you created are removed (because stopping
the InfiniBand network stack unloads the HCA driver, which is required for
the VNIC interfaces to be present).
In this case, do the following:
Stop the QLogic VNIC service with
Stop the InfiniBand stack again.
/etc/init.d/qlgc_vnic stop
.
5-22IB0056101-00 G
A
5 – Software Installation
OpenFabrics Drivers and Services Configuration and Startup
If you want to restart the QLogic VNIC interfaces, run the following
command:
# /etc/init.d/qlgc_vnic restart
You can get information about the QLogic VNIC interfaces by using the following
script:
# ib_qlgc_vnic_info
This information is collected from the
/sys/class/infiniband_qlgc_vnic/interfaces/ directory, under which
there is a separate directory corresponding to each VNIC interface.
VNIC interfaces can be deleted by writing the name of the interface to the
/sys/class/infiniband_qlgc_vnic/interfaces/delete_vnic file. For
example, to delete interface veth0, run the following command:
More information for configuration, starting and stopping the interface, and basic
troubleshooting is available in the QLogic OFED+ User Guide.
OpenSM
OpenSM is an optional component of the OpenFabrics project that provides a
subnet manager for InfiniBand networks. You do not need to use OpenSM if any
of your InfiniBand switches provide a subnet manager, or if you are running a
host-based SM.
After installing the opensm package, OpenSM is configured to be on after the next
machine reboot. Install OpenSM on the machine that will act as a subnet manager
in your cluster, or if it has been installed on more than one machine, use the
chkconfig command (as root) to disable it on all the nodes except for the one on
which it will be used. Use this method:
# chkconfig opensmd off
The command to enable it on reboot is:
# chkconfig opensmd on
You can start opensmd without rebooting your machine by typing:
# /etc/init.d/opensmd start
You can stop opensmd again like this:
# /etc/init.d/opensmd stop
If you want to pass any arguments to the OpenSM program, modify the following
file, and add the arguments to the
OPTIONS variable:
/etc/init.d/opensmd
IB0056101-00 G5-23
5 – Software Installation
OpenFabrics Drivers and Services Configuration and Startup
For example:
# Use the UPDN algorithm instead of the Min Hop algorithm.
OPTIONS="-R updn"
SRP
SRP stands for SCSI RDMA Protocol. It was originally intended to allow the SCSI
protocol to run over InfiniBand for Storage Area Network (SAN) usage. SRP
interfaces directly to the Linux file system through the SRP Upper Layer Protocol.
SRP storage can be treated as another device.
In this release, two versions of SRP are available: QLogic SRP and OFED SRP.
QLogic SRP is available as a separate download, and has its own installation and
configuration instructions.
SRP has been tested on targets from Engenio™ (now LSI Logic
Networks™.
NOTE:
Before using SRP, the SRP targets must already be set up by your system
administrator.
S
®
) and DataDirect
Using OFED SRP
To use OFED SRP, follow these steps:
1.Add ib_srp to the module list in /etc/sysconfig/infinipath to have
2.Discover the SRP devices on your fabric by running this command (as root):
it automatically loaded.
# ibsrpdm
In the output, look for lines similar to these:
GUID: 0002c90200402c04
ID: LSI Storage Systems SRP Driver 200400a0b8114527
5.You can look for the newly created devices in the /proc/partitions file.
The file will look similar to this example (the partition names may vary):
# cat /proc/partitions
major minor #blocks name
8 64 142325760 sde
8 65 142319834 sde1
8 80 71162880 sdf
8 81 71159917 sdf1
8 96 20480 sdg
8 97 20479 sdg1
6.Create a mount point (as root) where you will mount the SRP device. For
example:
# mkdir /mnt/targetname
# mount /dev/sde1 /mnt/targetname
IB0056101-00 G5-25
5 – Software Installation
Other Configuration: Changing the MTU Size
NOTE:
Use sde1 rather than sde. See the mount(8) man page for more
information on creating mount points.
MPI over uDAPL
Some MPI implementations, such as Intel MPI and HP-MPI, can be run over
uDAPL. uDAPL is the user mode version of the Direct Access Provider Library
(DAPL). Examples of these types of MPI implementations are Intel MPI and one
option on Open MPI.
S
If you are running this type of MPI implementation, the
modules will need to be loaded. To load these modules, use these commands (as
root):
# modprobe rdma_cm
# modprobe rdma_ucm
To ensure that the modules are loaded whenever the driver is loaded, add
rdma_cm and rdma_ucm to the OPENFABRICS_MODULES assignment in
/etc/sysconfig/infinipath.
rdma_cm and rdma_ucm
Other Configuration: Changing the MTU Size
The Maximum Transfer Unit (MTU) is set to 4K and enabled in the driver by
default. To change the driver default back to 2K MTU, add this line (as root) in
/etc/modprobe.conf (or in /etc/modprobe.conf.local on SLES):
options ib_ipath mtu4096=0
Restart the driver as described in “Starting and Stopping the InfiniPath Software”
on page 5-26.
NOTE:
The switch must also have the default set to 4K.
Starting and Stopping the InfiniPath Software
The InfiniPath driver software runs as a system service, usually started at system
startup. Normally, you will not need to restart the software, but you may want to
after installing a new InfiniPath release, after changing driver options, or when you
are doing manual testing.
Use the following commands to check or configure the state. These methods will
not reboot the system.
5-26IB0056101-00 G
A
5 – Software Installation
Starting and Stopping the InfiniPath Software
To check the configuration state, use the command:
$ chkconfig --list infinipath
To enable the driver, use the command (as root):
# chkconfig infinipath on 2345
To disable the driver on the next system boot, use the command (as root):
# chkconfig infinipath off
NOTE:
This command does not stop and unload the driver if the driver is already
loaded.
You can start, stop, or restart (as root) the InfiniPath support with:
# /etc/init.d/infinipath [start | stop | restart]
This method will not reboot the system. The following set of commands shows
how to use this script. Note the following:
If OpenSM is configured and running, it must be stopped before the
infinipath stop command, and must be started after the infinipath
start
command. Omit the commands to start/stop opensmd if you are not
running it on that node.
Omit the ifdown and ifup step if you are not using ipath_ether on that
node.
The sequence of commands to restart infinipath are as follows. Note that this
example assumes that ipath_ether is configured as eth2.
Stopping or restarting InfiniPath terminates any QLogic MPI processes, as
well as any OpenFabrics processes that are running at the time. Processes
using networking over
You can check to see if opensmd is running by using the following command; if
there is no output, opensmd is not configured to run:
# /sbin/chkconfig --list opensmd | grep -w on
ipath_ether will return errors.
S
You can check to see if ipath_ether is running by using the following command.
If it prints no output, it is not running.
$ /sbin/lsmod | grep ipath_ether
If there is output, look at the output from this command to determine if it is
configured:
$ /sbin/ifconfig -a
When you need to determine which InfiniPath and OpenFabrics modules are
running, use the following command:
$ lsmod | egrep ’ipath_|ib_|rdma_|findex’
Rebuilding or Reinstalling Drivers After a Kernel
Upgrade
If you upgrade the kernel, then you must reboot and then rebuild or reinstall the
InfiniPath kernel modules (drivers).
To rebuild the drivers, do the following (as root):
# cd /usr/src/infinipath/drivers
# ./make-install.sh
# /etc/init.d/infinipath restart
To reinstall the InfiniPath kernel modules, do the following (as root):
Rebuilding or Reinstalling Drivers if a Different Kernel is Installed
Rebuilding or Reinstalling Drivers if a Different
Kernel is Installed
Installation of the InfiniPath driver RPM (infinipath-kernel-2.2-xxx-yyy)
builds kernel modules for the currently running kernel version. These InfiniPath
modules will work only with that kernel. If a different kernel is booted, then you
must reboot and then re-install or rebuild the InfiniPath driver RPM.
Here is an example. These commands must be done as root:
Reinstallation instructions for the InfiniPath kernel modules are provided in
“Rebuilding or Reinstalling Drivers After a Kernel Upgrade” on page 5-28.
Further Information on Configuring and Loading
Drivers
See the modprobe(8), modprobe.conf(5), and lsmod(8) man pages for more
information. Also see the file:
/usr/share/doc/initscripts-*/sysconfig.txt
LED Link and Data Indicators
The LEDs function as link and data indicators once the InfiniPath software has
been installed, the driver has been loaded, and the fabric is being actively
managed by a subnet manager.
Table 5-5 describes the LED states. The green LED indicates the physical link
signal; the amber LED indicates the link. The green LED will normally illuminate
first. The normal state is Green On, Amber On for adapters other than the
QLE7240 and QLE7280, which have an additional state, as shown in Tabl e 5- 5.
Table 5-5. LED Link and Data Indicators
LED StatesIndication
Green OFF
Amber OFF
The switch is not powered up. The software is neither
installed nor started. Loss of signal.
Verify that the software is installed and configured with
ipath_control -i. If correct, check both cable connectors.
IB0056101-00 G5-29
5 – Software Installation
Adapter Settings
S
Table 5-5. LED Link and Data Indicators (Continued)
LED StatesIndication
Green ON
Amber OFF
Green ON
Amber ON
Green BLINKING (quickly)
Amber ON
Green BLINKING
Amber BLINKING
Table Notes
a
This feature is available only on the QLE7240 and QLE7280 adapters.
a
Signal detected and the physical link is up. Ready to talk
to SM to bring the link fully up.
If this state persists, the SM may be missing or the link
may not be configured.
Use ipath_control -i to verify the software state. If
all HCAs are in this state, then the SM is not running.
Check the SM configuration, or install and run opensmd.
The link is configured, properly connected, and ready.
Signal detected. Ready to talk to an SM to bring the link
fully up.
The link is configured. Properly connected and ready to
receive data and link packets.
Indicates traffic
Locates the adapter
This feature is controlled by ipath_control -b [On
| Off]
Adapter Settings
The following adapter settings can be adjusted for better performance.
Use taskset to tune CPU affinity on Opteron systems with the
QLE7240, QLE7280, and QLE7140. Latency will be slightly lower for the
Opteron socket that is closest to the PCI Express bridge. On some chipsets,
bandwidth may be higher on this socket. See the QLogic HCA and InfiniPath
Software User Guide for more information on using taskset. Also see the
taskset(1) man page.
Use an IB MTU of 4096 bytes instead of 2048 bytes, if available, with
the QLE7240, QLE7280, and QLE7140. 4K MTU is enabled in the
InfiniPath driver by default. A switch that supports and is configured for 4KB
MTU is also required to use this setting. To change this setting, see “Other
Configuration: Changing the MTU Size” on page 5-26.
5-30IB0056101-00 G
A
Use a PCIe MaxReadRequest size of at least 512 bytes with the
QLE7240 and QLE7280. QLE7240 and QLE7280 adapters can support
sizes from 128 bytes to 4096 byte in powers of two. This value is typically
set by the BIOS.
Use the largest available PCIe MaxPayload size with the QLE7240 and
QLE7280. The QLE7240 and QLE7280 adapters can support 128, 256, or
512 bytes. This value is typically set by the BIOS as the minimum value
supported both by the PCIe card and the PCIe root complex.
Adjust the MTRR setting. Check that the BIOS setting for MTRR is
Discrete. For more information, see “MTRR Mapping and Write
Combining” on page A-2.
Customer Acceptance Utility
ipath_checkout is a bash script that verifies that the installation is correct and
that all the nodes of the network are functioning and mutually connected by the
InfiniPath fabric. It must be run on a front end node, and requires specification of a
nodefile. For example:
5 – Software Installation
Customer Acceptance Utility
$ ipath_checkout [options] nodefile
The nodefile lists the hostnames of the nodes of the cluster, one hostname per
line. The format of
hostname1
hostname2
...
NOTE:
The hostnames in the nodefile are Ethernet hostnames, not IPv4 addresses.
ipath_checkout
1.Executes the
from the front end.
2.Executes the
ssh.
3.Gathers and analyzes system configuration from the nodes.
4.Gathers and analyzes RPMs installed on the nodes.
5.Verifies QLogic hardware and software status and configuration. Includes
tests for link speed, PIO bandwidth (incorrect MTRR settings), and MTU
size.
nodefile is as follows:
performs the following seven tests on the cluster:
ping command to all nodes to verify that they all are reachable
ssh command to each node to verify correct configuration of
6.Verifies the ability to mpirun jobs on the nodes.
7.Runs a bandwidth and latency test on every pair of nodes and analyzes the
results.
S
The options available with
ipath_checkout are shown in Table 5-6.
Table 5-6.
ipath_checkout Options
Command Meaning
-h, --helpThese options display help messages describing how a command is used.
-v, --verbose
-vv, --vverbose
-vvv, --vvver-
bose
-c, --continueWhen this option is not specified, the test terminates when
-k, --keepThis option keeps intermediate files that were created while
--workdir=DIRUse DIR to hold intermediate files created while running
These options specify three successively higher levels of
detail in reporting test results. There are four levels of detail
in all, including the case where none of these options are
given.
any test fails. When specified, the tests continue after a failure, with failing nodes excluded from subsequent tests.
performing tests and compiling reports. Results will be saved
in a directory created by mktemp and named
infinipath_XXXXXX or in the directory name given to
--workdir.
tests. DIR must not already exist.
--run=LISTThis option runs only the tests in LIST. See the seven tests
listed previously. For example, --run=123 will run only
tests 1, 2, and 3.
--skip=LISTThis option skips the tests in LIST. See the seven tests listed
previously. For example, --skip=2457 will skip tests 2, 4, 5,
and 7.
-d, --debugThis option turns on the -x and -v flags in bash(1).
In most cases of failure, the script suggests recommended actions. Please see
ipath_checkout man page for more information and updates.
the
Also refer to the Troubleshooting appendix in the QLogic HCA and InfiniPath Software User Guide.
5-32IB0056101-00 G
A
Removing Software Packages
Removing Software Packages
This section provides instructions for uninstalling or downgrading the InfiniPath
and OpenFabrics software.
Uninstalling InfiniPath and OpenFabrics RPMs
To uninstall the InfiniPath software packages on any node, type the following
command (as root) using a
This InfiniPath release supports Lustre. Lustre is a fast, scalable Linux cluster file
system that interoperates with InfiniBand. To use Lustre, you need:
A Linux kernel that is one of the supported kernels for this release, patched
with Lustre-specific patches
Lustre modules compiled for the above kernel
Lustre utilities required for configuration
The InfiniPath Release Notes provide information about the Lustre patches. For
general instructions on downloading, installing, and using Lustre, go to:
http://www.lustre.org
.
Installing QLogic MPI in an Alternate Location
QLogic MPI can be installed in an alternate installation location by using the
--prefix option with rpm. This option is useful for installations that support
multiple concurrent MPIs, for other site-specific requirements, or for smooth
integration of QLogic MPI with mpi-selector.
S
The mpi-selector utility and QLogic MPI mpi-selector registration are
provided by RPMs in the OtherMPIs directory.
When this option is used, the argument passed to --prefix replaces the default
/usr prefix: QLogic MPI binaries, documentation, and libraries are installed under
that prefix. However, a few configuration files are installed in /etc regardless of
the desired --prefix.
Additionally, installations that are maintained in alternate locations must ensure
that the environment variable $MPICH_ROOT is always set to the same prefix that
was used to install the RPMs with the --prefix option. When set, the $MPICH_ROOT variable allows QLogic MPI to correctly locate header and library
files for MPI compilation and running parallel jobs.
NOTE:
In InfiniPath 2.2, $MPICH_ROOT replaces the environment variable
$INFINIPATH_ROOT, which is now deprecated. In InfiniPath 2.1,
$INFINIPATH_ROOT assumed an RPM install --prefix of
$INFINIPATH_ROOT/usr.
5-34IB0056101-00 G
A
5 – Software Installation
Additional Installation Instructions
For example, install all RPMs that relate to QLogic MPI in /usr/mpi/qlogic.
Leave all remaining InfiniPath libraries and tools in their default installation
location (/usr). This approach leaves InfiniPath libraries (such as
libpsm_infinipath.so and libinfinipath.so) in standard system
directories so that other MPIs can easily find them in their expected location. Also,
this scenario leaves InfiniPath-specific utilities such as ipath_checkout and ipath_control in standard system search paths.
For this example, unpack the InfiniPath tarball as shown in “Unpacking the
InfiniPath tar File” on page 5-4. Then move all RPMs that will be prefixed to a new
directory called InfiniPath-MPI-prefixed. This includes all QLogic MPI
development headers, libraries, runtime and documentation RPMs, as well as
The desired prefix should be made available in the $MPICH_ROOT environment
variable, either by global shell configuration files or through third-party
environment management utilities such as mpi-selector or the Environment
Modules.
See the Using QLogic MPI section in the QLogic HCA and InfiniPath Software
User Guide for more information on setting $MPICH_ROOT and using the
mpi-selector utility.
Installing on an Unsupported Distribution
If you are running a kernel that does not match a supported kernel/distribution
pair, you may need to provide an override during the install of the
infinipath-kernel RPM. This override may be needed if you have a completely
unsupported distribution, or if you have upgraded the kernel without upgrading the
rest of the distribution. The InfiniPath install determines the distribution from either
the
/etc/redhat-release file or the /etc/SuSE-release file.
Using the override may not result in a buildable or working driver if your
distribution/kernel combination is not similar enough to a tested and
supported distribution/kernel pair.
The following example installation is for a Red Hat Enterprise Linux 4 Update 4
compatible kernel, where the
distribution. If you are a
# export IPATH_DISTRO=2.6.9_U4
Follow this with your normal rpm install commands, or run (as root):
# /usr/src/infinipath/drivers/make-install.sh
You can examine the list of supported distributions for this InfiniPath release by
looking at this script:
/usr/src/infinipath/drivers/build-guards.sh
S
/etc/redhat-release file indicates another
bash or sh user, type:
This is the current list of strings that can be passed to IPATH_DISTRO. The
distributions to which each string is applied are shown in brackets:
2.6.22_FC6 [Fedora Core 6, 2.6.22 kernel]
2.6.18_EL5 [RHEL5, Scientific Linux 5.0, CentOS 5.0]
2.6.18_EL5.1 [RHEL5.1, Scientific Linux 5.1, CentOS 5.1]
2.6.16_sles10 [SLES10 GM]
2.6.16_sles10_sp1 [SLES10 SP1]
2.6.9_U4 [RHEL4 U4, CentOS 4.4, Scientific Linux 4.4]
2.6.9_U5 [RHEL4 U5, CentOS 4.5, Scientific Linux 4.5]
2.6.9_U6 [RHEL4 U6, CentOS 4.6, Scientific Linux 4.6]
If you try to install on an unsupported distribution or an unsupported
distribution/kernel pair, you will see an error message. This example shows a
case where 4.3 is still in
Linux 4.4:
# rpm -Uv infinipath-kernel-2.2-3187.376_rhel4_psc.x86_64.rpm
Preparing packages for installation...
infinipath-kernel-2.2-3187.376_rhel4_psc
*** 2.6.9-42.0.10.ELsmp Scientific Linux SL release 4.3
(Beryllium) is not a supported InfiniPath distribution
error: %post(infinipath-kernel-2.2-3187.376_rhel4_psc.x86_64)
scriptlet failed, exit status 1
/etc/redhat-release, although the kernel is Scientific
Managing and Installing Software Using Rocks
Rocks is a distribution designed for managing clusters from the San Diego
Supercomputer center (SDSC).
5-36IB0056101-00 G
A
5 – Software Installation
Additional Installation Instructions
Rocks is a way to manage the kickstart automated installation method created by
Red Hat. By using the Rocks conventions, the installation process can be
automated for clusters of any size. A Roll is an extension to the Rocks base
distribution that supports different cluster types or provides extra functionality.
QLogic extends the normal Rocks compute node appliance xml file by adding two
functions: one function installs the QLogic InfiniPath software, and the other
function loads the drivers after kickstart reboots the machine.
This section provides an overview of one way of building a Rocks cluster using the
recommend rolls, and a sample xml file that describes the contents of a kickstart
config file. By reading this section and following the instructions on the Rocks web
site, you can then install the InfiniPath RPMs on the required cluster nodes.
NOTE:
There are many ways to use Rocks to manage clusters. Familiarize yourself
first with kickstart and then Rocks before using this method to install the
InfiniPath RPMs.
Installing Rocks and InfiniPath RPMs
The following instructions are for building a Rocks 4.2.1 cluster, and then for
installing InfiniPath. These instructions are only guidelines; see the material on the
Rocks web site to complete an installation. If you want to use later versions of
Rocks, these instructions will serve as general procedural steps.
1.Download the required rolls from the Rocks web site:
http://www.rocksclusters.org/
Follow the Downloads link to the following CDs:
Core Roll: Rocks 4.2.1 X86_64 ISO
(Area51+Base+Ganglia+grid+hpc+java+sge+web-server)
OS Roll - Disk 1
OS Roll - Disk 2
You may also need updates; look for the latest files with the service-pack
prefix.
2.Build the front end node with the above CDs. For more details, see the
Rocks installation documentation on the Rocks web site.
3.After building the front end node, make sure that the Rocks tools work. Add
users and reload nodes. For more details, see the Rocks installation
documentation on the Rocks web site.
4.In the directory
file:
extend-compute.xml
IB0056101-00 G5-37
/home/install/site-profiles/4.2.1/nodes, create the
A skeleton XML node file. This file is only a template
and is intended as an example of how to customize your
Rocks cluster and use InfiniPath InfiniBand software and
MPI.
We want to extend....
</description>
<changelog>
</changelog>
S
<main>
<!-- kickstart ’main’ commands go here,-->
<!--e.g., partitioning info -->
</main>
<!-!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! -->
<!-- Many older OFED packages may be included -->
<!-- with CentOS/RHEL. We don’t want to mix these -->
<!-- packages with QLogic packages -->
<!-- About 26 of them could cause collisions. -->
<!-- The - in front of the package name causes kickstart -->
<!-- to ignore them, so we negate the following and -->
<!-- install the QLogic package set in the -->
<!-- post-97-infinipath script. -->
<!-- This list is from CentOS; RedHat packages may be -->
<!-- slightly different.-->
<package>-dapl-devel</package>
<package>-dapl</package>
<package>-kernel-ib</package>
<package>-libibcm</package>
<package>-libibcommon-devel</package>
<package>-libibcommon</package>
<package>-libibmad-devel</package>
<package>-libibmad</package>
<package>-libibumad-devel</package>
<package>-libibumad</package>
<package>-libibverbs-devel</package>
<package>-libibverbs</package>
<package>-libibverbs-utils</package>
echo ’# INFINIPATH_ULIMIT_INSTALL_COMMENT; used by InfiniPath
install, do not remove’ >> /etc/initscript
echo ’# This allows locking of up to 128MB of memory per
process. It only’ >> /etc/initscript
echo ’# takes effect after a reboot, or "init 1", followed by
"init 3"’ >> /etc/initscript
echo "ulimit -l 131072" >> /etc/initscript
echo ’eval exec "$4"’ >> /etc/initscript
# the kernel must match the smp kernel you give below.
uname -r
export KVER="2.6.9-42.0.2.ELsmp"
/usr/src/infinipath/drivers/make-install.sh >> $IPLOG
chkconfig --level 123456 cpuspeed off
<!--
Inserting a post installation script here. This
code will be executed on the destination node AFTER the
packages have been installed.
# Install any other OpenFabrics (OFED) packages that you want,
# then restart
/etc/init.d/infinipath restart
#Either move this file or remove it
mv /etc/rc.d/rocksconfig.d/post-97-infinipath /tmp
#rm /etc/rc.d/rocksconfig.d/post-97-infinipath
</file>
S
</post>
</kickstart>
In this file, the installation of the InfiniPath drivers is done in the <post>
section, as it is a live install. This file can be used as a guideline: it may be
cut and pasted, then modified to suit your needs.
5.On the front end node, download the InfiniPath 2.2 tarball for your
distribution,
After downloading the tarball, type the following, where
RHEL4 or SLES10.
xxx is the
distribution identifier, and yyy is the platform architecture:
$ tar zxvf InfiniPath2.2-xxx-yyy.tgz
The tar command will create a directory based on the tar file name and
place the RPMs there. Next, copy the RPMs you want to install to the
directory:
/home/install/contrib/4.2.1/x86_64/RPMS
It is easiest to copy all of the RPMs to this directory, then use the commands
in the extend-compute.xml file to install the desired packages.
5-40IB0056101-00 G
A
NOTE:
If you intend to use OpenFabrics and are using RHEL4 or RHEL5, make
sure you install
rhel4-ofed-fixup-2.2-4081.772.rhel4_psc.noarch.rpm, which is in
OpenFabrics directory. This RPM will fix two conflicts. See
the
“OpenFabrics Library Dependencies” on page A-4 for more information.
6.The completion of the installation is done using the extend-compute.xml
file. More specific instructions for completing the install process can be
found under the Documentation link on the Rocks web site:
http://www.rocksclusters.org/
Further Information on Rocks and kickstart
QLogic recommends checking the Rocks web site for updates. Extensive
documentation on installing Rocks and custom rolls is on the Rocks web site. For
more information on Rocks, see:
http://www.rocksclusters.org/
5 – Software Installation
Additional Installation Instructions
To find more information on Red Hat Enterprise Linux 4, and on using kickstart,
see:
The following sections contain information about issues that may occur during
installation. Some of this material is repeated in the Troubleshooting appendix of
the QLogic HCA and InfiniPath Software User Guide.
Many programs and files are available that gather information about the cluster,
and can be helpful for debugging. See appendix D, Useful Programs and Files, in
the QLogic HCA and InfiniPath Software User Guide.
Hardware Issues
Some of the hardware issues that may occur during installation are described in
the following sections. Use the LEDs, as described in “LED Link and Data
Indicators” on page 5-29, to help diagnose problems.
Node Spontaneously Reboots
If a node repeatedly and spontaneously reboots when attempting to load the
InfiniPath driver, it may be because the QLogic adapter is not installed correctly in
the HTX or PCI Express slot.
Some HTX Motherboards May Need Two or More CPUs in Use
Some HTX motherboards may require that two or more CPUs be in use for the
QLogic adapter to be recognized. This is most evident in four-socket
motherboards.
BIOS Settings
This section covers issues related to BIOS settings.The two most important
settings are:
Advanced Configuration and Power Interface (ACPI). This setting must be
enabled. If ACPI is disabled, it may cause initialization problems, as
described in the Troubleshooting section of the QLogic HCA and InfiniPath Software User Guide.
Memory Type Range Registers (MTRR) mapping must be set to Discrete.
Using a different setting may result in reduced performance. See “MTRR
Mapping and Write Combining” on page A-2 for details.
IB0056101-00 GA-1
A – Installation Troubleshooting
BIOS Settings
MTRR Mapping and Write Combining
MTRR is used by the InfiniPath driver to enable write combining to the QLogic
on-chip transmit buffers. Write combining improves write bandwidth to the QLogic
chip by writing multiple words in a single bus transaction (typically 64 bytes). Write
combining applies only to x86_64 systems. To see if write combining is working
correctly and to check the bandwidth, run the following command:
$ ipath_pkt_test -B
With write combining enabled, the QLE7140 and QLE7240 report in the range
of 1150–1500 MBps; the QLE7280 reports in the range of 1950–2960 MBps. The
QHT7040/7140 adapters normally report in the range of 2300–2650 MBps.
S
You can also use
In some cases, the InfiniPath driver cannot configure the CPU write combining
attributes for QLogic InfiniPath. This case is normally seen with a new system, or
after the system’s BIOS has been upgraded or reconfigured.
If this error occurs, the interconnect operates, but in a degraded performance
mode. The latency typically increases to several microseconds, and the
bandwidth may decrease to as little as 200 MBps.
Upon driver startup, you may see these errors:
ib_ipath 0000:04:01.0: infinipath0: Performance problem: bandwidth
to PIO buffers is only 273 MiB/sec
infinipath: mtrr_add(feb00000,0x100000,WC,0) failed (-22)
infinipath: probe of 0000:04:01.0 failed with error -22
If you do not see any of these messages on your console, but suspect this
problem, check the /var/log/messages file. Some systems suppress driver
load messages but still output them to the log file.
Two suggestions for fixing this problem are described in “Edit BIOS Settings to Fix
MTRR Issues” on page A-2 and “Use the ipath_mtrr Script to Fix MTRR Issues”
on page A-3.
See the Troubleshooting section of the QLogic HCA and InfiniPath Software User Guide for more details on a related performance issue.
ipath_checkout (use option 5) to check bandwidth.
Edit BIOS Settings to Fix MTRR Issues
You can edit the BIOS setting for MTRR Mapping. The BIOS setting looks similar
to:
MTRR Mapping [Discrete]
For systems with very large amounts of memory (32GB or more), it may also be
necessary to adjust the BIOS setting for the PCI hole granularity to 2GB. This
setting allows the memory to be mapped with fewer MTRRs, so that there will be
one or more unused MTRRs for the InfiniPath driver.
A-2IB0056101-00 G
A
Some BIOS’ do not have the MTRR mapping option. It may have a different
name, depending on the chipset, vendor, BIOS, or other factors. For example, it is
sometimes referred to as 32 bit memory hole. This setting must be enabled.
If there is no setting for MTRR mapping or 32 bit memory hole, and you have
problems with degraded performance, contact your system or motherboard
vendor and ask how to enable write combining.
Use the ipath_mtrr Script to Fix MTRR Issues
QLogic also provides a script, ipath_mtrr, which sets the MTRR registers,
enabling maximum performance from the InfiniPath driver. This Python script is
available as a part of the InfiniPath software download, and is contained in the
infinipath* RPM. It is installed in /bin.
To diagnose the machine, run it with no arguments (as root):
# ipath_mtrr
The test results will list any problems, if they exist, and provide suggestions on
what to do.
A – Installation Troubleshooting
Software Installation Issues
To fix the MTRR registers, use:
# ipath_mtrr -w
Restart the driver after fixing the registers.
This script needs to be run after each system reboot.
See the ipath_mtrr(8) man page for more information on other options.
Issue with Supermicro® H8DCE-HTe and QHT7040
The QLogic adapter may not be recognized at startup when using the Supermicro
H8DCE-HT-e and the QHT7040 adapter. To fix this problem, set the operating
system selector option in the BIOS for Linux. The option will look like:
OS Installation [Linux]
Software Installation Issues
Some problems can be found by running ipath_checkout. Run
ipath_checkout before contacting technical support.
IB0056101-00 GA-3
A – Installation Troubleshooting
Version Number Conflict with opensm-* on RHEL5 Systems
Version Number Conflict with opensm-* on
RHEL5 Systems
The older opensm-* packages that come with the RHEL5 distribution have a
version number (3) that is greater the InfiniPath version number (2.2). This
prevents the newer InfiniPath packages from being installed. You may see an
error message similar to this when trying to install:
Preparing packages for installation...
package opensm-3.1.8-0.1.ofed20080130 (which is newer than
opensm-2.2-33596.832.3_1_10.rhel4_qlc) is already installed
package opensm-devel-3.1.8-0.1.ofed20080130 (which is newer than
opensm-devel-2.2-33596.832.3_1_10.rhel4_qlc) is already installed
These older packages should be manually uninstalled before installing the
InfiniPath and OpenFabrics 2.2 packages. Run the following command (as root):
If a package was not installed, you may see a warning message similar to:
S
rpm: no packages given for erase
OpenFabrics Dependencies
Install sysfsutils for your distribution before installing the OpenFabrics RPMs,
as there are dependencies. If
error messages like this:
error: Failed dependencies:
libsysfs.so.1()(64bit) is needed by
libipathverbs-2.2-28581.811.1_1.rhel4_psc.x86_64
libsysfs.so.1()(64bit) is needed by
libibverbs-utils-2.2-28581.811.1_1.rhel4_psc.x86_64
/usr/include/sysfs/libsysfs.h is needed by
libibverbs-devel-2.2-28581.811.1_1.rhel4_psc.x86_64
Check your distribution’s documentation for information about sysfsutils.
sysfsutils has not been installed, you will see
OpenFabrics Library Dependencies
There are two issues with OpenFabrics on RHEL4 and RHEL5 distributions.
Some of the OpenFabrics RPMs (most notably
dependencies that are distribution-specific. Additionally, the OpenFabrics
packages contained in this release are newer than the ones that are distributed
with RHEL4 U5 and RHEL5, and can cause conflicts during installation. An RPM,
rhel4-ofed-fixup-2.2-xxx.rhel4_psc.noarch.rpm (where xxx is the build
identifier), has been added to the RHEL tarball to work around the conflicts
problem. It is contained in the OpenFabrics directory, and is installed when the
other OpenFabrics RPMs are installed. Installation of this RPM will not affect FC6
users.
ibutils) have library
A-4IB0056101-00 G
A
Missing Kernel RPM Errors
Install the kernel-source, kernel-devel, and, if using an older release,
kernel-smp-devel RPMs for your distribution before installing the InfiniPath
RPMs, as there are dependencies. Use uname -a to find out which kernel is
currently running, to make sure that you install the version with which it matches.
If these RPMs have not been installed, you will see error messages like this when
installing InfiniPath:
Building and installing InfiniPath modules for 2.6.16_sles10
2.6.16.21-0.8-debug kernel
*** ERROR: /lib/modules/2.6.16.21-0.8-debug/build/.config is
missing.
*** Is the kernel-source rpm for 2.6.16.21-0.8-debug
installed?
================
================
Building and installing InfiniPath modules for 2.6.9_U4
2.6.9-42.ELsmp kernel
*** ERROR: /lib/modules/2.6.9-42.ELsmp/build/.config is missing.
*** Is the kernel-smp-devel rpm for 2.6.9-42.ELsmp
installed?
================
.
.
.
A – Installation Troubleshooting
Version Number Conflict with opensm-* on RHEL5 Systems
Install the correct RPMs by using the yum or yast commands, for example:
# yum install kernel-devel
NOTE:
Check your distribution’s documentation for more information on installing
these RPMs, and for usage of yum or yast.
Next, the infinipath-kernel package must be re-installed, with the
--replacepkgs option included. Then InfiniPath can be restarted. To do so,
type the following (as root):
# rpm -U --replacepkgs infinipath-kernel-*
# /etc/init.d/infinipath restart
IB0056101-00 GA-5
A – Installation Troubleshooting
Version Number Conflict with opensm-* on RHEL5 Systems
Resolving Conflicts
Occasionally, conflicts may arise when trying to install "on top of" an existing set of
files that may come from a different set of RPMs. For example, if you install the
QLogic MPI RPMs after having previously installed Local Area Multicomputer
(LAM)/MPI, there will be conflicts, since both installations have versions of some
of the same programs and documentation. You would see an error message
similar to the following:
This has no effect on ipath_ether, so this error message can be safely ignored.
Lockable Memory Error on Initial Installation of InfiniPath
During the first installation of InfiniPath software, /etc/initscript is created or
modified to increase the amount of lockable memory (up to 128 MB) for normal
users. This change will not take effect until the system is rebooted, and jobs may
fail with error messages about locking memory or failing mmap. This error is
described in the QLogic MPI Troubleshooting section Lock Enough Memory on
Nodes When Using a Batch Queuing System
Software User Guide.
This is not an issue when upgrading to a newer version of InfiniPath software.
in the QLogic HCA and InfiniPath
IB0056101-00 GA-7
A – Installation Troubleshooting
Version Number Conflict with opensm-* on RHEL5 Systems
Notes
S
A-8IB0056101-00 G
BConfiguration Files
Table B-1 contains descriptions of the configuration and configuration template
files used by the InfiniPath and OpenFabrics software.
/etc/modprobe.confSpecifies options for modules when added
or removed by the modprobe command.
Also used for creating aliases.
For Red Hat systems
/etc/modprobe.conf.localSpecifies options for modules when added
or removed by the modprobe command.
Also used for creating aliases.
For SLES systems
/etc/sysconfig/hardware/hwcfgeth-id-$MAC
/etc/sysconfig/ics_inic.cfgFile where VNIC Ethernet interfaces are
/etc/sysconfig/infinipathThe primary configuration file for Infini-
/etc/sysconfig/network/ifcfg<NAME>
Network module file for ipath_ether
network interface on SLES systems. $MAC
is replaced by the MAC address.
defined
Path, OFED modules, and other modules
and associated daemons. Automatically
loads additional modules or changes IPoIB
transport type.
Network configuration file for network interfaces. When used for ipath_ether,
<NAME> is in the form ethX, where X is
the number of the device, typically, 2, 3,
etc.
If used for VNIC configuration, name is in
the form
of the device.
configuration file for each interface defined
in /etc/sysconfig/ics_inic.cfg.
For SLES systems
eiocX, where X is the number
There will be one interface
IB0056101-00 GB-1
B – Configuration Files
S
Table B-1. Configuration Files (Continued)
Configuration File Name Description
/etc/sysconfig/network-scripts/ifcfg-<NAME>
Network configuration file for network interfaces. When used for ipath_ether,
<NAME> is in the form ethX, where X is
the number of the device, typically, 2, 3,
etc.
When used for VNIC configuration, the
name is in the form
the number of the device.
one interface configuration file for each
interface defined in /etc/syscon-fig/ics_inic.cfg.
File that explains many of the entries in the
configuration files
There will be
B-2IB0056101-00 G
CRPM Descriptions
The following sections contain detailed descriptions of the RPMs for InfiniPath and
OpenFabrics software.
InfiniPath and OpenFabrics RPMs
For ease of installation, QLogic recommends that all RPMs are installed on all
nodes. However, some RPMs are optional. Since cluster nodes can be used for
different functions, it is possible to selectively install RPMs. For example, you can
install the opensm package for use on the node that will act as a subnet manager.
If you want to selectively install the RPMs, see the following tables for a
comparison of required and optional packages.
Different Nodes May Use Different RPMs
In a cluster environment, different nodes may be used for different functions, such
as launching jobs, software development, or running jobs. These nodes are
defined as follows:
Front end node. This node launches jobs. It is referred to as the front end
node throughout this document.
Compute node. These nodes run jobs.
Developmentorbuild node. These are the machines on which examples
or benchmarks can be compiled.
Any machine can serve any combination of these three purposes, but a typical
cluster has many compute nodes and just a few (or only one) front end nodes.
The number of nodes used for development will vary. Although QLogic
recommends installing all RPMs on all nodes, not all InfiniPath software is
required on all nodes. See Tab le C- 1, Table C-2, Table C-3, or Ta bl e C- 4 for
information on installation of software RPMs on specific types of nodes.
RPM Version Numbers and Identifiers
The InfiniPath RPMs that are shipped have the InfiniPath release number and the
build versions contained in the RPM name. The architecture is designated by
x86_64, noarch, or i386, and is dependent upon the distribution. For example:
infinipath-2.2-33597.832_rhel4_qlc.x86_64.rpm
In this RPM, 2.2 is the InfiniPath release, and 33597.832 are the build versions.
IB0056101-00 GC-1
C – RPM Descriptions
RPM Organization
Non-InfiniPath components may also have their own version number:
In all of the tables in this appendix, the build identifier is xxx and the distribution
identifier is yyy. Using this convention, the previous RPMs would be listed as:
The tables in the following sections show the sample contents of each of the
above directories.
To generate a list of the InfiniPath software package contents on each RPM, type:
$ rpm -qlp rpm_file_name
Documentation and InfiniPath RPMs
The documentation/RPMs for InfiniPath are listed in Tab le C -1 .
Table C-1. InfiniPath Documentation/RPMs
RPM NameFront EndComputeDevelopment
infinipath-doc-2.2-xxx_yyy.noarch.rpm
InfiniPath man pages and other documents
mpi-doc-2.2-xxx_yyy.noarch.rpm
Man pages for MPI functions and other
MPI documents
OptionalOptionalOptional
OptionalOptionalOptional
C-2IB0056101-00 G
A
C – RPM Descriptions
Documentation and InfiniPath RPMs
The InfiniPath/RPMs are listed in Tab le C -2 .
Table C-2. InfiniPath/RPMs
RPM NameFront EndComputeDevelopment
infinipath-2.2-xxx_yyy.x86_64.rpm
Utilities and source code
InfiniPath configuration files
Contains ipath_checkout and
ipathbug-helper
infinipath-kernel-2.2-xxx_yyy.x86_64.rpm
InfiniPath drivers, OpenFabrics kernel
modules
infinipath-libs-2.2-xxx_yyy.i386.rpm
InfiniPath protocol shared libraries for
32-bit and 64-bit systems
Table Notes
a
If you want to use ipath_checkout and ipathbug-helper, install this RPM wherever
you install mpi-frontend.
a
OptionalRequiredOptional
OptionalRequiredOptional
OptionalRequiredOptional
The InfiniPath-Devel/RPMs are listed in Tab le C -3 .
Table C-3. InfiniPath-Devel/RPMs
RPM NameFront EndComputeDevelopment
infinipath-devel-2.2-xxx_yyy.x86_64.rpm
Development files for InfiniPath
mpi-devel-2.2-xxx_yyy.noarch.rpm
Source code for the MPI development
environment, including headers and
libs, MPI examples and benchmarks.
Use to build the examples or rebuild the
benchmarks.
IB0056101-00 GC-3
OptionalOptionalOptional
OptionalOptionalRequired
C – RPM Descriptions
OpenFabrics RPMs
The InfiniPath-MPI/RPMs are listed in Table C-4.
S
Table C-4. InfiniPath-MPI/RPMs
RPM NameFront EndComputeDevelopment
mpi-benchmark-2.2-xxx_yyy.x86_64.rpm
MPI benchmark binaries
mpi-frontend-2.2-xxx_yyy.i386.rpm
MPI job launch scripts and binaries,
including mpirun and MPD
mpi-libs-2.2-xxx_yyy.i386.rpm
Shared libraries for MPI
OpenFabrics RPMs
OpenFabrics and OpenSM are optional components. For ease of installation,
QLogic recommends that all of the OpenFabrics RPMs listed in Table C-5 and
Table C-6 be installed on all nodes. The development RPMs in Table C-7 are only
needed on the nodes where OFED programs are compiled. The opensm package
in Ta bl e C -8 should be installed only on the node that will be used as a subnet
manager. The packages in Tab le C- 9 should be installed only if other HCAs are
used. The packages in Table C-11 should be installed if other MPI
implementations are desired.
Table C-5. OpenFabrics Documentation/RPMs
OptionalRequiredOptional
RequiredRequiredOptional
OptionalRequiredRequired
RPM NameComments
ofed-docs-2.2-xxx.1_3.yyy.x86_64.rpm
OpenFabrics documentation
Table Notes
OpenFabrics documentation is installed in the same directory as the InfiniPath documentation.
Optional for
OpenFabrics
Table C-6. OpenFabrics/RPMs
RPM NameComments
dapl-2.2-xxx.1_2_5.yyy.x86_64.rpm
uDAPL 1.2 support
dapl-2.2-xxx.2_0_7.yyy.x86_64.rpm
uDAPL 2.0 support
C-4IB0056101-00 G
Optional for
OpenFabrics
Optional for
OpenFabrics
A
C – RPM Descriptions
OpenFabrics RPMs
Table C-6. OpenFabrics/RPMs (Continued)
RPM NameComments
dapl-utils-2.2-xxx.2_0_7.yyy.x86_64.rpm
uDAPL support
ib-bonding-2.2-xxx.0_9_0.yyy.x86_64.rpm
Utilities to manage and control the driver operation
ibsim-2.2-xxx.0_4.yyy.x86_64.rpm
Voltaire InfiniBand Fabric Simulator
ibutils-2.2-xxx.1_2.yyy.x86_64.rpm
ibutils provides InfiniBand (IB) network and path diagnostics.
ibvexdmtools-2.2-xxx.0_0_1.yyy.x86_64.rpm
Discover and use QLogic Virtual NIC devices via VNIC protocol over
InfiniBand
infiniband-diags-2.2-xxx.1_3_6.yyy.x86_64.rpm
Diagnostic tools
iscsi-initiator-utils-2.2-xxx.2_0.yyy.x86_64.rpm
Server daemon and utility programs for iSCSI. Also iSER support
For Red Hat systems
libibcm-2.2-xxx.1_0_2.yyy.x86_64.rpm
Along with the OpenFabrics kernel drivers, libibcm provides a userspace IB connection management API.
a
Optional for
OpenFabrics
Optional for
OpenFabrics
Optional for
OpenFabrics
Optional for
OpenFabrics
Optional for
OpenFabrics
Optional for
OpenFabrics
Optional for
OpenFabrics
Optional for
OpenFabrics
libibcommon-2.2-xxx.1_0_8.yyy.x86_64.rpm
Common utility functions for IB diagnostic and management tools
libibmad-2.2-xxx.1_1_6.yyy.x86_64.rpm
Low-layer IB functions for use by the IB diagnostic and management
programs. These include management datagrams (MADs), SA, SMP,
and other basic IB functions.
libibumad-2.2-xxx.1_1_7.yyy.x86_64.rpm
Provides the user MAD library functions that sit on top of the user MAD
modules in the kernel. These functions are used by IB diagnostic and
management tools, including OpenSM.
IB0056101-00 GC-5
Required for
OpenSM
Required for
OpenSM
Required for
OpenSM
C – RPM Descriptions
OpenFabrics RPMs
S
Table C-6. OpenFabrics/RPMs (Continued)
RPM NameComments
libibverbs-2.2-xxx.1_1_1.yyy.x86_64.rpm
Library that allows userspace processes to use InfiniBand verbs as
described in the InfiniBand Architecture Specification. This library
includes direct hardware access for fast path operations. For this
library to be useful, a device-specific plug-in module must also be
installed.
libibverbs-utils-2.2-xxx.1_1_1.yyy.x86_64.rpm
Useful libibverbs example programs such as ibv_devinfo, which
displays information about IB devices
libipathverbs-2.2-xxx.1_1.yyy.x86_64.rpm
Provides device-specific userspace driver for QLogic HCAs
librdmacm-2.2-xxx.1_0_6.yyy.x86_64.rpm
Support for the new connection manager
librdmacm-utils-2.2-xxx.1_0_6.yyy.x86_64.rpm
Utilities for the new connection manager
libsdp-2.2-xxx.1_1_99.yyy.x86_64.rpm
Can be LD_PRELOAD-ed to have a sockets application use IB Sockets Direct Protocol (SDP) instead of TCP, transparently and without
recompiling the application
Required for
OpenFabrics
Required for
OpenFabrics
Required for
OpenFabrics
Optional for
OpenFabrics
Optional for
OpenFabrics
Required for
OpenFabrics
ofed-scripts-2.2-xxx.1_3.yyy.x86_64.rpm
OpenFabrics scripts
openib-diags-2.2-xxx.1_2_7.yyy.x86_64.rpm
Useful programs for troubleshooting and checking the state of the
adapter, IB fabric, and its components
open-iscsi-2.2-xxx.2_0.yyy.x86_64.rpm
Transport independent, multi-platform implementation of RFC3720
iSCSI with iSER support
For SLES systems
opensm-libs-2.2-xxx.3_1_10.yyy.x86_64.rpm
Provides the library for OpenSM
perftest-2.2-xxx.1_2.yyy.x86_64.rpm
IB performance tests
C-6IB0056101-00 G
a
Optional for
OpenFabrics
Optional for
OpenFabrics
Optional for
OpenFabrics
Required for
OpenSM
Optional for
OpenFabrics
Loading...
+ hidden pages
You need points to download manuals.
1 point = 1 manual.
You can buy points or you can get point for every manual you upload.