Information furnished in this manual is believed to be accurate and reliable. However, QLogic Corporation assumes no
responsibility for its use, nor for any infringements of patents or other rights of third parties which may result from its use.
QLogic Corporation reserves the right to change product specifications at any time without notice. Applications described
in this document for any of these products are for illustrative purposes only. QLogic Corporation makes no representation
nor warranty that such applications are suitable for the specified use without further testing or modification. QLogic
Corporation assumes no responsibility for any errors that may appear in this document.
No part of this document may be copied nor reproduced by any means, nor translated nor transmitted to any magnetic
medium without the express written consent of QLogic Corporation. In accordance with the terms of their valid PathScale
agreements, customers are permitted to make electronic and paper copies of this document for their own exclusive use.
Linux is a registered trademark of Linus Torvalds.
QLA, QLogic, SANsurfer, the QLogic logo, PathScale, the PathScale logo, and InfiniPath are registered trademarks
of QLogic Corporation.
Red Hat and all Red Hat-based trademarks are trademarks or registered trademarks of Red Hat, Inc.
SuSE is a registered trademark of SuSE Linux AG.
All other brand and product names are trademarks or registered trademarks of their respective owners.
Document Revision History
Rev. 1.0, 8/20/2005
Rev. 1.1, 11/15/05
Rev. 1.2,02/15/06
Rev. 1.3 Beta 1, 4/15/06
Rev. 1.3, 6/15/06
Rev. 2.0 Beta, 9/25/06,
QLogic Rev IB6054601 A
Rev. 2.0 Beta 2, 10/15/06,
QLogic Rev IB6054601 B
Rev. 2.0, 11/30/06,
QLogic Rev IB6054601 C
Rev. 2.0, 3/23/07,
QLogic Rev IB6054601 D
Rev. D ChangeDocument Sections Affected
Added metadata to pdf document onlyPDF metadata
Rev. C ChangesDocument Sections Affected
Updated Preface and Overview by combining into single section, now
called Introduction. Same as introduction in Install Guide.
Added SLES 9 as new supported distribution1.7
Revised info about MTRR mapping in BIOS. Some BIOS’ don’t have it,
or call it something else.
Corrected usage of ipath_core, replacing with ib_ipath2
Added more options to mpirun man page description3.5.10
Added new section on Environment for Multiple Versions of InfiniPath
or MPI
Added info on support for multiple MPIs3.6
C.1, C.2.1, C.2.2, C.2.3
1
3.5.8.1
Q
Page iiIB6054601-00 D
Q
InfiniPath User Guide
Version 2.0
Added info about using MPI over uDAPL. Need to load modules
rdma_cm and rdma_ucm.
Added section: Error messages generated by mpirun. This explains
more about the types of errors found in the sub-sections. Also added
error messages related to failed connections between nodes
Added mpirun error message about stray processes to error message
section
Added driver and link error messages reported by MPI programsC.8.12.3
Added section about errors occurring when different runtime/compile
time MPI versions are used
2.0 mpirun incompatible with 1.3 librariesC.8.1
Added glossary entry for MTRRE
Added new index entries for MPI error messages format, corrected
index formatting
This chapter describes the objectives, intended audience, and organization of the
InfiniPath User Guide.
The InfiniPath User Guide is intended to give the end users of an InifiniPath cluster
what they need to know to use it. In this case, end users are understood to include
both the cluster administrator and the MPI application programmers, who have
different but overlapping interests in the details of the technology.
For specific instructions about installing the InfiniPath QLE7140 PCI Express™
adapter, the QMI7140 adapter, or the QHT7140 /QHT7040 HTX™ adapters, and
the initial installation of the InifiniPath Software, see the InfiniPath Install Guide.
1.1
Who Should Read this Guide
This guide is intended both for readers responsible for administration of an InfiniPath
cluster network and for readers wanting to use that cluster.
This guide assumes that all readers are familiar with cluster computing, that the
cluster administrator reader is familiar with Linux administration and that the
application programmer reader is familiar with MPI.
Section 1
Introduction
1.2
How this Guide is Organized
The InfiniPath User Guide is organized into these sections:
■ Section 1"Introduction". This section.
■ Section 2 “InfiniPath Cluster Administration” describes the lower levels of the
supplied InfiniPath software. This would be of interest mainly to an InfiniPath
cluster administrator.
■ Section 3 “Using InfiniPath MPI” helps the MPI programmer make best use of
the InfiniPath MPI implementation.
■ Appendix A “Benchmark Programs”
■ Appendix B “Integration with a Batch Queuing System”
■ Appendix C “Troubleshooting”. The Troubleshooting section provides
information for troubleshooting installation, cluster administration, and MPI.
■ Appendix D “Recommended Reading”
IB6054601-00 D1-1
1 – Introduction
Interoperability
1.3
Overview
Q
■ Appendix E Glossary of technical terms
■ Index
In addition, the InfiniPath Install Guide contains information on InfiniPath hardware
and software installation.
The material in this documentation pertains to an InfiniPath cluster. This is defined
as a collection of nodes, each attached to an InfiniBand™-based fabric through the
InfiniPath Interconnect. The nodes are Linux-based computers, each having up to
eight processors.
The InfiniPath interconnect is InfiniBand 4X, with a raw data rate of 10 Gb/s (data
rate of 8Gb/s).
InfiniPath utilizes standard, off-the-shelf InfiniBand 4X switches and cabling.
InfiniPath OpenFabrics software is interoperable with other vendors’ InfiniBand
HCAs running compatible OpenFabrics releases. There are two options for Subnet
Management in your cluster:
■ Use the Subnet Manager on one or more managed switches supplied with your
Infiniband switches.
■ Use the OpenSM component of OpenFabrics.
1.4
Switches
The InfiniPath interconnect is designed to work with all InfiniBand-compliant
switches. Use of OpenSM as a subnet manager is now supported. OpenSM is part
of the OpenFabrics component of this release.
1.5
Interoperability
InfiniPath participates in the standard InfiniBand Subnet Management protocols for
configuration and monitoring. InfiniPath OpenFabrics (including IPoIB) is
interoperable with other vendors’ InfiniBand HCAs running compatible OpenFabrics
releases. The InfiniPath MPI and Ethernet emulation stacks (
interoperable with other InfiniBand Host Channel Adapters (HCA) and Target
Channel Adapters (TCA). Instead, InfiniPath uses an InfiniBand-compliant
vendor-specific protocol that is highly optimized for MPI and TCP between
InfiniPath-equipped hosts.
ipath_ether) are not
1-2IB6054601-00 D
Q
NOTE:OpenFabrics was known as OpenIB until March 2006. All relevant
references to OpenIB in this documentation have been updated to reflect
this change. See the OpenFabrics website at http://www.openfabrics.org
for more information on the OpenFabrics Alliance.
1.6
What’s New in this Release
QLogic Corp. acquired PathScale in April 2006. In this 2.0 release, product names,
internal program and output message names now refer to QLogic rather than
PathScale.
The new QLogic and former PathScale adapter model numbers are shown in the
table below.
Table 1-1. PathScale-QLogic Adapter Model Numbers
Former
PathScale
Model Number
HT-400IBA6110Single Port 10GBS InfiniBand to HTX ASIC
PE-800IBA6120Single Port 10GBS InfiniBand to x8 PCI Express
HT-460QHT7040Single Port 10GBS InfiniBand to HTX Adapter
HT-465QHT7140Single Port 10GBS InfiniBand to HTX Adapter
PE-880QLE7140Single Port 10GBS InfiniBand to x8 PCI Express
PE-850QMI7140Single Port 10GBS InfiniBand IBM Blade Center
New QLogic Model
Number
1 – Introduction
What’s New in this Release
Description
ROHS
ASIC ROHS
Adapter
Adapter
This version of InfiniPath provides support for all QLogic’s HCAs, including:
■ InfiniPath QLE7140, which is supported on systems with PCIe x8 or x16 slots
■ InfiniPath QMI7140, which runs on Power PC systems, particularly on the IBM®
BladeCenter H processor blades
■ InfiniPath QHT7040 and QHT7140, which leverage HTX™. The InfiniPath
QHT7040 and QHT7140 are exclusively for motherboards that support
HTXcards. The QHT7140 has a smaller form factor than the QHT7040, but is
otherwise the same. Unless otherwise stated, QHT7140 will refer to both the
QHT7040 and QHT7140 in this documentation.
Expanded MPI scalability enhancements for PCI Express have been added. The
QHT7040 and QHT7140 can support 2 processes per context for a total of 16. The
QLE7140 and QMI7140 also support 2 processes per context, for a total of 8.
IB6054601-00 D1-3
1 – Introduction
Supported Distributions and Kernels
Support for multiple versions of MPI has been added. You can use a different version
of MPI and achieve the high-bandwidth and low-latency performance that is
standard with InfiniPath MPI.
Also included is expanded operating system support, and support for the latest
OpenFabrics software stack.
Multiple InfiniPath cards per node are supported. A single software installation works
for all the cards.
Additional up-to-date information can be found on the QLogic web site:
http://www.qlogic.com
1.7
Supported Distributions and Kernels
The InfiniPath interconnect runs on AMD Opteron, Intel EM64T, and IBM Power
Blade Center H) systems running Linux. The currently supported distributions and
associated Linux kernel versions for InfiniPath and OpenFabrics are listed in the
following table. The kernels are the ones that shipped with the distributions, unless
otherwise noted.
Q
Table 1-2. InfiniPath/OpenFabrics Supported Distributions and Kernels
InfiniPath/OpenFabrics supported
Distribution
Fedora Core 3 (FC3)2.6.12 (x86_64)
Fedora Core 4 (FC4)2.6.16, 2.6.17 (x86_64)
Red Hat Enterprise Linux 4 (RHEL4)2.6.9-22, 2.6.9-34, 2.6.9-42(U2/U3/U4)
(x86_64)
CentOS 4.2-4.4 (Rocks 4.2-4.4)2.6.9 (x86_64)
SUSE Linux 9.3 (SUSE 9.3)2.6.11 (x86_64)
SUSE LInux Enterprise Server (SLES 9)2.6.5 (x86_64)
SUSE LInux Enterprise Server (SLES 10)2.6.16 (x86_64 and ppc64)
NOTE:IBM Power systems run only with the SLES 10 distribution.
The SUSE10 release series is no longer supported as of this InfiniPath 2.0 release.
Fedora Core 4 kernels prior to 2.6.16 are also no longer supported.
kernels
1-4IB6054601-00 D
Q
1.8
Software Components
The software provided with the InfiniPath Interconnect product consists of:
■ InfiniPath driver (including OpenFabrics)
■ InfiniPath ethernet emulation
■ InfiniPath libraries
■ InfiniPath utilities, configuration, and support tools
■ InfiniPath MPI
■ InfiniPath MPI benchmarks
■ OpenFabrics protocols, including Subnet Management Agent
■ OpenFabrics libraries and utilities
1 – Introduction
Software Components
OpenFabrics kernel module support is now built and installed as part of the InfiniPath
RPM install. The InfiniPath release 2.0 runs on the same code base as OpenFabrics
Enterprise Distribution (OFED) version 1.1. It also includes the OpenFabrics
1.1-based library and utility RPMs. InfiniBand protocols are interoperable between
InfiniPath 2.0 and OFED 1.1.
This release provides support for the following protocols:
■ IPoIB (TCP/IP networking)
■ SDP (Sockets Direct Protocol)
■ OpenSM
■ UD (Unreliable Datagram)
■ RC (Reliable Connection)
■ UC (Unreliable Connection)
■ SRQ (Shared Receive Queue)
■ uDAPL (user Direct Access Provider Library)
This release includes a technology preview of:
■ SRP (SCSI RDMA Protocol)
Future releases will provide support for:
■ iSER (iSCSI Extensions for RDMA)
No support is provided for RD.
IB6054601-00 D1-5
1 – Introduction
Documentation and Technical Support
NOTE:32 bit OpenFabrics programs using the verb interfaces are not supported
in this InfiniPath release, but will be supported in a future release.
1.9
Conventions Used in this Document
This Guide uses these typographical conventions:
Table 1-3. Typographical Conventions
ConventionMeaning
commandFixed-space font is used for literal items such as commands,
functions, programs, files and pathnames, and program
output;
variableItalic fixed-space font is used for variable names in programs
and command lines.
conceptItalic font is used for emphasis, concepts.
user inputBold fixed-space font is used for literal items in commands or
constructs that you type in.
$Indicates a command line prompt.
#Indicates a command line prompt as root when using bash or
sh.
[ ]Brackets enclose optional elements of a command or
program construct.
...Ellipses indicate that a preceding element can be repeated.
>Right caret identifies the cascading path of menu commands
used in a procedure.
2.0The current version number of the software is included in the
RPM names and within this documentation.
NOTE:Indicates important information.
Q
1.10
Documentation and Technical Support
The InfiniPath product documentation includes:
■ The InfiniPath Install Guide
■ The InfiniPath User Guide
■ Release Notes
■ Quick Start Guide
1-6IB6054601-00 D
Q
1 – Introduction
Documentation and Technical Support
■ Readme file
The Troubleshooting Appendix for installation, InfiniPath and OpenFabrics
administration, and MPI issues is located in the InfiniPath User Guide.
Visit the QLogic support Web site for documentation and the latest software updates.
http://www.qlogic.com
IB6054601-00 D1-7
1 – Introduction
Documentation and Technical Support
Notes
Q
1-8IB6054601-00 D
This chapter describes what the cluster administrator needs to know about the
InfiniPath software and system administration.
2.1
Introduction
The InfiniPath driver ib_ipath, layered Ethernet driver ipath_ether, OpenSM,
and other modules and the protocol and MPI support libraries are the components
of the InfiniPath software providing the foundation that supports the MPI
implementation.
Figure 2-1, below, shows these relationships.
Section 2
InfiniPath Cluster Administration
MPI Application
InfiniPath Channel (ADI Layer)
InfiniPath Protocol Library
InfiniPath Hardw are
2.2
Installed Layout
The InfiniPath software is supplied as a set of RPM files, described in detail in the
InfiniPath Install Guide. This section describes the directory structure that the
installation leaves on each node’s file system.
The InfiniPath shared libraries are installed in:
/usr/lib for 32-bit applications
/usr/lib64 for 64-bit applications
MPI programming examples and source for several MPI benchmarks are in:
/usr/share/mpich/examples
InfiniPath utility programs, as well as MPI utilities and benchmarks are installed in:
/usr/bin
The InfiniPath kernel modules are installed in the standard module locations in:
/lib/modules (version dependent)
They are compiled and installed when the infinipath-kernel RPM is installed.
They must be rebuilt and re-installed when the kernel is upgraded. This can be done
by running the script:
The following is a preliminary guideline for estimating the memory footprint of the
InfiniPath adapter on Linux x86_64systems. Memory consumption is linear based
2-2IB6054601-00 D
Q
2 – InfiniPath Cluster Administration
Memory Footprint
on system configuration. OpenFabrics support is under development and has not
been fully characterized. This table summarizes the guidelines.
Table 2-1. Memory Footprint of the InfiniPath Adapter on Linux x86_64 Systems
Adapter
component
InfiniPath DriverRequired9 MBIncludes accelerated IP
MPIOptional71 MB per process with
OpenFabricsOptional1~6 MB
Required/
optional
Memory FootprintComment
support. Includes tables
space to support up to
1000 node systems.
Clusters larger than 1000
nodes can also be
configured.
+ 32 MB per node when
multiple processes
communicate via shared
memory
+ 264 Bytes per MPI node
on the subnet
+ ~500 bytes per QP
+ TBD bytes per MR
+ ~500 bytes per EE
Context
+ OpenFabrics stack from
openfabrics.org (size not
included in these
guidelines)
parameters (sendbufs,
recvbufs and size of the
shared memory region)
are tunable if reduced
memory footprint is
desired.
This not been fully
characterized as of this
writing.
Here is an example for a 1024 processor system:
■ 1024 cores over 256 nodes (each node has 2 sockets with dual-core processors)
■ 1 adapter per node
■ Each core runs an MPI process, with the 4 processes per node communicating
via shared memory.
■ Each core uses OpenFabrics to connect with storage and file system targets
using 50 QPs and 50 EECs per core.
IB6054601-00 D2-3
2 – InfiniPath Cluster Administration
Configuration and Startup
This breaks down to a memory footprint of 331MB per node, as follows:
Table 2-2. Memory Footprint, 331 MB per Node
ComponentFootprint (in MB)Breakdown
Driver9Per node
MPI316 4*71 MB (MPI per process)
OpenFabrics66 MB + 200 KB per node
2.4
Configuration and Startup
2.4.1
BIOS Settings
A properly configured BIOS is required. The BIOS settings, which are stored in
non-volatile memory, contain certain parameters characterizing the system,. These
parameters may include date and time, configuration settings, and information about
the installed hardware.
Q
+ 32 MB (shared memory
per node)
There are currently two issues concerning BIOS settings that you need to be aware
of:
■ ACPI needs to be enabled
■ MTRR mapping needs to be set to “Discrete”
MTRR (Memory Type Range Registers) is used by the InfiniPath driver to enable
write combining to the InfiniPath on-chip transmit buffers. This improves write
bandwidth to the InfiniPath chip by writing multiple words in a single bus transaction
(typically 64). This applies only to x86_64 systems.
However, some BIOSes don’t have the MTRR mapping option. It may be referred
to in a different way, dependent upon chipset, vendor, BIOS, or other factors. For
example, it is sometimes referred to as "32 bit memory hole", which should be
enabled.
If there is no setting for MTRR mapping or 32 bit memory hole, please contact your
system or motherboard vendor and inquire as to how write combining may be
enabled.
ACPI and MTRR mapping issues are discussed in greater detail in the
Troubleshooting section of the InfiniPath User Guide.
NOTE:BIOS settings on IBM Blade Center H (Power) systems do not need
adjustment.
2-4IB6054601-00 D
Q
You can check and adjust these BIOS settings using the BIOS Setup Utility. For
specific instructions on how to do this, follow the hardware documentation that came
with your system.
2.4.2
InfiniPath Driver Startup
The ib_ipath module provides low level InfiniPath hardware support. It does
hardware initialization, handles infinipath-specific memory management, and
provides services to other InfiniPath and OpenFabrics modules. It provides the
management functions for InfiniPath MPI programs, the ipath_ether ethernet
emulation, and general OpenFabrics protocols such as IPoIB, and SDP. It also
contains a Subnet Management Agent.
The InfiniPath driver software is generally started at system startup under control
of these scripts:
/etc/init.d/infinipath
/etc/sysconfig/infinipath
2 – InfiniPath Cluster Administration
Configuration and Startup
These scripts are configured by the installation. Debug messages are printed with
the function name preceding the message.
The cluster administrator does not normally need to be concerned with the
configuration parameters. Assuming that all the InfiniPath and OpenFabrics
software has been installed, the default settings upon startup will be:
■ InfiniPath ib_ipath is enabled
■ InfiniPath ipath_ether is not running until configured
■ OpenFabrics IPoIB is not running until configured
■ OpenSM is enabled on startup. Disable it on all nodes except where it will be
used as subnet manager.
2.4.3
InfiniPath Driver Software Configuration
The ib_ipath driver has several configuration variables which provide for setting
reserved buffers for the software, defining events to create trace records, and setting
debug level. See the
2.4.4
ib_ipath man page for details.
InfiniPath Driver Filesystem
The InfiniPath driver supplies a filesystem for exporting certain binary statistics to
user applications. By default, this filesystem is mounted in the
when the infinipath script is invoked with the "start" option (e.g. at system startup)
IB6054601-00 D2-5
/ipathfs directory
2 – InfiniPath Cluster Administration
Configuration and Startup
and unmounted when the infinipath script is invoked with the "stop" option (e.g. at
system shutdown).
The layout of the filesystem is as follows:
atomic_stats
00/
01/
...
The atomic_stats file contains general driver statistics. There is one numbered
directory per InfiniPath device on the system. Each numbered directory contains
the following files of per-device statistics:
atomic_counters
node_info
port_info
The atomic_counters file contains counters for the device: examples would be
interrupts received, bytes and packets in and out, and so on. The
contains information such as the device’s GUID. The
information for each port on the device. An example would be the port LID.
Q
node_info file
port_info file contains
2.4.5
Subnet Management Agent
Each node in an InfiniPath cluster runs a Subnet Management Agent (SMA), which
carries out two-way communication with the Subnet Manager (SM) running on one
or more managed switches. The Subnet Manager is responsible for network
initialization (topology discovery), configuration, and maintenance. The Subnet
Manager also assigns and manages InfiniBand multicast groups, such as the group
used for broadcast purposes by the
the SMA are to keep the SM informed whether a node is alive and to get the node’s
assigned identifier (LID) from the SM.
2.4.6
Layered Ethernet Driver
The layered Ethernet component ipath_ether provides almost complete Ethernet
software functionality over the InfiniPath fabric. At startup this is bound to some
Ethernet device
transparent way, except that Ethernet multicasting is not supported. Broadcasting
is supported. You can use all the usual command line and GUI-based configuration
tools on this Ethernet. Configuration of
These instructions are for enabling TCP-IP networking over the InfiniPath link. To
enable IPoIB networking, see section 2.4.7.1.
ethx. All Ethernet functions are available through this device in a
ipath_ether driver. The primary functions of
ipath_ether is optional.
2-6IB6054601-00 D
2 – InfiniPath Cluster Administration
Q
You must create a network device configuration file for the layered Ethernet device
on the InfiniPath adapter. This configuration file will resemble the configuration files
for the other Ethernet devices on the nodes. Typically on servers there are two
Ethernet devices present, numbered as 0 (eth0) and 1 (eth1). This examples
assumes we create a third device, eth2.
NOTE:When multiple InfiniPath chips are present, the configuration for eth3,
eth4, and so on follow the same format as for adding eth2 in the examples
below.
Two slightly different procedures are given below for the ipath configuration; one
for Fedora and one for SUSE, SLES9, or SLES 10.
Many of the entries that are used in the configuration directions below are explained
in the file sysconfig.txt. To familiarize yourself with these, please see:
/usr/share/doc/initscripts-*/sysconfig.txt
2.4.6.1
ipath_ether Configuration on Fedora and RHEL4
Configuration and Startup
These configuration steps will cause the ipath_ether network interfaces to be
automatically configured when you next reboot the system. These instructions are
for the
Typically on servers there are two Ethernet devices present, numbered as 0 (eth0)
and 1 (eth1). This example assumes we create a third device, eth2.
NOTE:When multiple InfiniPath chips are present, the configuration for eth3,
Fedora Core 3, Fedora Core 4 and Red Hat Enterprise Linux 4 distributions.
eth4, and so on follow the same format as for adding eth2 in the
examples below.
1. Check for the number of Ethernet drivers you currently have by either one of
the two following commands :
$ ifconfig -a
$ ls /sys/class/net
As mentioned above we assume that two Ethernet devices (numbered 0 and
1) are already present.
2. Edit the file
alias eth2 ipath_ether
3. Create or edit the following file (as root).
/etc/sysconfig/network-scripts/ifcfg-eth2
/etc/modprobe.conf (as root) by adding the following line:
IB6054601-00 D2-7
2 – InfiniPath Cluster Administration
Configuration and Startup
If you are using DHCP (dynamic host configuration protocol), add the following
lines to ifcfg-eth2:
If you are using static IP addresses, use the following lines instead, substituting
your own IP address for the sample one given here
netmask is shown.
# QLogic Interconnect Ethernet
DEVICE=eth2
BOOTPROTO=static
ONBOOT=YES
IPADDR=192.168.5.101 #Substitute your IP address here
NETMASK="255.255.255.0"#Normal matching netmask
TYPE=Ethernet
This will cause the ipath_ether Ethernet driver to be loaded and configured during
system startup. To check your configuration, and make the
driver available immediately, use the command (as root):
Q
.The normal matching
ipath_ether Ethernet
# /sbin/ifup eth2
4. Check whether the Ethernet driver has been loaded with:
$ lsmod | grep ipath_ether
5. Verify that the driver is up with:
$ ifconfig -a
2.4.6.2
ipath_ether Configuration on SUSE 9.3, SLES 9, and SLES 10
These configuration steps will cause the ipath_ether network interfaces to be
automatically configured when you next reboot the system. These instructions are
for the
Typically on servers there are two Ethernet devices present, numbered as 0 (eth0)
and 1 (eth1). This example assumes we create a third device, eth2.
NOTE:When multiple InfiniPath chips are present, the configuration for eth3,
SUSE 9.3, SLES 9 and SLES 10 distributions.
eth4, and so on follow the same format as for adding eth2 in the
examples below. Similarly , in step 2, add one to the unit number, so
replace
and so on.
.../00/guid with /01/guid for the second InfiniPath interface,
2-8IB6054601-00 D
Q
2 – InfiniPath Cluster Administration
Configuration and Startup
Step 3 is applicable only to SLES 10; it is required because SLES 10 uses a newer
version of the
NOTE:The MAC address (media access control address) is a unique identifier
The following steps must all be executed as the root user.
# sed ’s/^\(..:..:..\):..:../\1/’ \
/sys/bus/pci/drivers/ib_ipath/00/guid
NOTE:Care should be taken when cutting and pasting commands such as
udev subsystem.
attached to most forms of networking equipment. Step 2 below determines
the MAC address to use, and will be referred to as $MAC in the
subsequent steps. $MAC must be replaced in each case with the string
printed in step 2.
the above from PDF documents, as quotes are special characters
and may not be translated correctly.
The output should appear similar to this (6 hex digit pairs, separated by colons):
Note that removing the middle two 00:00 octets from the GUID in the above
output will form the MAC address
If either step 1 or step 2 fails in some fashion, the problem must be found and
corrected before continuing. Verify that the RPMs are installed correctly, and
that infinipath has correctly been started. If problems continue, run
ipathbug-helper and report the results to your reseller or InfiniPath support
organization.
3. Skip to Step 4 if you are using SUSE 9.3 or SLES 9. This step is only done on
SLES 10 systems. Edit the file:
/etc/udev/rules.d/30-net_persistent_names.rules
If this file does not exist, skip to Step 4.
IB6054601-00 D2-9
2 – InfiniPath Cluster Administration
Configuration and Startup
Check each of the lines starting with SUBSYSTEM=, to find the highest numbered
interface. (For standard motherboards, the highest numbered interface will
typically be 1.)
Add a new line at the end of the file, incrementing the interface number by one.
In this example, it becomes eth2. The new line will look like this:
Make sure that you substitute your own IP address for the sample IPADDR
shown here. The BROADCAST, NETMASK, and NETWORK lines need to
match for your network.
2-10IB6054601-00 D
Q
6. To verify that the configuration files are correct, you will normally now be able
to run the commands:
# ifup eth2
# ifconfig eth2
Note that it may be necessary to reboot the system before the configuration
changes will work.
2.4.7
OpenFabrics Configuration and Startup
In the prior InfiniPath 1.3 release the InfiniPath (ipath_core) and OpenFabrics
(ib_ipath) modules were separate. In this release there is now one module,
ib_ipath, which provides both low level InfiniPath support and management
functions for OpenFabrics protocols. The startup script for ib_ipath is installed
automatically as part of the software installation, and normally does not need to be
changed.
2 – InfiniPath Cluster Administration
Configuration and Startup
However, the IPoIB network interface and OpenSM components of OpenFabrics
can be configured to be on or off. IPoIB is off by default; OpenSM is on by default.
IPoIB and OpenSM configuration is explained in greater detail in the following
sections.
NOTE:The following instructions work for FC4, SUSE9.3, SLES 9, and SLES 10.
2.4.7.1
Configuring the IPoIB Network Interface
Instructions are given here to manually configure your OpenFabrics IPoIB network
interface. This example assumes that you are using sh or bash as your shell, and
that all required InfiniPath and OpenFabrics RPMs are installed, and your startup
scripts have been run, either manually or at system boot.
For this example, we assume that your IPoIB network is 10.1.17.0 (one of the
networks reserved for private use, and thus not routable on the internet), with a /8
host portion, and therefore requires that the netmask be specified.
This example assumes that no hosts files exist, and that the host being configured
has the IP address 10.1.17.3, and that DHCP is not being used.
NOTE:We supply instructions only for this static IP address case. Configuration
methods for using DHCP will be supplied in a later release.
Type the following commands (as root):
# ifconfig ib0 10.1.17.3 netmask 0xffffff00
IB6054601-00 D2-11
2 – InfiniPath Cluster Administration
Configuration and Startup
To verify the configuration, type:
# ifconfig ib0
The output from this command should be similar to this:
ib0 Link encap:InfiniBand HWaddr
00:00:00:02:FE:80:00:00:00:00:00:00:00:00:00:00:00:00:00:00
inet addr:10.1.17.3 Bcast:10.1.17.255 Mask:255.255.255.0
UP BROADCAST RUNNING MULTICAST MTU:2044 Metric:1
RX packets:0 errors:0 dropped:0 overruns:0 frame:0
TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:128
RX bytes:0 (0.0 b) TX bytes:0 (0.0 b)
Next, type:
# ping -c 2 -b 10.1.17.255
The output of the ping command should be similar to that below, with a line for
each host already configured and connected:
WARNING: pinging broadcast address
PING 10.1.17.255 (10.1.17.255) 517(84) bytes of data.
174 bytes from 10.1.17.3: icmp_seq=0 ttl=174 time=0.022 ms
64 bytes from 10.1.17.1: icmp_seq=0 ttl=64 time=0.070 ms (DUP!)
64 bytes from 10.1.17.7: icmp_seq=0 ttl=64 time=0.073 ms (DUP!)
Q
2.4.8
OpenSM
The IPoIB network interface is now configured.
NOTE:The configuration must be repeated each time the system is rebooted.
OpenSM is an optional component of the OpenFabrics project that provides a
subnet manager for InfiniBand networks. This package can be installed on all
machines, but only needs to be enabled on the machine in your cluster that is going
to act as a subnet manager. You do not need to use OpenSM if any of your InfiniBand
switches provide a subnet manager.
After installing the opensm package, OpenSM is configured to be on on the next
machine reboot. It only needs to be enabled on the node which acts as the subnet
manager, so use the chkconfig command (as root) to disable it on the other nodes:
# chkconfig opensmd off
The command to enable it on reboot is:
# chkconfig opensmd on
You can start opensmd without rebooting your machine as follows:
# /etc/init.d/opensmd start
2-12IB6054601-00 D
Loading...
+ 92 hidden pages
You need points to download manuals.
1 point = 1 manual.
You can buy points or you can get point for every manual you upload.