Extreme Networks Alpine 3804, BlackDiamond 6808, Alpine 3808, Summit1i, Summit5i Troubleshooting Manual

...
Advanced System Diagnostics and Troubleshooting Guide
Extreme Networks, Inc.
3585 Monroe Street
Santa Clara, California 95051
(888) 257-3000
http://www.extremenetworks.com
Published: March 2005
Part number: 100189-00 Rev. 01
A E P S
©2005 Extreme Networks, Inc. All rights reserved. Extreme Networks, ExtremeWare, Alpine, and BlackDiamond are registered trademarks of Extreme Networks, Inc. in the United States and certain other jurisdictions. ExtremeWare Vista, ExtremeWorks, ExtremeAssist, ExtremeAssist1, ExtremeAssist2, PartnerAssist, EPICenter, Extreme Standby Router Protocol, ESRP, SmartTraps, Summit, Summit1i, Summit5i, Summit7i,Summit48i, Summit48si, SummitPx, Summit 200, Summit Virtual Chassis, SummitLink, SummitGbX, SummitRPS and the Extreme Networks logo are trademarks of Extreme Networks, Inc., which may be registered or pending registration in certain jurisdictions. The Extreme Turbodrive logo is a service mark of Extreme Networks, which may be registered or pending registration in certain jurisdictions. Specifications are subject to change without notice.
NetWare and Novell are registered trademarks of Novell, Inc. Merit is a registered trademark of Merit Network, Inc. Solaris is a trademark of Sun Microsystems, Inc. F5, BIG/ip, and 3DNS are registered trademarks of F5 Networks, Inc. see/IT is a trademark of F5 Networks, Inc.
“Data Fellows”, the triangle symbol, and Data Fellows product names and symbols/logos are trademarks of Data Fellows.
F-Secure SSH is a registered trademark of Data Fellows.
All other registered trademarks, trademarks and service marks are property of their respective owners.
uthors: ditor: roduction: pecial Thanks:
2
Contents
Preface
Introduction 9 Terminology 9 Conventions 9 Related Publications 10
Chapter 1 Introduction
Introduction 11 Diagnostics: A Brief Historical Perspective 12 Overview of the ExtremeWare Diagnostics Suite 13 Supported Hardware 13 Applicable ExtremeWare Versions 13
Chapter 2 Hardware Architecture
Diagnostics Support 15 The BlackDiamond Systems 16
BlackDiamond 6800 Series Hardware Architecture Differences 16 The BlackDiamond Backplane 17 BlackDiamond I/O Modules 18 Management Switch Modules 19
BlackDiamond MSM Redundancy 20 Causes of MSM Failover and System Behavior 20
Alpine Systems 22 Summit Systems 23
Advanced System Diagnostics and Troubleshooting Guide 3
Contents
Chapter 3 Packet Errors and Packet Error Detection
Overview 25 Definition of Terms 26 Standard Ethernet Detection for Packet Errors on the Wire 27 Extreme Networks’ Complementary Detectio n of Packet Errors Between Wires 27
Hardware System Detection Mechanisms 28 Software System Detection Mechanisms 29
Failure Modes 30
Transient Failures 30 Systematic Failures 30
Soft-State Failures 30 Permanent Failures 31 Responding to Failures 31
Error Messages for Fabric Checksums 31
Error Message Format 32 Fabric Checksum Error Message Logging 33 Checksum Message Examples 34
Panic/Action Error Messages 35
Panic/Action Message Example 36
Chapter 4 Software Exception Handling
Overview of Software Exception Handling Features 37
System Watchdog Behavior 37 System Software Exception Recovery Behavior 38 Redundant MSM Behavior 38
Configuring System Recovery Actions 40
Related Commands 40 Configuring System Recovery Actions 40 Usage Notes 41
Configuring Reboot Loop Protection 42
Related Commands 42 Configuring Reboot Loop Protection 42
Dumping the System Memory 44
Related Commands 44 Configuring an Automatic System Dump During System Recovery 44 Initiating a Manual System Dump 45
Chapter 5 Diagnostics
Diagnostic Test Functionality 47
How Diagnostic Tests are Run 47 How the Test Affects the Switch 48
4 Advanced System Diagnostics and Troubleshooting Guide
Contents
System Health Checks: A Diagnostics Suite 51
Diagnostic Suite Components 51 The Role of Memory Scanning and Memory Mapping 52
Modes of Operation 53
The Role of Processes to Monitor System Operation 54
Power On Self Test (POST) 55
Related Commands 55 Configuring the Boot-Up Diagnostics 55
Runtime (On-Demand) System Diagnostics 56
Normal System Diagnostics 56
Related Commands 56 Running the Normal Diagnostics on BlackDiamond Systems 57 Running the Normal Diagnostics on Alpine Systems 57 Running the Normal Diagnostics on Summit Systems 57 System Impact of Running the Normal Diagnostics 57
Extended System Diagnostics 58
Related Commands 58 Running the Extended Diagnostics on BlackDiamond Systems 58 Running the Extended Diagnostics on Alpine Systems 58 Running the Extended Diagnostics on Summit Systems 59 System Impact of Running the Extended Diagnostics 59
On Demand Packet Memory Scan 59
Related Commands 59 Running the Packet Memory Scan Diagnostics on BlackDiamond Systems 59 Running the Packet Memory Scan Diagnostics on Alpine Systems 59 Running the Packet Memory Scan Diagnostics on Summit Systems 60
Automatic Packet Memory Scan (via sys-health-check) 60
Memory Scanning and Memory Mapping Behavior 60 Limited Operation Mode 63 Effects of Running Memory Scanning on the Switch 63
Summit, Alpine, or BlackDiamond with a Single MSM 63 BlackDiamond System with Two MSMs 64
Interpreting Memory Scanning Results 65
Per-Slot Packet Memory Scan on BlackDiamond Switches 66
Related Commands 66 Configuring the Packet Memory Scan Recovery Mode 66 System Impact of Per-Slot Packet Memory Scanning 67 Network Impact of Per-Slot Packet Memory Scanning 67
System (CPU and Backplane) Health Check 69
Health Check Packet Types 69 Backplane Health Check States 69 Related Commands 69 Health Check Functionality 70
Alarm-Level Response Action 70 Auto-Recovery Response Action 70
Backplane Health Check 71
Advanced System Diagnostics and Troubleshooting Guide 5
Contents
Viewing Backplane Health Check Results—show log Command 71 Viewing Backplane Health Check Diagnostic Results—show diagnostics Command 71 Analyzing the Results 76
CPU Health Check 77
Viewing CPU Health Check Results—show log Command 77 Viewing CPU Health Check Diagnostic Results—show diagnostics Command 78 Analyzing the CPU Health Check Results 78
Transceiver Diagnostics 80
Usage Guidelines 80 Related Commands 80 Configuring the Transceiver Diagnostics 80 System Impacts of the Transceiver Diagnostics 81 Network Impact of the Transceiver Diagnostics 81 Viewing Diagnostics Results 82
Example Log Messages for Transceiver Diagnostic Failures 82 Examples, show diagnostics Command 82 Example—show switch Command 84 Transceiver Diagnostic Res ult Analysis 84
FDB Scan 85
Usage Guidelines 85 Related Commands 86
Enabling FDB Scanning 86 Disabling FDB Scanning 86 Configuring the FDB Scan Diagnostics 86
System Impact of the FDB Scan Diagnostic 87 Network Impact of the FDB Scan Diagnostic 87 Viewing Diagnostics Results 88
Example Log Messages for FDB Scan Diagnostic Failures 88 Example FDB Scan Results from the show diagnostics Command 88 Example Output from the show switch command 89 Example Output from the show fdb remap Command 89
Chapter 6 Additional Diagnostics Tools
Temperature Logging 92
Related Commands 92 System Impacts of Temperature Logging 92 Network Impact of Temperature Logging 92
Syslog Servers 93
Related Commands 93
Enabling Logging to Remote Syslog Server Targets 93 Disabling Logging to Remote Syslog Server Ta rgets 93 Adding a Syslog Server 94 Deleting a Remote Syslog Server 94
System Impact of the Syslog Server Facility 94 Network Impact of the Syslog Server Facility 94
6 Advanced System Diagnostics and Troubleshooting Guide
Chapter 7 Troubleshooting Guidelines
Contacting Extreme Technical Support 95
Americas TAC 95 Asia TAC 96 EMEA TAC 96 Japan TAC 96
What Information Should You Collect? 97 Analyzing Data 97 Diagnostic Troubleshooting 98 Extreme Networks’ Recommendations 99 Using Memory Scanning to Screen I/O Modules 101
Appendix A Limited Operation Mode and Minimal Operation Mode
Limited Operation Mode 103
Triggering Limited Operation Mode 104 Bringing a Switch Out of Limited Operation Mode 104
Contents
Minimal Operation Mode 104
Triggering Minimal Operation Mode 104 Bringing a Switch Out of Minimal Operation Mode 105
Appendix B Reference Documents
General Information 107 Other Documentation Resources 107
Index
Advanced System Diagnostics and Troubleshooting Guide 7
Contents
8 Advanced System Diagnostics and Troubleshooting Guide
Preface
This Preface provides an overview of this guide, describes guide conventions, and lists other publications that might be useful.
Introduction
This guide describes how to use the ExtremeWare hardware diagnostics suite to test and validate the operating integrity of Extreme Networks switches. The tools in the diagnostic suite are used to detect, isolate, and treat faults in a system.
This guide is intended for use by network designers, planners, and operations staff.
Terminology
When features, functionality, or operation is specific to a modular or stand-alone switch fami ly, the family name is used. Explanations about features and operations that are the same across all product families simply refer to the product as the “switch.”
Conventions
Table 1 and Table 2 list conventions that are used throughout this guide.
Table 1: Notice Icons
Icon Notice Type Alerts you to...
Note Important features or instructions.
Caution Risk of personal injury, system damage, or loss of data.
Warning Risk of severe personal injury.
Advanced System Diagnostics and Troubleshooting Guide 9
Preface
Table 2: Te x t C o nventi o ns
Convention Description
Screen displays This typeface indicates command syntax, or represents information as it appears on the
screen.
The words “enter” and “type”
[Key] names Key names are written with brackets, such as [Return] or [Esc].
Letter in bold type Letters within a command that appear in bold type indicate th e keyboard shortcut for a
Words in italicized type Italics emphasize a point or denote new terms at the place where th ey are defined in
When you see the word “enter” in this guide, you must type something, and then press the Return or Enter key. Do not press the Return or Enter key when an instruction simply says “type.”
If you must press two or more keys simultaneously, the key names are linked with a plus sign (+). Example:
Press [Ctrl]+[Alt]+[Del].
command. When entering the command, you can use just the bol ded letters instead of the entire word.
the text.
Related Publications
The publications related to this one are:
ExtremeWare Software User Guide, Software Version 6.2.2.
ExtremeWare Software Command Reference, Software Version 6.2.2.
ExtremeWare Error Message Decoder.
Documentation for Extreme Networks products is available on the World Wide We b at the following location:
http://www.extremenetworks.com/
10 Advanced System Diagnostics and Troubleshooting Guide
1 Introduction
This guide describes how to use the ExtremeWare hardware diagnostics suite to test and validate the operating integrity of Extreme Networks switches. The tools in the diagnostic suite are used to detect, isolate, and treat faults in a system.
This chapter contains the following sections:
Diagnostics: A Brief Historical Perspective on page 12
Overview of the ExtremeWare Diagnostics Suite on page 13
Supported Hardware on page 13
Applicable ExtremeWare Versions on page 13
Introduction
The purpose of this guide is to provide information and guidelines to assist you in implementing the diagnostic suite within ExtremeWare. The Extreme Networks diagnostic software is intended to identify poss ible faul ty h ardwa re or s oftw are e rror c ondi tions and—depending on how the various diagnostics features are configured—take the appropriate preconfigured action. The action might be to enable the switch to write informative error messages to the switch log, attempt to recover itself and continue operating, or simply remove the suspect system component from service.
It is important to note that while each diagnostic test—by itself—is not complicated, sys tem configuration, as a whole, must be evaluated to ensure ongoing, expected behavior both with the switch and across the network itself. For example, in implementing the diagnostic suite, you must take into consideration these operational variables:
Redundancy mechanisms implemented
Levels of redundancy within the network
Acceptable outage windows (scheduled and unscheduled)
Advanced System Diagnostics and Troubleshooting Guide 11
Introduction
Diagnostics: A Brief Historical Perspective
Diagnostic utility programs were created to aid in troubleshooting system problems by detecting and reporting faults so that operators or administrators could go fix the problem. While this approach does help, it has some key limitations:
It is, at its base, reactive, meaning a failure must occur before the diagnostic test can be used to look
for a cause for the failure.
It can be time consuming, because the ability to troubleshoot a failure successfully based on the
information provided by the diagnostics test depends greatly on the types of information reported by the test and the level of detail in the information.
Because users of mission-critical networks and network applications are becoming increasingly dependent on around-the-clock network access and highest performance levels, any downtime or service degradation is disruptive and costly. So time lost to an unexpected failure, compounded by more time lost while someone attempts to track down and fix the failure, has become increasingly less acceptable.
The process of improving diagnostic tests to minimize failures and their impact is a kind of feedback system: What you learn through the use of the diagnostics improves your understanding of hardware failure modes; what you learn from an improved understanding of hardware failure modes improves your understanding of the diagnostics.
The goal of the current generation of ExtremeWare diagnostics is to help users achieve the highest levels of network availability and performance by providing a suite of diagnostic tests that moves away from a reactive stance—wherein a problem occurs and then you attempt to determine what caused the problem—to a proactive state—wherein the system hardware, sof tware , an d dia gno stic s wor k tog eth er to reduce the total number of failures and downtime through:
More accurate reporting of errors (fewer false notifications; more information about actual errors)
Early detection of conditions that lead to a failure (so that corrective action can be taken before the
failure occurs)
Automatic detection and correction of packet memory errors in the system’s control and data planes Administrators will now find a greatly reduced MTTR (mean time to repair) due to fast and accurate
fault identification. Multiple modules will no longer need to be removed and tested; faulty components will usually be identified directly. Over time, there should be a significant reduction in the number of problems found.
NOTE
In spite of the improved ExtremeWare hardware diagnostics, some network events might still occur, because software is incapable of detecting and preventing every kind of failure.
12 Advanced System Diagnostics and Troubleshooting Guide
Overview of the ExtremeWare Diagnostics Suite
Overview of the ExtremeWare Diagnostics Suite
The ExtremeWare diagnostic suite includes the following types of tools for use in detecting, isolating, and treating faults in a switch. Each of these diagnostic types is summarized below, but is described in greater detail in later sections of this guide.
Power-on self test (POST)—A sequence of hardware tests that run automatically each time the switch
is booted, to validate basic system integrity. The POST runs in either of two modes: normal (more thorough, but longer-running test sequence) or FastPOST (faster-running basic test sequence).
On-demand system hardware diagnostics—Run on demand through user CLI commands; runs in
either of two modes: normal (faster-running basic test sequence) or extended (more thorough, but longer-running test sequence).
The extended diagnostics include the packet memory scan, which checks the packet memory area of the switch fabric for defects and maps out defective blocks. This test can be run by itself, as part of the slot-based extended diagnostics, or can be invoked from within the system health checks.
Switch-wide communication-path packet error health checks—An integrated diagnostic
subsystem—the system health check feature—that consists of a number of different test types, operating proactively in the background to detect and respond to packet error problems in module memory or on communication paths. The system health checks include the following kinds of tests:
Backplane health checksCPU health checksSwitch fabric checksum validationDynamic memory scanning and memory mappingTransceiver diagnosticsForwarding database (FDB) scan
Suppor ted Hardware
The ExtremeWare diagnostic suite applies only to Extreme Networks switch products based on the “inferno” series chipset. Equipment based on this chipset are referred to as being “inferno” series or “i” series products: the BlackDiamond family of core chassis switches (6804, 6808, and 6816), the Alpine systems (3802, 3804, 3808), and the Summit “i”-series stackables (Summit1i, Summit5i, Summit7i, Summit 48i, and Summit48Si).
Applicable ExtremeWare Versions
The information in this guide is based on the features and feature attributes found in ExtremeWare Version 6.2.2b134 or later.
Advanced System Diagnostics and Troubleshooting Guide 13
Introduction
14 Advanced System Diagnostics and Troubleshooting Guide
2 Hardware Architecture
This chapter provides a brief summary of the switch hardware features most relevant to understanding the use of the Extreme Networks diagnostic suite.
This chapter contains the following sections:
Diagnostics Support on page 15
The BlackDiamond Systems on page 16
Alpine Systems on page 22
Summit Systems on page 23
Diagnostics Suppor t
The ExtremeWare diagnostic suite applies only to Extreme Networks switch products based on the “inferno” series chipset. Equipment based on this chipset are referred to as being “inferno” series or “i” series products: the BlackDiamond family of core chassis switches (6804, 6808, and 6816), the Alpine systems (3802, 3804, 3808), and the Summit “i”-series stackables (Summit1i, Summit5i, Summit7i, Summit 48i, and Summit48Si).
NOTE
These switches and the switch modules use naming conventions ending with an “i” to identify them as “inferno” series or “i” series products. For the most current list of products supporting the “i” chipset, such as the MSM-3 and other “3”-series modules, such as the G16X
Unless otherwise specified, a feature requiring the “i” chipset requires the use of both an “i” chipset-based management module, such as the MSM64i, and an “i” chipset-based I/O module, such as the G8Xi.
3
, consult your release notes.
Advanced System Diagnostics and Troubleshooting Guide 15
Hardware Architecture
The BlackDiamond Systems
In the context of the system diagnostics, the BlackDiamond family of core chassis switches share the same fundamental hardware architecture: a multislot modular chassis containing a passive backplane that supports redundant load-sharing, hot-swappable switch fabric modules. On BlackDiamond systems, each I/O module and MSM represents an individual switch containing its own switching fabric and packet memory.
Figure 1: BlackDiamond 6800 Series architecture, general block diagram
I/O Module MSM
OTP
PQM
OTP
PQM
PHY
PHY
PHY
PHY
MAC
MAC
(Quake)
Data Bus
(Twister)
AFQM
ASIC
FDB
PM
SE
ASIC
VPST
MAC
Data Bus
MAC
PHY
PHY
PHY
PHY
Management Bus
PHY
Backplane
PHY
PHY
PHY
NVRAM
E-NET
Console
PCMCIA
MAC
MAC
PCI Bus
(Quake)
Data Bus
(Twister)
Subassembly
AFQM
ASIC
FDB
PM
SE
ASIC
CPU
VPST
Data Bus
BlackDiamond 6800 Series Hardware Architecture Differences
In the context of understanding the ExtremeWare diagnostics and their use in troubleshooting system problems, these are the key hardware distinctions between the BlackDiamond 6816, 6808, and 6804.
BlackDiamond 6816—Modular chassis with passive backplane; sixteen chassis slots for I/O modules;
four chassis slots for MSMs.
BlackDiamond 6808—Modular chassis with passive backplane; eight chassis slots for I/O modules;
two chassis slots for MSMs.
BlackDiamond 6804—Modular chassis with passive backplane; four chassis slots for I/O modules;
two chassis slots for MSMs.
16 Advanced System Diagnostics and Troubleshooting Guide
DN_024B
A
The BlackDiamond Systems
The BlackDiamond Backplane
The BlackDiamond backplane is a passive backplane, meaning that all the active components such as CPUs, ASICs, and memory have been moved onto plug-in modules, such as the I/O modules and MSMs.
Figure 2: BlackDiamond passive backplane architecture (BlackDiamond 6808 shown)
Switch
1
Module
Switch
2
Module
64 Gbps
Switch
Fabric
Eight Load-Shared Gigabit Links
Switch
3
Module
64 Gbps
Switch
Fabric
Switch
8
Module
A
Fault-Tolerant Switch Fabric and System Management
B
DN_001
The BlackDiamond backplane provides inter-slot electrical connections for both network data traffic and a separate control bus for switch fabric management. Data traffic is carried on four AUI links between each MSM and each I/O slot on BlackDiamond 6804 and BlackDia mond 6808 systems, and on two AUI links between each MSM and each I/O slot on BlackDiamond 6816 systems. Device management occurs on a 32-bit PCI bus connecting MSMs and I/O modules. The number of backplane slots for I/O modules and MSMs determines the BlackDiamond system type (6804, 6808, 6816).
The chief advantages of a passive backplane are:
The absence of active components yields a much lower possibility of backplane failure.
You can remove and replace system modules faster, making upgrades and repairs easier, faster, and
cheaper.
NOTE
One disadvantage of a passive backplane is that a problem on one switch module might cause other switch modules to fail. More information on this possibility is covered in later chapters of this guide.
Advanced System Diagnostics and Troubleshooting Guide 17
Hardware Architecture
BlackDiamond I/O Modules
Each BlackDiamond I/O module has a built-in switching fabric (see Figure 3) giving the module the capability to switch local traffic on the same module. Traffic that is destined for other modules in the chassis travels across the backplane to the MSMs, where it is switched and sent to its destination I/O module.
Figure 3: BlackDiamond I/O module architecture (G8Xi 32 Gb fabric shown)
G8Xi (32 Gb Fabric)
PHY
MAC
MAC
PHY
PBUS
SE
ASIC
SE
ASIC
SE
ASIC
SE
ASIC
SRAM (Packet Mem + FDB)
AFQM
ASIC
OTP RAM
PQ RAM
VPST RAM
MAC
MAC
4x GbE to MSM-A
4x GbE to MSM-B
DN_026B
Each BlackDiamond I/O module has eight load-shared Gigabit Ethernet links to both MSMs through the backplane. The load sharing algorithm distributes traffic across different channels through the backplane’s Gigabit Ethernet links, providing bi-directional communication.
Each BlackDiamond I/O module is equipped with the following kinds of hardware components:
PHY: An industry-standard ASIC responsible for physical layer (layer 1) signal, clocking, etc.
MAC: The MAC handles the standard MAC layer functions as well as some other functions to
prepare a packet for transmission to the switch fabric or to the ext ernal network, including 802.1p and DiffServ examination, VLAN insertion, MAC substitution, TTL decrement, 802.1p and DiffServ replacement, etc.
Each I/O module has both external MACs and internal MACs. External MACs handle the interface to the external ports; internal MACs handle the interface to the BlackDiamond backplane. Each MSM provides four Gigabit Ethernet links to each I/O module.
PBUS: The packet data bus that transfers packets between the MAC and the packet memory.
Switch engine (distributed packet processor) ASIC (Twist er) and its associated memories: packet
RAM and FDB RAM. The SE ASIC impl ements a high-speed, parallel data transfer bus for transferring packets from MACs to packet memory and back.
Address filtering and queue management ASIC (Quake) and its associated memories: OTP RAM, PQ
RAM, and VPST RAM.
When a data packet is received by the PHY, the PHY passes the packet to the MAC. The MAC handles the layer 2 tasks, such as tagging and the MAC address, then transfers the packet across the PBUS to the
18 Advanced System Diagnostics and Troubleshooting Guide
The BlackDiamond Systems
packet memory for temporary storage. Based on the information in memory, such as the FDB, the address filtering and queue management ASIC makes a forwarding decision. If the next hop is a local port (on the same module), the packet is forwarded to the external MAC and PHY for the exit port. If the packet is destined for another module (as either slow path traffic or fast path traffic), the packet is transferred to the internal MAC and then on to the MSM (CPU).
All I/O modules share the management bus on the backplane, and use it to communicate to each other and to the MSMs.
Management Switch Modules
As its name indicates, the Management Switch Fabric Module ( MSM) serves a dual role in the system: it is equipped to act as the internal switch fabric for data that is being transferred between I/O modules in the chassis, and to handle the upper-layer processing and system management functions for the switch. (See Figure 4.)
Figure 4: BlackDiamond MSM architecture (MSM64i shown)
CPU
Subsystem
CPU
256 MB
DRAM
Flash
NVRAM
PCMCI
Slot
CPU
AFQM
ASIC
OTP RAM
PQ RAM
VPST RAM
AFQM
ASIC
OTP RAM
PQ RAM
VPST RAM
MSM64i (64 Gb Fabric)
SE
ASIC
SE
ASIC
SRAM (Packet Mem + FDB)
SE
ASIC
SE
ASIC
SE
ASIC
SE
ASIC
SE
ASIC
SE
ASIC
PBUS
MAC
MAC
MAC
MAC
MAC
MAC
MAC
MAC
DN_027B
An MSM is equipped with the following hardware components: CPU subsystem (dual CPUs, DRAM, NVRAM, flash memory, and PCMCIA slot) and switch fabric subsystem (Quake ASICs, OTP, PQ, and VPST RAM, packet memory and FDB SRAM, Twister ASICs, PBUS, and MAC ASICs).
Advanced System Diagnostics and Troubleshooting Guide 19
Hardware Architecture
BlackDiamond MSM Redundancy
The CPU subsystems on a pair of BlackDiamond MSMs operate in a mast er-slave relationship. (See Figure 5.)
Figure 5: BlackDiamond MSM redundancy scheme
Fault Tolerant Switch Fabric
and
System
Management
MSM64i (A)
CPU Sub-
system
MSM64i (B)
CPU Sub-
system
CPU Control Path Data Path
Switching
Fabric
Switching
Fabric
Backplane
I/O Module
I/O Module
I/O Module
I/O Module
I/O Module
I/O Module
I/O Module
I/O Module
DN_028A
The master MSM CPU subsystem actively manages the switch and the t ask of switching packets in the CPU control (or management) path. The slave MSM CPU subsystem is in standby mode, but is checked periodically by the master MSM CPU (via EDP) to determine wh ether it is still available.
The master MSM also guarantees that management operations occur in a synchronized manner. For example, if you make a configuration change and need to save it, the master MSM ensures that the configuration is saved to both the master MSM and the slave MSM at the same time. That way, the updated slave is ready to take over as the master if the master MSM fails.
Despite the master-slave switch management role, the switch fabrics on both MSMs actively switch core traffic in a load-shared fashion. Load-sharing switches core traffic from different I/O modules.
All MSMs share the control (or management) bus on the backplane, and use it to communicate to each other and to installed I/O modules to perform system health checks and status polling.
Causes of MSM Failover and System Behavior
A number of events can cause an MSM failover to occur, including:
Software exception; system watchdog timer expiry
Diagnostic failure (extended diagnostics, transceiver check/scan, FDB scan failure/remap)
Hot removal of the master MSM or hard-reset of the master MSM
20 Advanced System Diagnostics and Troubleshooting Guide
The BlackDiamond Systems
The MSM failover behavior depends on the following factors:
Platform type and equippage (Summit vs. Alpine vs. BlackDiamond)
Software configuration settings for the software exception handling options such as system
watchdog, system recovery level, and reboot loop protection. (For more information on the configuration settings, see Chapter 4, “Software Exception Handling.”)
In normal operation, the master MSM continuously resets the watchdog timer. If the watchdog timer expires, the slave MSM will either 1) reboot the chassis and take over as the master MSM (when the switch is equipped with MSM-64i modules), or 2) initiate a hitless failover (when the switch is equipped with MSM-3 modules). The watchdog is a software watchdog timer that can be enabled or disabled through CLI commands. The watchdog timer is reset as long as ExtremeWare is functioning well enough to return to the main software exception handling loop where the critical software exception handling tasks, such as tBGTask, handle the process of resetting the watchdog timer and creating log entries.
Software configuration settings for the system health check feature, or for any of the diagnostic tests
that you might choose to run manually. For example, in the context of memory scanning and mapping, Chapter 5, “Diagnostics,” contains
three tables that describe the behavior of the switch for different platform types and diagnostics configuration:
Table 5: Auto-recovery memory scanning and memory mapping behaviorTable 6: Manual diagnostics memory scanning and memory mapping behavior, normalTable 7: Manual diagnostics memory scanning and memory mapping behavior, extended
NOTE
On switches equipped with MSM64i modules, you should per iodically use the synchronize command to ensure that the slave MSM and master MSM are using matched images and configurations. If not synchronized, the slave MSM might attempt to use the image it has loaded in conjunction with the configuration from the master MSM, a mismatch that will most likely cause the switch to behave differently after an MSM failover, thereby defeating the intended purpose of redundant peer MSMs.
If you need to insert a new MSM, you can duplicate the contents of the NVRAM and flash memory from an existing MSM to the newly-installed MSM using one CLI synchronization command.
NOTE
The MSM-3 uses new technology to provide “hitless” failover, meaning the MSM-3 transitions through a failover with no traffic loss and no switch downtime, while it maintains active links and preserves layer 2 state tables. Contrast this performance to normal failover with MSM64i modules, which can take the switch down for approximately 30 seconds. The MSM-3 makes hitless upgrades possible. It is supported in ExtremeWare release 7.1.1 and later.
Advanced System Diagnostics and Troubleshooting Guide 21
Hardware Architecture
Alpine Systems
Like the BlackDiamond systems, the Alpine systems are also based on a multislot modular chassis that uses the inferno chipset, but the Alpine switches differ from the BlackDiamond switches on these points (see Figure 6):
• Active backplane—Alpine switches use an active backplane that uses the same basic set of ASICs
(the switch engine ASIC and the address filtering and queue management ASIC) and memory (packet memory for storing packets; OTP RAM, PQ RAM, and VPST RAM) that are used on the BlackDiamond MSMs and I/O modules, so it offers wire-speed switching.
But unlike the BlackDiamond MSM, the Alpine backplane has no CPU and no MAC. It does provide PBUS links to all I/O modules. The number of backplane slots for I/O modules determines the Alpine system type (3802, 3804, 3808).
• SMMi processor module instead on MSM—The SMMi processor module is similar to the CPU
subsystem of the BlackDiamond MSM in that it is equipped with the following hardware components: CPU subsystem, DRAM, NVRAM, and flash memory, console port connectors, management interface, and a PCMCIA slot. But unlike the MSM, the SMMi contains no switching fabric.
• I/O modules provide PHY and MAC functionality, but no onboard switching fabric—Each
standard I/O module has two PBUS links to the switching fabric on the Alpine backplane.
Figure 6: Alpine architecture (Alpine 3804 shown)
PHY
PHY
PHY
PHY
WAN Subsystem
CPU DRAM
2x 10/100base-TX
GM-4Xi
MACPHY
FM-32Ti
MAC
MAC
WM-4T1
PBUS
Alpine 3804 - (32 Gb Fabric)
SE
ASIC
SE
ASIC
SE
ASIC
SE
ASIC
CPU Subsystem
Flash
NVRAM
SMMi
SRAM (Packet Mem + FDB)
256 MB
DRAM
AFQM
ASIC
OTP RAM
PQ RAM
VPST RAM
CPU
CPU
DN_029B
22 Advanced System Diagnostics and Troubleshooting Guide
Summit Systems
Summit Systems
Unlike the BlackDiamond and Alpine systems, the Summit “i”-series stackables are not modular systems: all of the system components are integrated into one unit. (See Figure 7.)
Figure 7: Summit architecture
PHY
PHY
PHY
PHY
PHY
PHY
PHY
PHY
MACPHY
MAC
MAC
PBUS
SE
ASIC
SE
ASIC
SE
ASIC
SE
ASIC
Flash
NVRAM
SRAM (Packet Mem + FDB)
256 MB
DRAM
AFQM
ASIC
OTP RAM
PQ RAM
VPST RAM
CPU
CPU
DN_030A
The Summit CPU subsystem is similar to the CPU subsystem of the BlackDiamond MSM and the Alpine SMMi in that it is equipped with the same basic hardware components: dual CPUs, memory (DRAM, NVRAM, and flash memory), console port connectors, management interface, and a PCMCIA slot.
The Summit switching fabric subsystem uses the same basic set of ASICs (the switch engine ASIC and the address filtering and queue management ASIC) and memory (packet memory for storing packets; OTP RAM, PQ RAM, and VPST RAM) that are used on the BlackDiamond and Alpine switches, so it, too, offers wire-speed switching.
The Summit I/O subsystem provides PHY and MAC functionality in a variety of port config urations (types of ports and numbers of ports).
Advanced System Diagnostics and Troubleshooting Guide 23
Hardware Architecture
24 Advanced System Diagnostics and Troubleshooting Guide
3 Packet Errors and Packet Error
Detection
This chapter describes some of the factors that might result in packet errors in the switch fabric and the kinds of protection mechanisms that are applied to ensure that packet error events are minimized and handled appropriately.
This chapter contains the following sections:
Overview on page 25
Definition of Terms on page 26
Standard Ethernet Detection for Packet Errors on the Wire on page 27
Extreme Networks’ Complementary Detection of Packet Errors Between Wires on page 27
Failure Modes on page 30
Error Messages for Fabric Checksums on page 31
Panic/Action Error Messages on page 35
Overview
Complex, wire-speed switch fabrics are subject to electronic anomalies that might result in packet errors. The Ethernet standard contains built-in protections to detect packet errors on the link between devices, but these mechanisms cannot always detect packet errors occurring in the switch fabric of a device. Extreme Networks has incorporated many protection mechanisms to ensure that packet error events are minimized and handled properly.
These protection mechanisms include the following:
Ethernet CRC (detects packet errors between switches)
Switch fabric checksums (automatic detection of live packet errors)
Packet memory scanning (offline detection of packet memory errors)
System health check (automatic test of various CPU and data paths)
FDB scan (background scan process scans entire FDB RAM pool on all switch fabrics)
Transceiver check (background detects packet errors on internal control paths)
Advanced System Diagnostics and Troubleshooting Guide 25
Packet Errors and Packet Error Detection
Definition of Ter ms
To establish a basis for the descriptions in this chapter, Table 3 lists and defines terms that are used repeatedly throughout this chapter and those that follow. When any of these terms are used for their precise meaning, they are shown emphasi zed in bold type.
Table 3: Data Error Terms
Term Description
Packet error event When the contents of a network data or control packet are modified by the
transmission medium or a network device in a way that is not indicated by the rules of standard network behavior such that the contents of the packet will be considered invalid by upper layer protocols or applications, we say that a packet error event has occurred.
Note that the term applies only to packet changes initiated by layer 1 interaction; that is, if an error in the electrical or optical processing of the bit-level data in the packet results in a change to the packet, we consider this a packet error event.
The term does not extend to systematic software or hardware errors that result in valid but incorrect changes to the packet at higher OSI layers, such as inserti ng the wrong next-hop MAC destination address into the packet head er because of an erroneous entry in the hardware forwarding database.
Checksum A value computed by running actual packet data through a polynomial formula.
Checksums are one of the tools used by Extreme Networks in attempts to detect and manage packet error events.
Packet checksum A checksum value that is computed by the MAC chip when the packet is transferred
from the MAC chip to the switch fabric. This checksum value precedes the packet as it transits the switch fabric.
Verification checksum A checksum value that is computed by the MAC chip when the packet is transferred
Checksum error When a packet exits the switch fabric, the packet checksum that follows the packet
System health check A series of system tests and associated reporting mechanisms that are used to notify
System health check error This term refers to error messages in the system log that are generated by the
Transient errors Errors that occur as one-time events during normal system processing. These types
Soft-state errors These types of error events are characterized by a prolonged period of reported
Permanent errors These type s of errors result from permanent hardware defects that might, or might
from the switch fabric to the MAC chip for transmission.
must match the verification checksum computed a s the packet leaves the switch fabric. If the checksums do not match, then a checksum error results.
network operators of potential system problems and to isolate and diagnose faulty components when problems occur. The checksum error reporting mechanism is a part of the system health check system.
system health check system. Error messages generated by the system health check system are prefaced by the text string “Sys-health-check.” Checksum error
messages are a subset of the system health check error messages.
of errors will occur as single events, or might recur for short durations, but do not have a noticeable impact on network functionality and require no user interventi on to correct.
error messages and might, or might not, be accompanied by noticeable de gradation of network service. These events require user intervention to correct, but are resolved without replacing hardware.
Error messages of this type are the result of software or hardware systems entering an abnormal operating state in which normal switch operation mi ght, or might not, be impaired.
not, affect normal switch operation. They cannot be resolved by user interventi on and will not resolve themselves. You must replace hardware to resolve permanent errors.
26 Advanced System Diagnostics and Troubleshooting Guide
Table 3: Data Error Terms (continued)
Term Description
Fast path This term refers to the data path for a packet that traverses a switch and does not
require processing by the CPU. Fast path packets are handled entirely by ASICs and are forwarded at wire rate.
Slow path This term refers to the data path for packets that must be processed by the switch
CPU, whether they are generated by the CPU, removed from the network by the CPU, or simply forwarded by the CPU.
Standard Ethernet Detection for Packet Errors on theWire
Standard Ethernet Detection for Packet Errors on the Wire
Data transiting from one switch to another is checked for packet errors using the Ethernet Cyclic Redundancy Check (CRC) built into the IEEE 802.3 specification.
As the sending switch assembles a frame, it performs a CRC calculation on the bits in that frame and stores the results of that calculation in the frame check sequence field of the frame. At the receiving end, the switch performs an identical CRC calculation and compares the result to the value stored in the frame check sequence field of the frame. If the two values do not match, the receiving switch assumes that packet data has been illegally modified between CRC calculation and CRC validation and discards the packet, and increments the CRC error counter in the MAC device associated with that port. In Extreme Networks devices, ExtremeWare polls the MAC CRC error count registers and makes that information available through the output of the
show port rxerrors CLI command.
Extreme Networks’ Complementary Detection of Packet Errors Between Wires
The 802.3 Ethernet specification provides a CRC check for validation of data on the wire, but offers no guidance for handling data validation in the devices between the wires. Because these devices are far more complicated than the wires connected to them, common sense indicates the requirement for some mechanism for checking internal data integrity. To complement the Ethernet CRC data validation scheme, Extreme Networks switches implement an internal data checksum validation scheme referred to as the fabric checksum.
The switch fabric in a switch is essentially an extremely high-speed data path connecting multiple ports and using a set of programmable lookup tables to make intelligent forwarding decisions when moving data from point to point inside the switch. Figure 8 uses a generalized block diagram of a switch to illustrate data movement within a switch.
Advanced System Diagnostics and Troubleshooting Guide 27
Packet Errors and Packet Error Detection
Figure 8: Generalized switch block diagram
4
5
3 6
ASIC ASIC
DMAC
2
1
1 PHY port and MAC device layer 4 Control bus 2 Packet bus (PBUS) 5 Packet memory 3 Forwarding ASICs 6 CPU subsystem
MAC MAC
CPU
sub-system
CG_002B
The following sections describe the hardware and software components that work together to detect and manage packet error incidents within the Extreme Networks switch.
Hardware System Detection Mechanisms
All Extreme Networks switches based on the “i”-series switch fabric validate data integrity internal to the switch fabric using a common checksum verification algorithm. Using Figure 8 as a generalized model, when a packet is received at an Ethernet network interface, the receiving MAC ASIC verifies the Ethernet CRC: it computes a CRC value by applying the same algorithm used to compute the CRC value appended to the received packet data by the transmitting switch. If the algorithm and the data it is applied to are the same on both ends of the Ethernet link, the CRC values should match. If they do not, the packet is assumed to have been damaged and is discarded.
If the CRC values match, the MAC ASIC must then transfer the packet to the internal switch fabric. Before doing this, however, it produces a checksum value based on the packet data being passed to the switch fabric. This checksum value becomes the packet checksum. It is prepended to the packet and both the packet checksum and packet are passed on to the switch fabric.
After the switch fabric is finished processing the packet and has made a decision regarding where the packet is to be forwarded, it passes the packet to the transmitting MAC ASIC. The transmitting MAC ASIC performs the reverse of the process performed by the receiving MAC ASIC. It first computes a checksum value based on the packet data received from the switch fabric. We will call this value the verification checksum.
The transmitting MAC ASIC then compares the verification checksum against the packet checksum. If the two values do not match, the result is a checksum error. The MAC ASIC maintains a count of every checksum error that occurs on every port. When a packet is found to have a checksum error, it is still
28 Advanced System Diagnostics and Troubleshooting Guide
Extreme Networks’ Complementary Detection of Packet Errors Between Wires
transmitted, but an invalid CRC value is included with the packet. Therefore, the receiving device will detect an invalid CRC value and will drop the packet.
In Summit stackable switches, the packet checksum is calculated by the MAC ASIC on the receiving port and is compared against the verification checksum calculated by the MAC ASIC on the transmitting port, as described above.
In Alpine 3800 series switches, the packet checksum is calculated by the MAC ASIC on the receiving port on the I/O module on which the packet is received. The packet checksum and packet are passed to the switch fabric, which is on the Alpine switch backplane, and then from the switch fabric to the transmitting MAC ASIC on the I/O module on which the packet is to be transmitted. There, the verification checksum is computed and compared against the packet checksum.
In BlackDiamond 6800 series switches, the packet checksum is computed by the MAC ASIC on the receiving port on the I/O module on which the packet is received. The packet checksum and the packet traverse the switch fabric on the I/O module and are handed off to either an external MAC ASIC, connected to a network port, or to an internal MAC ASIC, connected to a BlackDiamond backplane link.
In either case, the behavior of the MAC ASIC is the same: it computes the verification checksum and compares it against the packet checksum to detect any changes in packet data. Similarly, whether the packet is transmitted out the external port to the network, or out the internal port to the BlackDiamond backplane, the packet is accompanied by an Ethernet CRC.
The behavior of the BlackDiamond MSM module is identical to that of the BlackDiamond I/O module, except that all MAC ASICs on the MSM are internal (not to network ports). Regardless, the behavior of the receiving and transmitting MAC ASICs is the same for packets traversing an MSM module as for packets traversing an I/O module.
Thus far, all of the systems described have been involved in fast-path forwarding. Therefore, any checksum errors detected using the mechanisms described above are referred to as fast-path checksum errors.
Packets that must be processed by the switch CPU are also validated by checksum values. When a packet is received, it might be destined specifically for the CPU (as in the case of protocol packets) or it might be passed to the CPU for assistance in making a forwarding decision (if the switch fabric lacks the information required to forward the packet correctly). In either case, the receiving MAC ASIC still computes and prepends a packet checksum just as it does for fast-path packets, but because the packet is not passed to a transmitting MAC ASIC before it is forwarded, the switch fabric itself is responsible for computing the verification checksum and comparing it against the packet checksum. If a mismatch is found, the switch fabric reports the checksum error condition to the CPU as it passes the packet up to the CPU. These types of checksum errors are one instance of a class of checksum errors known as slow-path checksum errors.
Software System Detection Mechanisms
As described above, each MAC ASIC maintains a port-by-port count of every checksum error detected. ExtremeWare contains mechanisms that can retrieve the checksum error counts from the MAC ASICs in the switch and act on it. Current versions of ExtremeWare retrieve the checksum error counts from all MAC ASICs in the switch at twenty-second intervals. The counts at the end of the twenty-second interval are compared with the counts at the beginning of the twenty-second interval on a port-by-port basis. If, for any given port, the count is found to be different, then ExtremeWare is said to have detected a checksum error. Depending on the ExtremeWare version, the configuration settings, the frequency and count of checksum errors, and a variety of other factors, ExtremeWare will initiate one of several actions, described in the section “System (CPU and Backplane) Health Check” on page 69. For
Advanced System Diagnostics and Troubleshooting Guide 29
Packet Errors and Packet Error Detection
example, the system health check facility can be configured such that ExtremeWare will enter a message into the system log that a checksum error has been detected.
Failure Modes
Although packet errors are extremely rare events, packet errors can occur anywhere along the data path, along the control path, or while stored in packet memory. A checksum mismatch might occur due to a fault occurring in any of the components between the ingress and egress points—including, but not limited to, the packet memory (SRAM), ASICs, MAC, or bus transceiver components.
There are many causes and conditions that can lead to packet error events. These causes and conditions can fall into one of these categories:
Transient errors
Systematic errorsSoft-state errorsPermanent errors
The failure modes that can result in the above categories are described in the sections that follow.
Transient Failures
Transient failures are errors that occur as one-time events during normal system processing. These types of errors will occur as single events, or might recur for short durations. Because these transient events usually occur randomly throughout the network, there is usually no single locus of packet errors. They are temporary (do not persist), do not have a noticeable impact on network functionality, and require no user intervention to correct: There is no need to swap a hardware module or other equipment.
Systematic Failures
Systematic errors are repeatable events: some hardware device or component is malfunctioning in such a way that it persistently exhibits incorrect behavior. In the context of the ExtremeWare Advanced System Diagnostics, the appearance of a checksum error message in the system log—for example—indicates that the normal error detection mechanisms in the switch have detected that the data in a packet has been modified inappropriately. While checksums provide a strong check of data integrity, they must be qualified according to their risk to the system and by what you can do to resolve the problem.
Systematic errors can be subdivided into two subgroups:
Soft-state failures
Permanent, or hard failures
Soft-State Failures
These types of error events are characterized by a prolonged period of reported error messages and might, or might not, be accompanied by noticeable degradation of network service. These events require user intervention to correct, but are resol ved w ith out re plac ing hardw are.
Failures of this type are the result of software or hardware systems entering an abnormal operating state in which normal switch operation might, or might not, be impaired.
30 Advanced System Diagnostics and Troubleshooting Guide
Loading...
+ 82 hidden pages