ATTO Technology Diamond Array V Installation And Operation Manual

Diamond Storage Array
V-Class
Installation and Operation Manual
© 2005 ATTO Technology Inc. All rights reserved. All brand or product names are trademarks of their respective holders. No part of this manual may be reproduced in any form or by any means without the express written permission of ATTO Technology Inc.
2/05
6.4 PRMA-0338-000
Contents
Preface
1.0 Product Overview ...........................................................................1
Features Fibre Channel model SCSI model
2.0 Technical Overview ........................................................................3
ATA disk drives ADXT Powers ATA to New Levels
3.0 Installation Instructions .................................................................5
Three major steps are required to set up the Diamond Storage Array Step 1: physical setup Step 2a: set up the Ethernet connection Step 2b: connect to Ethernet If the Diamond Storage Array is attached to a DHCP server If the Diamond Storage Array is not attached to a DHCP server and you wish to change the defaults Step 3: configure drives
3.1 Components .........................................................................7
Floor model Rack mount
3.2 Physical Setup ......................................................................9
Floor Model Rack Mount General Instructions
3.2.1 Connecting a Fibre Channel Array .............................11
Autoconfiguration Manual configuration
3.2.2 Connecting a SCSI Array ............................................13
To connect the SCSI Diamond Storage Array
3.3 Determining Drive and Sled Designations .........................15
Numbering conventions Physical numbering Logical Numbering Examples Unique serial number for each LUN
ATTO Technology Inc.
4.0 Accessing the Array .......................................................................17
Command Line Interface ATTO ExpressNAV In-band SCSI over Fibre Channel RS-232 port Ethernet port SNMP
4.1 In-band CLI Over Fibre Channel .........................................19
I/O details
4.2 Serial Port Access ................................................................21
4.3 Ethernet Access: Telnet and SNMP Protocols ..................23
To connect to the Ethernet port To use Telnet To use SNMP
4.4 ATTO ExpressNAV: Browser-based Interface ...................25
Browser Compatibility To optimize ExpressNAV in Internet Explorer To open an ExpressNAV session To navigate ExpressNAV
4.4.1 ExpressNAV Pages ......................................................27
Status Ethernet SNMP Serial Port Fibre Channel Storage Management RAID Clear Data Logical Units Partitions Zoning Rebuild Configuration Advanced To use the Advanced Page CLI commands Restart Help
4.5 CLI: Interface via ASCII-based Commands .......................29
4.5.1 Summary of CLI Commands .......................................31
Diamond Storage Array Installation and Operation Manual
4.5.2 General Use Commands .............................................34
FirmwareRestart Help PartitionCommit SaveConfiguration SystemSN VerboseMode ZoneCommit
4.5.3 Fibre Channel Configuration Commands ..................35
FcConnMode FcDataRate FcFairArb FcFrameLength FcFullDuplex FcHard FcHardAddress FcPortInfo FcPortList FcSCSIBusyStatus FcWWName
4.5.4 Serial Port Configuration Commands ........................37
SerialPortBaudRate SerialPortEcho SerialPortHandshake SerialPortStopBits
4.5.5 Ethernet Commands ....................................................39
EthernetSpeed FTPPassword IPAddress IPDHCP IPGateway IPSubnetMask SNMPTrapAddress SNMPTraps SNMPUpdates Te ln e t Pa s sw or d TelnetTimeout TelnetUsername
4.5.6 Diagnostic Commands ................................................41
AudibleAlarm DiamondModel DiamondName DriveCopyStatus DriveInfo FcNodeName FcPortList FcPortName Help IdentifyDiamond Info LUNInfo PartitionInfo
ATTO Technology Inc.
RAID5ClearDataStatus RAIDRebuildStatus SerialNumber SledFaultLED SMARTData Temperature VirtualDriveInfo ZoneInfo
4.5.7 Drive Configuration Commands .................................43
ATA Di sk St a t e AutoRebuild ClearDiskReservedAreaData DriveCopy DriveCopyHalt DriveCopyResume DriveCopyStatus DriveInfo DriveSledPower DriveWipe IdeTransferRate LUNInfo LUNState PartitionCommit PartitionInfo PartitionMerge PartitionSplit QuickRAID0 QuickRAID1 QuickRAID5 QuickRAID10 RAID5ClearData RAID5ClearDataStatus RAIDInterleave RAIDHaltRebuild RAIDManualRebuild RAIDRebuildState RAIDRebuildStatus RAIDResumeRebuild RebuildPriority ResolveLUNConflicts RestoreModePages SledFaultLED VirtualDriveInfo ZoneAddDevice ZoneAddHost ZoneAddPort ZoneClearAll ZoneCommit ZoneCreate ZoneDelete ZoneInfo ZoneRemoveDevice ZoneRemoveHost ZoneRemovePort ZoneRetrieve ZoneState
Diamond Storage Array Installation and Operation Manual
4.5.8 Maintenance Services Commands .............................48
FcScsiBusyStatus FirmwareRestart MaxEnclTempAlrm MinEnclTempAlrm Temperature Zmodem ZoneRetrieve
5.0 Configuring Drives ..........................................................................49
JBOD (Just a Bunch of Disks) RAID Level 0 RAID Level 1 RAID Level 10 RAID Level 5 Interleave Partitions Zones Hot Spare sleds Enhancing performance
5.1 JBOD .....................................................................................51
To set up the JBOD configuration
5.2 RAID Level 0 .........................................................................52
Sled-based versus disk-based To set up RAID Level 0 groups To remove RAID Level 0 groups from the array
5.3 RAID Level 1 .........................................................................55
To set up RAID Level 1 groups To set up RAID Level 1 with Hot Spare sleds To remove RAID groups
5.4 RAID Level 5 .........................................................................57
Configuring a fully-populated array To set up one RAID Level 5 group with one Hot Spare sled To set up two RAID Level 5 groups with two Hot Spare sleds Configuring a partially-populated array To set up one RAID Level 5 group with one Hot Spare sled Removing RAID groups
5.5 RAID Level 10 .......................................................................61
To set up RAID Level 10 groups To remove RAID groups To set up RAID Level 10 with Hot Spare sleds:
5.6 Rebuilding RAID Level Configurations ..............................63
To reset LUN status To synchronize mirrored drives automatically Rebuild priority To synchronize mirrored drives manually
ATTO Technology Inc.
5.7 RAID Interleave .....................................................................65
To change the RAID Interleave parameter
5.8 Creating Partitions ...............................................................67
To create a partition To merge partitions
5.9 Creating Zones .....................................................................69
Principles of Zoning Factors to consider Status and Sense Data Configuring Zones To create a zone To remove zones To change current zones Other operations Errors
6.0 Copying Drives ................................................................................73
7.0 Updating Firmware .........................................................................75
Updating firmware via the RS-232 serial port Updating firmware via the optional Ethernet card
8.0 System Monitoring and Reporting ................................................77
8.1 Troubleshooting ...................................................................81
Windows 2000 special instructions Error Messages System Fault LED Command Line Interface messages ERROR. Wrong/Missing Parameters ERROR. Invalid Command. Type ‘help’ for command list ERROR. Command Not Processed. Audible Alarm Specific situations and suggestions If a drive fails to respond If a power supply fails To determine if the problem exists with the Host Interface Card or the connection If you can’t access the array CLI via Ethernet If you do not see the appropriate number of LUNs on the host machine
8.2 Resetting Defaults ................................................................83
Default Return to Default settings Factory Default To reset to Factory Defaults, firmware version 2.5.3 or higher
9.0 Hardware Maintenance ...................................................................85
Diamond Storage Array Installation and Operation Manual
9.1 Hot Swap Operating Instructions .......................................87
Disk Drives The following method is the safest way to perform a hot swap of a drive Power Supplies Blower Assemblies To replace a blower assembly
9.2 Optional Hot Spare Sled ......................................................89
To set up RAID Level 1 with Hot Spare sleds To set up RAID Level 10 with Hot Spare sleds To set up one RAID Level 5 group with one Hot Spare sled To set up two RAID Level 5 groups with Hot Spare sleds
Glossary .................................................................................................i
Fibre Channel technology SAN technology SCSI protocol
Appendix A ATA Disk Technology .......................................................iii
Appendix B Information command returns .........................................iv
Driveinfo LUNInfo PartitionInfo ZoneInfo
Appendix C Sample Zoning Command Sequences ............................x
First time configuration (after download) Simple Symmetric Model Asymmetric Model Combined Symmetric/Asymmetric Model
Appendix D Product Safety ...................................................................xiii
Safety compliances EMC specifications Radio and television interference
Appendix E Specifications ....................................................................xiv
Environmental and physical Rack mount dimensions Floor mount dimensions
Appendix F Part numbers .....................................................................xv
Appendix G Warranty ............................................................................xvi
Manufacturer limited warranty Contact ATTO Technology, Inc.
ATTO Technology Inc.
Diamond Storage Array Installation and Operation Manual
Preface
This guide will take the technology-savvy user through the installation and maintenance of the Diamond Storage Array.
The Diamond Storage Array was designed to meet your need for large amounts of easily accessible storage using proprietary Aggregated Data Transfer Technology (ADXT ATA disk drives with the sustained data transfer rates required by sophisticated computer users.
Your comments help us improve and update our products. Contact us:
ATTO Technology, Inc. 155 CrossPoint Parkway Amherst, New York 14068 (716) 691-1999 • voice (716) 691-9353 • fax http://www.attotech.com/diamond
ATTO Technology can also be reached via e-mail at the following addresses:
Sales Support: sls@attotech.com Technical Support: techsupp@attotech.com
Disclaimer
Although reasonable efforts have been made to assure the accuracy of the information contained herein, this publication could include technical inaccuracies or typographical errors. Manufacturer expressly disclaims liability for any error in this information and for damages, whether direct, indirect, special, exemplary, consequential or otherwise, that may result from such error including but not limited to loss of profits resulting from the use or misuse of the manual or information contained therein (even if Manufacturer has been advised of the possibility of such damages). Any questions or comments regarding this document or its contents should be addressed to Manufacturer.
J
) to merge the power of multiple, high performance
Manufacturer provides this publication as is, without warranty of any kind, either express or implied, including, but not limited to, the implied warranties for merchantability or fitness for a particular purpose.
Information in this publication is subject to change without notice and does not represent a commitment on the part of Manufacturer. Changes may be made periodically to the information herein; these changes will be incorporated in new editions of the publication. Manufacturer reserves the right to make improvements and/or changes at any time in product(s) and/or program(s) described in this publication.
ATTO Technology Inc.
Diamond Storage Array Installation and Operation Manual
1.0 Product Overview
The Diamond Storage Array offers up to 24 ATA disk drives in a rack mount or floor model configuration.
The Diamond Storage Array is ideally suited for data intensive applications such as web hosting, e-mail servers, on-line transaction processing, digital video editing, medical imaging and digital audio editing. Virtually any high performance computing system with a growing need for storage capacity can use the power of the array.
With the cost effective approach of using ATA disk drives, you can add more storage capacity as your needs grow without the costs of other disk storage technologies. You can also improve the performance and capacity of the array cabinet as technology progresses by simply replacing disk drive sleds and host interface cards.
The array is operating system independent and supports all popular computer hardware platforms and network environments.
Three interface options are available: a 1-gigabit Fibre Channel interface, a 2-gigabit Fibre Channel interface, and an Ultra160 SCSI interface.
The array is a fully populated, turnkey solution with drives pre-installed. It is fully supported by a highly trained customer service and engineering staff.
The Diamond Storage Array uses Aggregated Data Transfer Technology (ADXT
J
) to merge the performance of multiple ATA drives to achieve sustained, full bandwidth data transfer rates. ADXT provides end users with the power and sophisticated data control needed to take moderately priced ATA disk drives, combine them in a disk storage array, and power them to the performance levels of SCSI or Fibre Channel disk arrays.
Features
• Up to 7.2 Terabytes initial configuration
(expandable with future drive technology)
• 24 ATA disk drive capacity
• Aggregated Data Transfer Technology
(ADXTJ) for high performance/scalability
• Ultra ATA 100 megahertz (Mhz)
• JBOD, RAID Level 0, RAID Level 1, RAID Level
10 and RAID Level 5 configurable
• Partitioning capability
• Zoning capability compatible with third party
servers, switches and with deliverables from industry standards organizations.
• Hot spare sleds: replace degraded sleds with
spares on the fly using software
• Staggered drive spin-up to reduce peak power
demand
1
ATTO Technology Inc.
• Tagged command queuing to process up to 255 simultaneous data requests
• RS-232 management for local management control; Ethernet option available for setup connection only
• ExpressNAV
• Two power supplies capable of 85-264 V (rated 100-240V AC) operation (340 watts each)
• UL, TUV and CE marked and compliant
• Internal thermal and power management
• Redundant hot swappable power supplies with integrated thermal and power management
• Floor model cabinet or 19” 3U rack mount
TM
browser-based user interface
Fibre Channel model
• 2 gigabit Fibre Channel Port (single or dual channel)
• SFP-based Fibre Channel interface supports long wave and short wave optical cables
• Built-in hub for daisy-chaining
• Up to 9,500 I/Os per second per port
• Up to 240 MB/sec. sustained Fibre Channel transfer rates per interface
SCSI model
• Ultra 160 SCSI bus.
• Dual stacked VHDCI connectors for daisy­chaining and termination
• SCSI Target ID selection switch
• Support for single-ended and LVD SCSI
• No onboard termination
Exhibit 1.0-1 Back of rack mount model, Diamond Storage Array. Left: 2 gigabit Fibre Channel. Right: SCSI.
2
Diamond Storage Array Installation and Operation Manual
2.0 Technical Overview
The Diamond Storage Array uses Aggregated Data Transfer Technology (ADXT) to achieve the high data transfer performance you need. ADXT bandwidth Fibre Channel data transfer rates.
merges the performance of multiple ATA drives together to achieve sustained, full
Unlike other storage arrays which use expensive SCSI or Fibre Channel disk drives to achieve performance, the Diamond Series uses lower cost ATA disk drives combined with an intelligent midplane to create a storage array with price and performance characteristics. The intelligent midplane contains hardware and software which provide the proprietary ADXT switched data management and data movement technology. The storage array delivers faster sustained
Exhibit 2.0-1 Data pathways and architecture for Fibre Channel operation
,
a
data transfer rates as well as impressive I/Os per second.
The array is made up of dual SCSI or Fibre Channel host interface cards, the intelligent midplane, a system management card, and 12 independent disk drive sleds containing up to 24 ATA disk drives.
ATA disk drives
ATA disk drives were known originally as Integrated Drive Electronics (IDE), a low end disk interface. The original IDE interface was low performance, single threaded (no simultaneous I/O requesting), contained minimal error detection and was unsuitable for
3
ATTO Technology Inc.
computer applications requiring high performance and high reliability. As IDE was refined and acquired important capabilities, its name was changed to ATA, Advanced Technology Attachment.
• UltraDMA transfer protocol similar to high performance SCSI disk protocol operating at 66 MB/sec.
• Double-clocking of data transfers, doubling disk data transfer rates
• CRC (Cyclic Redundancy Check) code allowing full error detection and data reliability
• Multi-threaded I/O support
• Overlapped Command Support: allows commands to be simultaneously active on multiple drives on the same ATA bus.
• Command Queuing which allows simultaneous multiple read/write commands to be sent to each drive, reducing command overhead and allowing the drive to service commands in the most efficient manner: similar to the SCSI feature Tagged Command Queuing.
• Faster drive speeds (5400/7200 RPM) with higher media transfer rates
• A communication protocol and interface with a fundamental lower cost structure than SCSI or FC interfaces
ATA disk drives operate at performance and data integrity levels similar to those that were previously available only on SCSI or Fibre Channel disk drives.
ADXT Powers ATA to New Levels
The original notion of RAID was to build high capacity, reliable storage subsystems using large numbers of inexpensive disk drives. Thus its original definition: Over time that definition became
Independent Drives
inherent cost advantage in a RAID system was lost.
Intelligent Midplane
heart of the Diamond Series storage array is the intelligent midplane with ADXT the data rates of individual ATA disk drives to create high data transfer rates. This
Redundant Array of Inexpensive Drives
Redundant Array of
and the
Disk 2
Disk 2
Disk 1
Disk 1
to sum or
The
aggregate
AIE2AIE3AIE4AIE5AIE6AIE7AIE8AIE9AIE10AIE11AIE
AIE
1
Disk 2
Disk 1
.
Disk 2
Disk 1
technology enables features such as serverless backup, advanced error protection, metadata storage techniques, virtualization software, thermal management and advanced enclosure services.
The midplane contains a combination of custom Application Specific Integrated Circuits (ASICs), processors and proprietary embedded software divided into three main processing sections which handle the data being read or written to the ATA disk drives from the Fibre Channel or SCSI host interfaces. The Virtual Device Manager (VDM), Data Routing Fabric (DRF) and ATA Interface Engines (AIE) organize data streams for storage or retrieval.
Virtual Drive Manager
Data is accessed through Virtual Drives using an implementation of the standard SCSI protocol controlled by the Virtual Device Manager.
Data Routing
Fabric incoming or outgoing data is routed between the ATA Interface Engines (AIE) and the Fibre Channel or SCSI interface by the custom Data Routing Fabric ASIC, a high speed, low latency transfer fabric with more than 2 GB/sec. of bandwidth supported by up to 512 Megabytes of memory.
ATA Interface Engine (AIE)
The interface to each pair of drives is through a custom ATA Interface Engine ASIC. The AIE implements the typically software-intensive ATA interface completely in silicon. Each AIE contains a dedicated ATA protocol processor to completely automate command and protocol processing.
Disk 2
Disk 2
Disk 2
Disk 2
Disk 2
Disk 2
Disk 2
Disk 1
Disk 1
Disk 1
Disk 1
Disk 1
Disk 1
Disk 1
DATA ROUTING FABRIC
text
VIRTUAL DEVICE MANAGER
Disk 2
Disk 1
12
FIBRECHANNEL
INTERFACE 1
FIBRECHANNEL
INTERFACE 2
Diamond Storage Array Installation and Operation Manual
4
3.0 Installation Instructions
If you are familiar with the Diamond Storage Array, Fibre Channel, SCSI and RAID configurations, you may set up and configure the array using these instructions. You will find details, illustrations and other guidance for more involved operations and special cases in the rest of this manual.
CAUTIONCAUTION
Before configuring the Diamond Storage Array, ensure that any computer data to be stored on the array is properly backed up and verified. The Manufacturer is not responsible for the loss of any data stored on the array under any circumstances and any special, incidental, or consequential damages that may result thereof.
Three major steps are required to set up the Diamond Storage Array
1 Physically set up Diamond Storage Array
2 Connect to Ethernet
3 Configure the drives
Step 1: physical setup
1 Make sure the Diamond Storage Array is
mounted properly and has adequate air flow around it.
2 Insert the appropriate connector into the
interface card in the back of the Diamond Storage Array.
3 Connect the Fibre Channel or SCSI cable from
your host computer system to the connector. To use ExpressNAV browser-based management interface and configure your Diamond Storage Array, you must connect to the Ethernet port.
4 To set up the Ethernet connection: connect a
cross-over cable (for a direct connection to a PC) or regular network cable from a network device to the RJ45 Ethernet port on the Ethernet management card on the front of the Diamond Storage Array.
Step 2a: set up the Ethernet connection
The Diamond Storage Array supports service operations over the RS-232 serial port using standard terminal emulation software available with most systems.
1 Connect a DB-9 null modem serial cable
between the port marked RS-232 on the front of
the Diamond Storage Array and the computer’s serial port. The cable must be no longer than three meters.
2 Make sure the power switches on the power
supplies on the rear of the unit are in the Stand­by position.
3 Plug in the power cords to the back of the unit,
then into an appropriate power source (100-240 VAC).
4 Reboot your host computer system.
5 Press the Stand-by power switch for each
power supply on the Diamond Storage Array to the ON position.
6 Upon successful power up and diagnostics, the
unit displays the POST (power up self test) information.
The Diamond is now in Command Line Interface mode. You may modify the setup of the Diamond Storage Array using the CLI (refer to
via ASCII-based Commands
on page 27), but the
CLI: Interface
easiest method to configure the array is by using ATTO ExpressNAV software, a graphical user management interface accessed through a standard Internet browser. Refer to
ExpressNAV: Browser-based Interface
ATTO
on page 23.
Step 2b: connect to Ethernet
If the Diamond Storage Array is attached to a DHCP server
1 At the Ready prompt after POST (refer to Step
6 above), type set IPDHCP enabled
2Type SaveConfiguration Restart
3At the Ready prompt after POST (see above),
type get IPAddress
4 Enter this address into your browser.
5 The ATTO ExpressNAV screen appears. Log in
using the Telnet defaults:
Username: Telnet Password: Diamond
6 Follow the screens to find information about the
array or to configure the array from the factory-
5
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
default settings. The Diamond Storage Array may be set up in a JBOD, RAID Level 0, 5 or 10 configuration with or without Hot Spare sleds
If the Diamond Storage Array is not attached to a DHCP server and you wish to change the defaults
1 At the Ready prompt after POST (see above),
type set IPAddress [desired IP address].
2Type set IPSubnetMask [desired IP Subnet
Mask]
3Type set IPGateway [desired IP Gateway]
4Type SaveConfiguration Restart to save the
configuration and restart the Diamond Storage Array
5 After the powerup and POST complete, type
the IP address from step 1 above into your browser.
6 The ATTO ExpressNAV screen appears. After
logging in (refer to Step 2b: connect to Ethernet, Step 5 on page 5), follow the screens to find information about the array or to
configure the array from the factory-default settings. The Diamond Storage Array may be set up in a JBOD, RAID Level 0, 1, 5 or 10 configuration with or without Hot Spare sleds.
Step 3: configure drives
The simplest way to set up configurations is to use the ExpressNAV interface. Refer to
ExpressNAV: Browser-based Interface
ATTO
on page 23 for more information on the interface. After logging in, follow the screens to find information about the array or to configure the array from the factory-default settings.
You may the CLI to set up RAID, partitions and zones.
Note
All arrays using RAID level 10 and Hot Spare sled options must be fully populated.
6
Quick start
3.1 Components
The Diamond Storage Array has been designed to be easy to use, maintain and upgrade. It features a durable steel outer case and modular components in either a floor or a rack mount model.
Immediately upon receipt, check the shipping carton for damage from mishandling. Contact us at once via the means that is easiest for you (refer to
Warranty
on page xvi) if the carton has been
mishandled or displays any signs of damage.
The front of the Diamond Storage Array provides access to the management card and disk drive sleds. The rear of the unit holds the host interface cards, power supplies and blower assemblies.
CAUTIONCAUTION
All modular components must be replaced by qualified personnel only.
Floor model
The management system card is at the top front of the case. At its center is a DB-9 serial RS-232 port, a connection for setup, monitoring and upgrade of the unit from any computer system with an RS-232 interface. The optional 10/100 BaseT Ethernet management services card enables Telnet-based monitoring and
management. It also provides the ability to update the firmware in the array via FTP.
LEDs to the port’s right indicate fault, unit ready, host interface cards A and B installation status, and the power status for each power supply.
Below the management system card are individual disk drive sleds which also have LEDs for each drive’s status. Each sled contains two hard drives. Up to 24 hard drives may be installed on the 12 sleds. Empty bays should be covered by blank faceplates or empty sleds. Access is provided by loosening two screws and gently pulling on the sled handle.
On the rear of the unit are blowers which support hard drive, cabinet and power supply cooling. The blowers are held in by removable screws. Correct operation is displayed by a LED at the top of each panel.
The power supplies for the array, also in the rear of the unit, are accessible by loosening two screws and pulling on the power supply module handle. The power standby on/off switch is at the top of each module. A yellow LED indicates and a green LED indicates
on
. The power cord
caution
socket is at the bottom of each power supply.
Between the power supplies and blower assemblies are two slots that hold the Host Interface cards. The HIC is the connection point into the array and is available in three options: 1­Gigabit Fibre Channel, 2-Gigabit Fibre Channel, or Ultra 160 SCSI. Host Interface cards have and
on-line
or
fault
and
activity
LED indicators,
fault
depending on the model.
SCSI
The SCSI card faceplate has a rotary binary-coded hex switch to set the SCSI ID of the array. The SCSI card also has an in channel, to connect via cable to the unit’s communication source, and an out channel, available for daisy­chaining arrays together or to complete termination using an external LVD terminator.
Rack mount
The system management card is at the left front of the case. At its center is a DB-9 serial RS-232 port
7
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
which allows a connection for setup, monitoring and upgrade of the unit from
System Management Card
any computer system with an RS-232 interface. The optional 10/100 BaseT Ethernet management services card enables Telnet-based monitoring and management. It also provides the ability to update the firmware in the
Disk drive sleds (12)
array via FTP. LEDs farthest to the left indicate
Power supplies
fault, unit ready, Host Interface cards A and B installation status, and the power status for each power supply.
To the right of the system management card are individual disk drive sleds which also have LED indicators for each drive’s status. Each sled contains two hard drives. Up to 24 hard drives may be installed on the 12 sleds. Empty bays should be covered by blank faceplates or filled with empty sleds to promote effective cooling. Access is provided by loosening two screws and pulling on the sled handle.
In the rear of the unit are the blower assemblies which support hard drive, cabinet and power supply cooling. Correct operation is displayed by a LED at the top of each panel. The blowers are held in place by removable screws.
The power supplies for the array are accessible by loosening two screws and pulling on the power supply module handle. The power standby on/off switch is at the top of each module. A yellow LED indicates
caution
and a green LED indicates on.
Intelligent midplane (internal)
Host interface cards
Blower assemblies
The power cord socket is at the bottom of each power supply.
Between the power supplies and blower assemblies are two slots that hold the Host Interface cards. The HIC is the connection point into the array and is available in 2-Gigabit Fibre Channel, or Ultra 160 SCSI. Host Interface cards
fault
and
have
on-line
or
fault
and
activity
LED
indicators, depending on the model.
SCSI
The SCSI card faceplate has a rotary binary-coded hex switch which allows you to set the SCSI ID of the array. The SCSI card also has an in channel, to connect by a cable to the unit’s communication source, and an out channel, available for daisy-chaining arrays together or to complete termination using an external LVD terminator.
8
Components
3.2 Physical Setup
The Diamond Storage Array is shipped completely assembled with two 120 VAC power cords for use in the United States and Canada.
Immediately upon receipt, check the shipping carton for damage from mishandling. Contact us at once by the means easiest for you (refer to
Warranty
on page xvi) if the carton has been
mishandled or displays any signs of damage.
Floor Model
The Diamond Storage Array is heavy (about 92 pounds fully loaded) and requires two people to lift and carry it safely. Place the array on a level surface and make sure there is adequate space in the front and back of the unit for proper cooling and airflow. Continue with the general instructions.
Rack Mount
The array fully loaded is heavy (about 86 pounds). The unit should be handled with care and requires two people to lift, carry and/or install it safely.
The array can be mounted via several different methods in a 19” rack with 3U (5.25”) of vertical space required. Air flow should not be restricted in any way.
Installation in a rack may create a differential between the room ambient temperature and the internal ambient temperature in the rack. While the maximum internal operating temperature of the array is 47°C, you should not run the system at the maximum temperature for extended periods. Therefore, ensure that the room ambient temperature is kept below 30°C
for best operation.
Each side of the rack mount array chassis has three pairs of mounting holes. One is located near the front of the rack, one near the unit’s center of gravity, and one near the rear of the rack. The holes accommodate 10/32 screws but the screws can protrude no farther than .375 inches into the rack.
Spaced rail pairs in some rackmount cabinets.
You can mount the array using two sets of rail pairs spaced to accommodate the overall length of the unit (approximately 23 inches). Mount using the rack mount brackets on the front and rear of the unit fastened to the rail pairs using 10/32 pan head screws with lock and flat washers.
Rack mount cabinets with stationary shelf or tray system.
The shelf or tray must be able to support at least 125 pounds. The shelf or tray must be installed and secured to the rack before installing the array. Secure the front of the array to the rack with 10/32 screws, locks and flat washers.
Sliding shelf or tray type systems
should never
be used under any circumstances.
Two point open rack system.
The rack must be strong enough to support the array properly. Mounting brackets should be moved to the
9
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
centermost mounting holes and secured using 10/32 screws.
CAUTIONCAUTION
Do not mount multiple arrays on a two-rail rack or mount the array above the midpoint of a two-rail rack system. Do not mount the array on any kind of rail-type system. The array is too heavy and does not have the proper hole pattern for rails.
Note
Insure the array has adequate air flow.
General Instructions
1 Insert the proper connector into the Host
Interface Card in the back of the array. (refer to
Power Sw itch Posit ions
On Stand-by
2 Connect the cable (Fibre Channel or SCSI)
from your host system to the Host Interface Card connector on the back of the array. The cable you use depends upon your application, the environment and distance.
Connecting a Fibre Channel
on page 11 for Fibre
Array
Channel and
SCSI Array
SCSI).
Connecting a
on page 13 for
3 Make sure the power switches on the power
supplies on the rear of the unit are in the stand­by position. Plug in the power cords to the back of the unit, then into an appropriate power source (100-240 VAC). The power source must be connected to a protective earth ground and comply with local electrical codes. Improper grounding may result in an electrical shock or damage to the unit.
4 Press the stand-by power switch for each
power supply to the ON position. When the green power LED on the back of the unit is lit, the power supply is fully operational and delivering power to the system. The power LED on the front of the array lights while the firmware executes.
When the power is turned on, the LEDs on the front of the array flash twice. Drives spin up in groups of three every one to two seconds. The individual LEDs blink. After all available drives have spun up, the individual drive LEDs stay lit. When all available drives are operational, the ready LED on the top front panel of the management card remains lit.
5 Reboot your computer
6 Determine the best configuration for your needs
(i.e. JBOD, RAID, etc.) and refer to the rest of this manual for configuration information.
Exhibit 3.2-1 Back side of a rack mount array.
10
Physical setup
3.2.1 Connecting a Fibre Channel Array
The Diamond Storage Array supports up to two Fibre Channel Host Interface Cards (HIC). Physical connections and CLI commands contribute to the Fibre Channel topology.
The cable you use depends upon your application, the environment and the distance required for your
2 Gb HIC uses
2 SFPs to connect up
to 2 FC cables
storage area network.
To comply with FCC and CE EMI for the 2-gigabit
Host Interface Card, use fiber optic cables.
Exhibit 3.2-1 Fibre Channel cable options
Cable length Cable type Cable size
<10 meters Unequalized copper
>10 <30 meters Equalized copper
Up to 175 meters multi mode fiber optic 62.5
Up to 500 meters multi mode fiber optic 50 micron
Up to 10 kilometers
single mode fiber optic 9 micron
micron
The Diamond Storage Array may have two Fibre Channel Host Interface Cards (HIC). In 2 gigabit Fibre Channel arrays, each HIC is connected by a Fibre Channel cable via a SFP (small form factor pluggable) module into a point-to-point or loop Fibre Channel topology.
Install the SFP according to the manufacturer’s instructions.
Each HIC has two ports and an on-board hub. Each port has an SFP module to connect to Fibre Channel. Each HIC is independent of the other, so that one may be connected into a point-to-point topology and the other into a loop. However, if one port in a HIC is connected into a point-to­point topology, the other port cannot be used.
Autoconfiguration
The array automatically determines which HICs are installed and if they are in loop or point-to­point topologies.
If you wish to see how the unit has been set up, go to the CLI commands and type
Info
or go to the
Status page of the ExpressNAV interface (refer to
ATTO ExpressNAV: Browser-based Interface
on page
25. The return displays the data rate and connection mode for each HIC (FC 0 and FC 1).
Manual configuration
You may manually configure the array using CLI commands (refer to
Commands
Channel
(refer to
on page 35.) or access the page of the ExpressNAV interface
ATTO ExpressNAV: Browser-based Interface
Fibre Channel Configuration
Fibre
on page 25).
FCConnMode
topology for both HICs on an array. Options are loop only (loop), point-to-point only (ptp), loop preferred (loop-ptp) or point-to-point preferred (ptp-loop).
Loop connects to either an FC-AL arbitrated loop or a fabric loop port (FL_Port) on a switch.
Point-to-point (ptp) connects to a direct fabric connection, such as an F port on a switch.
Loop-ptp allows the array to determine what kind of connection to use, but tries to connect in loop mode first, then point-to-point mode.
Ptp-loop allows the card to determine what kind of connection to use, but tries to connect in point-to­point mode first, then loop mode.
FcDataRate
at which both HICs on a Diamond operate. Choices are 1 gigabit, 2 gigabit and autodetection.
specifies the Fibre Channel
specifies the Fibre Channel data rate
One of the advantages of using loop topology for Fibre Channel connections is that it allows arrays to be daisy-chained together.
11
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
Exhibit 3.2-2 Possible 2 gigabit Fibre Channel physical connections depending on which Fibre Channel connection mode has been selected.
point-to-
point mode
loop
mode
Diamond Storage Array A Diamond Storage Array B
no
connection
possible
loop
mode
daisy-chain
loop
mode
loop
mode
loop
mode
loop
mode
12
Fibre Channel connections
3.2.2 Connecting a SCSI Array
The SCSI Diamond Storage Array uses a VHDCI connector and SCSI cables to connect to a host. It automatically detects the type of Host Interface Card it is using without any intervention.
To connect the SCSI Diamond Storage Array
1 Insert a SCSI VHDCI connector into the Host
Interface Card in the back of the array. If the SCSI array is the last device on the bus, you
must attach a VHDCI terminator to one connector of the SCSI Host Interface Card or connect a cable between the second connector and the next device on the SCSI bus.
Exhibit 3.2-1 SCSI cable options.
Bus speed, MB/sec.
max.
SCSI-1 5 8 6 - 8
Fast SCSI 10 8 6 - 8
Fast Wide SCSI 20 16 6 - 16
Wide Ultra SCSI 40 16 3 - 4
Wide Ultra SCSI 40 16 1.5 - 8
Wide Ultra 2 SCSI 80 16 - 12 16
Ultra 3 or Ultra160 SCSI
160 16 - 12 16
Bus width,
2 The SCSI Host Interface Card has a rotary
binary-coded hex switch which allows you to set the SCSI ID of the HIC. Be sure the selected ID is different from all other SCSI devices on the bus.
Note
If slower devices are connected on the same SCSI bus as the Ultra 160 array, the bus communicates at the rate of the slowest device.
Max. bus lengths,
bits
meters
Single-ended LVD
Max. device
support
Exhibit 3.2-2 SCSI interface cards: left without terminators attached; right with a terminator attached.
13
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
14
SCSI connections
3.3 Determining Drive and Sled Designations
g
The Diamond Storage Array has been designed with 12 sleds, each holding two drives. The easiest way to configure an array is to use all the drives on all the sleds. The firmware uses a numbering system to determine which drives and sleds it is affecting.
All sled slots should be filled contiguously, starting with the first slot next to the management card.
When configuring a array with fewer than 12 drive sleds, you must consider several factors: RAID level, number of physical drives/sleds and the end configuration you are trying to achieve. Review the information about each configuration to determine how each configuration would be affected by using fewer sleds.
Numbering conventions
The Diamond Storage Array with firmware version 3.1 and newer uses a unique numbering convention to orient its drives and sleds to the controlling firmware. Older versions do not use this convention. (refer to
Updating Firmware
on page 75 for information about updating the array firmware.)
Physical
refers to the physical drives in the array, the hardware that actually exists in a physical sense.
Logical (or virtual)
refers to what the host operating system recognizes as an entity. Two physical drives may be seen as one logical drive by the operating system.
Logical disks do not always map one-to-one with physical disks.
In RAID configurations, for example, several physical disk drives (or portions of several physical drives) are grouped into a logical disk, call a RAID Group or a Logical Unit
1 2 3 4 5 6 7 8 9 10 11 12
Disk 2
Disk 2
Disk 2
LUN 13
LUN 14
LUN 15
Disk 2
LUN 16
Rack Mount Drive Sleds
LUN 17
Disk 2
LUN 16
LUN 17
Disk 2
LUN 18
Disk 2
LUN 19
LUN 20
LUN 19
(LUN). Each RAID group is broken into logical blocks of 512 bytes each, numbered 0 through n (the Logical Block Number or LBN). A 100 GB LUN has approximately 200,000,000 logical blocks.
A RAID group is also referred to as a Virtual Drive.
A LUN is associated with a RAID group or Virtual Drive unless you are using partitions. If you have partitions, a LUN is associated with each partition. A RAID Group or Virtual Drive may then have multiple partitions or LUNs.
Physical numbering
The Diamond RAID Storage Array contains
• Up to 24 physical disk drives
• Two drives mounted on 12 physical drive sleds
• Sleds are numbered 1-12, starting at the top (floor units) or the left (rackmount units).
• Each sled is connected to its own internal ATA bus with two disk drives, numbered 1 and 2.
• Two green LEDs, labeled Drive 1 and Drive 2, indicate activity for the two drives. They remain solidly lighted when there is no activity.
Disk 2
Disk 2
Disk 2
Disk 2
LUN 20
LUN 21
LUN 22
LUN 23
Disk 2
LUN 24
Mana
Drive 1
LUN 1
Drive 1
LUN 2
Drive 1
LUN 3
Drive 1
LUN 4
Drive 1
LUN 5
Drive 1
LUN 6
Drive 1 LUN 7
Drive 1
LUN 8
Drive 1
LUN 9
Drive 1 LUN 10
Drive 1 LUN 6
ement car d
Drive 2
LUN 16
Drive 2 LUN 17
LUN 18
Drive 2
LUN 19
Drive 2 LUN 20
Drive 2 LUN 21
LUN 22
Drive 2 LUN 13
Drive 2
LUN 14
Drive 2 LUN 15
Drive 2
LUN 20
LUN 19 Drive 1 LUN 12
Drive 2
LUN 17
1 2 3 4 5 6 7 8 9 10 11 12
Floor Model Drive Sleds
Management Card
15
Disk 1 LUN 1
Disk 1 LUN 2
Disk 1 LUN 3
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
Disk 1 LUN 4
Disk 1 LUN 5
Disk 1 LUN 6
Disk 1 LUN 7
Disk 1 LUN 8
Disk 1 LUN 9
Disk 1
LUN 10
Disk 1
LUN 11
Disk 1
LUN 12
Drive 1 LUN 11
Drive 1
LUN 12
Drive 2 LUN 23
Drive 2 LUN 24
Logical Numbering
Logical numbering depends on the RAID configuration of the storage array. Current firmware includes RAID Level 0, RAID Level 5, RAID Level 1, RAID Level 10, Zoning and Partitioning capability and hot spare sleds. The default configuration is QuickRAID0 0 or JBOD (Just a Bunch of Disks), in a single zone.
Examples
JBOD
mode uses 24 LUNs. Each LUN is equivalent to one physical drive. The array can operate with several sleds missing, but the empty sled(s) are treated as offline and cannot be configured.
RAID Level 0 (QuickRAID0 [n])
With a fully populated array, RAID 0 may be configured as 1, 2, 3, 4, 6, or 12 LUNs. As RAID0 1, all 24 physical drives are configured as a single stripe group or LUN. You may also configure two LUNs of 12 drives each, three LUNs of eight drives each, four LUNs of six drives each, six LUNs of four drives each and 12 LUNS of two drives each. (See Exhibit 5.2-3 in
RAID Level 0
on page 52). The command assumes there are 24 drives available to configure the specified number of LUNS.
RAID Level 1 (QuickRAID1)
has no options: the array can be configured into six groups, LUNs 1­6, with each LUN containing two physical sleds; each sled in the LUN would be a mirror image of the other sled in the LUN.
RAID Level 10 (QuickRAID10)
The QuickRAID10 command first creates six mirrored groups, then stripes them into groups of one, two or three RAID 10 groups. Each group is a LUN. Issuing QuickRAID10 2, the 12 physical sleds are configured as six mirrored pairs, then striped into two LUNs.
RAID Level 5 (QuickRAID5)
with a fully populated Array may be configured as 1, 2, 3 or 4 LUNs. As QuickRAID5 1, all 24 physical drives are configured as a single RAID 5 LUN. You may also configure two LUNs of 12 drives each, three LUNs of eight drives each, or four LUNs of six drives each. RAID Level 5 parity reduces the usable capacity of each LUN by the equivalent of one drive sled.
Unique serial number for each LUN
Each LUN in a system has a unique 24-character serial number which is updated when a system configuration changes. It is stored in a Device Association Table on each drive.
When a drive or sled is replaced in QuickRAID configurations that allow for hot swap, a new serial number is computed and is stamped onto all member drives of the RAID group. The CLI command
LUNInfo
the ExpressNAV interface
or the
Logical Units page of
displays the serial
number for each drive.
Exhibit 3.3-1 Format of the 24 characters of the unique serial number for each LUN
Character Placement Valid values
0-19 Any ASCII char Drive Serial Number
20 1 for RAID1
A for RAID10 0 for RAID0 5 for RAID5 X for JBOD
21 A – Z LUN Configuration Iteration Character: starts at A and increments every time
22 A – Z Generation Character: starts at A and increments every time a new
23 0 Reserved for future use
Example of unique serial number for a JBOD configuration:
1231231231231231231XAC0
RAID Configuration Character
a member drive is replaced in a RAID Group. Reverts back to A any time the Generation Character is updated
configuration is stamped on the system. When it reaches Z, rolls over to A.
16
Drive, sled numbering
4.0 Accessing the Array
Communicate with the Diamond Storage Array through the Fibre Channel link, the RS-232 port or the Ethernet port using Command Line Interface commands or ATTO ExpressNAV,
management console
.
an integrated user
You may configure and tune the Diamond Storage Array for different environments and applications, update the firmware, monitor internal power and temperature status, report on hardware diagnostics and log failures.
Three avenues are available:
• In-band SCSI over Fibre Channel and over SCSI
• RS-232 port
• Telnet and SNMP over Ethernet
The following chapters describe how to access the array and use the Command Line Interface or ATTO ExpressNAV, an integrated user management console.
Command Line Interface
The CLI provides access to the array through ASCII command lines.
An initial display, after powering up the unit or restarting the firmware, contains the information in Exhibit 4.0-1. Once the initial display is complete, with the word
Ready
, you are in the
CLI mode.
Help
Type
to display a list of all commands
available.
ATTO ExpressNAV
ATTO ExpressNAV is an integrated configuration tool accessible through an Ethernet connection. Platform independent, ExpressNAV contains all the current capabilities of the CLI in a user­friendly GUI console. A menu on each page provides access to information and configuration operations.
Refer to
ATTO ExpressNAV: Browser-based Interface
on page 25 for more information on the program.
In-band SCSI over Fibre Channel
In-band SCSI commands (
) may be issued to the array to manage
Buffer
Write Buffer
and
Read
configuration via two mechanisms:
• In-band CLI over SCSI, where ASCII CLI
commands, may be issued via CLI commands except
ID/value
SCSI CDB (command descriptor block) to select the buffer ID of the configuration parameters to be affected, and the new value of the parameter. Most configuration options are available.
, where the application program uses a
Zmodem
Write Buffer
are supported.
. All
RS-232 port
The array provides remote service operations over the RS-232 serial port using standard terminal emulation software available with most systems.
Set the following serial parameters in your terminal program:
• Bits per second: 115200
• Data Bits: 8
• Parity: None
• Stop Bits: 1
• Flow Control: None.
• Terminal type: ASCII
• Echo: on.
Ethernet port
The 10/100 BaseT Ethernet port provides Telnet­or SNMP-based monitoring and management.
The default IP address is 10.0.0.1; the default subnet mask is 255.255.0.0. To change the defaults, first be configure the array for the network using the RS-232 port to establish the correct IP address. The management port provides TCP/IP-based monitoring and management.
17
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
SNMP
SNMP, or Simple Network Management Protocol, is an application layer protocol that allows network devices to exchange management information. Through a combination of standard and custom MIBs (Management Information
indications to an SNMP server, allowing the array to be managed with other devices in a complex system through a common interface.
Use CLI commands to configure up to six unique Trap addresses. A trap is a way for the array to tell the SNMP server that something has happened.
Base), the array provides status and error
Exhibit 4.0-1 The POST information displays after boot of the Diamond Storage Array.
Diamond Storage Array (c) 2004 ATTO Technology, Incorporated.
Firmware version 5.40 release date Mar 30 2004, 10:43:06 Build 021G
Power-On Self-Test Completion Status: GOOD 128 Megabytes of RAM Installed.
Interface Port 0 is not installed. Interface Port 1 is 1.0624 Gb/s Fibre Channel.
Interface 0 World Wide Name = 20 00 00 10 86 10 02 DC Interface 1 World Wide Name = 20 00 00 10 86 10 02 DC
Diamond Array Serial Number = "MIDP100197" Diamond Array Name = " " System Serial Number = "" Active Configuration = ATTO DiamondClass = (V)86 Internal Temperature = 23 C [5 - 47] ErrorLog Contents: NO ERRORS For help, type HELP.
Ready.
18
4.1 In-band CLI Over Fibre Channel
In-band Command Line Interface (CLI) configures and manages the Diamond Storage Array using SCSI-based CLI commands over a Fibre Channel port connection.
In-band CLI allows a programmer to configure the Diamond Storage Array while it is moving data. Using a programmer’s interface, CLI commands as described previously in this manual may be implemented.
In-band CLI is implemented as part of LUN 0. It uses a different LUN than the array, and reports as a Storage Enclosure Services (SES) device (device type 0x0D).
LUN 0 is visible on all Fibre ports but is actually a single unit. The default value for LUN 0 is 0x00.
LUN 0 must be reserved for each Write Buffer/Read Buffer pair, using the SCSI Reserve command to insure integrity of the in-band CLI session.
1 An initiator (host) sends a SCSI Reserve
command to LUN 0.
• If LUN 0 is not reserved by another initiator, LUN 0 is now reserved and available to begin a new CLI session.
• If the array configuration is reserved by a different CLI session (i.e. serial or Telnet), the in-band session does not allow modifications of the array configuration. If you try, the results buffer of LUN 0 returns:
Process X has the configuration reserved. ID of this session = Y Ready.
2 The initiator issues a SCSI Write Buffer
command to LUN 0. A must be accompanied by an ASCII buffer representing the CLI command string such as
set DiamondName Omega1
LUN 0 executes the command line and create feedback in the form of ASCII characters into a buffer. This buffer is 8 KB and circular. Retrieve the results by issuing a before issuing another
Write Buffe
Read Buffer
Write Buffer
r command
command
command.
3 A subsequent
the new command line and overwrite the previous results in the buffer with new results.
4 LUN 0 can be released by issuing a SCSI
Release
Write/Read Buffer pair, or multiple Write/Read Buffer pairs.
Initiator (Host) Diamond Storage Array
Reserve LUN 0 return: “ok”
Write Buffer LUN 0 bid ‘AA’ “get Temperature”
Read Buffer LUN 0 bid ‘AA’
Release LUN 0 return: “ok”
Write Buffer
command to the LUN after each
command executes
executes the CLI command, stores output in buffer
return:”Temperature=28C\r\n\ Ready.\r\n\0”
I/O details
The buffer sent to the Services LUN during the data out phase of a Write Buffer command must be:
• ASCII data
• maximum 80 bytes length
• terminated with either a carriage return character (0x0D), line feed character (0x0A) or NULL character (0x00)
• Characters following the first carriage return character, line feed character or NULL character are ignored.
The buffer retrieved from the Services LUN during the data-in phase of a
Read Buffer
command:
• ASCII data
• 8 KBytes (8192 bytes) in length
• terminated with a NULL character (0x00)
• Characters following the NULL character are meaningless.
19
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
Loading...
+ 87 hidden pages