Three major steps are required to set up the Diamond Storage Array
Step 1: physical setup
Step 2a: set up the Ethernet connection
Step 2b: connect to Ethernet
If the Diamond Storage Array is attached to a DHCP server
If the Diamond Storage Array is not attached to a DHCP server and you wish to change the
defaults
Step 3: configure drives
Status
Ethernet
SNMP
Serial Port
Fibre Channel
Storage Management
RAID
Clear Data
Logical Units
Partitions
Zoning
Rebuild
Configuration
Advanced
To use the Advanced Page CLI commands
Restart
Help
4.5 CLI: Interface via ASCII-based Commands .......................29
4.5.1 Summary of CLI Commands .......................................31
Diamond Storage Array Installation and Operation Manual
4.5.2 General Use Commands .............................................34
FirmwareRestart
Help
PartitionCommit
SaveConfiguration
SystemSN
VerboseMode
ZoneCommit
EthernetSpeed
FTPPassword
IPAddress
IPDHCP
IPGateway
IPSubnetMask
SNMPTrapAddress
SNMPTraps
SNMPUpdates
Te ln e t Pa s sw or d
TelnetTimeout
TelnetUsername
Configuring a fully-populated array
To set up one RAID Level 5 group with one Hot Spare sled
To set up two RAID Level 5 groups with two Hot Spare sleds
Configuring a partially-populated array
To set up one RAID Level 5 group with one Hot Spare sled
Removing RAID groups
5.9 Creating Zones .....................................................................69
Principles of Zoning
Factors to consider
Status and Sense Data
Configuring Zones
To create a zone
To remove zones
To change current zones
Other operations
Errors
Windows 2000 special instructions
Error Messages
System Fault LED
Command Line Interface messages
ERROR. Wrong/Missing Parameters
ERROR. Invalid Command. Type ‘help’ for command list
ERROR. Command Not Processed.
Audible Alarm
Specific situations and suggestions
If a drive fails to respond
If a power supply fails
To determine if the problem exists with the Host Interface Card or the connection
If you can’t access the array CLI via Ethernet
If you do not see the appropriate number of LUNs on the host machine
Diamond Storage Array Installation and Operation Manual
9.1 Hot Swap Operating Instructions .......................................87
Disk Drives
The following method is the safest way to perform a hot swap of a drive
Power Supplies
Blower Assemblies
To replace a blower assembly
9.2 Optional Hot Spare Sled ......................................................89
To set up RAID Level 1 with Hot Spare sleds
To set up RAID Level 10 with Hot Spare sleds
To set up one RAID Level 5 group with one Hot Spare sled
To set up two RAID Level 5 groups with Hot Spare sleds
Fibre Channel technology
SAN technology
SCSI protocol
Appendix A ATA Disk Technology .......................................................iii
Appendix B Information command returns .........................................iv
Driveinfo
LUNInfo
PartitionInfo
ZoneInfo
Appendix C Sample Zoning Command Sequences ............................x
First time configuration (after download)
Simple
Symmetric Model
Asymmetric Model
Combined Symmetric/Asymmetric Model
Appendix D Product Safety ...................................................................xiii
Safety compliances
EMC specifications
Radio and television interference
Appendix E Specifications ....................................................................xiv
Environmental and physical
Rack mount dimensions
Floor mount dimensions
Appendix F Part numbers .....................................................................xv
Appendix G Warranty ............................................................................xvi
Manufacturer limited warranty
Contact ATTO Technology, Inc.
ATTO Technology Inc.
Diamond Storage Array Installation and Operation Manual
Preface
This guide will take the technology-savvy user through the installation and maintenance of the Diamond Storage
Array.
The Diamond Storage Array was designed to meet your need for large amounts of easily accessible storage using
proprietary Aggregated Data Transfer Technology (ADXT
ATA disk drives with the sustained data transfer rates required by sophisticated computer users.
Your comments help us improve and update our products. Contact us:
ATTO Technology, Inc.
155 CrossPoint Parkway
Amherst, New York 14068
(716) 691-1999 • voice
(716) 691-9353 • fax
http://www.attotech.com/diamond
ATTO Technology can also be reached via e-mail at the following addresses:
Although reasonable efforts have been made to assure the accuracy of the information contained herein, this
publication could include technical inaccuracies or typographical errors. Manufacturer expressly disclaims
liability for any error in this information and for damages, whether direct, indirect, special, exemplary,
consequential or otherwise, that may result from such error including but not limited to loss of profits resulting
from the use or misuse of the manual or information contained therein (even if Manufacturer has been advised of
the possibility of such damages). Any questions or comments regarding this document or its contents should be
addressed to Manufacturer.
J
) to merge the power of multiple, high performance
Manufacturer provides this publication as is, without warranty of any kind, either express or implied, including,
but not limited to, the implied warranties for merchantability or fitness for a particular purpose.
Information in this publication is subject to change without notice and does not represent a commitment on the
part of Manufacturer. Changes may be made periodically to the information herein; these changes will be
incorporated in new editions of the publication. Manufacturer reserves the right to make improvements and/or
changes at any time in product(s) and/or program(s) described in this publication.
ATTO Technology Inc.
Diamond Storage Array Installation and Operation Manual
1.0 Product Overview
The Diamond Storage Array offers up to 24 ATA disk drives in a rack mount or floor model configuration.
The Diamond Storage Array is ideally suited for data
intensive applications such as web hosting, e-mail
servers, on-line transaction processing, digital video
editing, medical imaging and digital audio editing.
Virtually any high performance computing system
with a growing need for storage capacity can use the
power of the array.
With the cost effective approach of using ATA disk
drives, you can add more storage capacity as your
needs grow without the costs of other disk storage
technologies. You can also improve the performance
and capacity of the array cabinet as technology
progresses by simply replacing disk drive sleds and
host interface cards.
The array is operating system independent and
supports all popular computer hardware platforms and
network environments.
Three interface options are available: a 1-gigabit Fibre
Channel interface, a 2-gigabit Fibre Channel interface,
and an Ultra160 SCSI interface.
The array is a fully populated, turnkey solution with
drives pre-installed. It is fully supported by a highly
trained customer service and engineering staff.
The Diamond Storage Array uses Aggregated Data
Transfer Technology (ADXT
J
) to merge the
performance of multiple ATA drives to achieve
sustained, full bandwidth data transfer rates. ADXT
provides end users with the power and sophisticated
data control needed to take moderately priced ATA
disk drives, combine them in a disk storage array, and
power them to the performance levels of SCSI or Fibre
Channel disk arrays.
Features
• Up to 7.2 Terabytes initial configuration
(expandable with future drive technology)
• 24 ATA disk drive capacity
• Aggregated Data Transfer Technology
(ADXTJ) for high performance/scalability
• Ultra ATA 100 megahertz (Mhz)
• JBOD, RAID Level 0, RAID Level 1, RAID Level
10 and RAID Level 5 configurable
• Partitioning capability
• Zoning capability compatible with third party
servers, switches and with deliverables from
industry standards organizations.
• Hot spare sleds: replace degraded sleds with
spares on the fly using software
• Staggered drive spin-up to reduce peak power
demand
1
ATTO Technology Inc.
• Tagged command queuing to process up to 255
simultaneous data requests
• RS-232 management for local management
control; Ethernet option available for setup
connection only
• ExpressNAV
• Two power supplies capable of 85-264 V (rated
100-240V AC) operation (340 watts each)
• UL, TUV and CE marked and compliant
• Internal thermal and power management
• Redundant hot swappable power supplies with
integrated thermal and power management
• Floor model cabinet or 19” 3U rack mount
TM
browser-based user interface
Fibre Channel model
• 2 gigabit Fibre Channel Port (single or dual
channel)
• SFP-based Fibre Channel interface supports
long wave and short wave optical cables
• Built-in hub for daisy-chaining
• Up to 9,500 I/Os per second per port
• Up to 240 MB/sec. sustained Fibre Channel
transfer rates per interface
SCSI model
• Ultra 160 SCSI bus.
• Dual stacked VHDCI connectors for daisychaining and termination
• SCSI Target ID selection switch
• Support for single-ended and LVD SCSI
• No onboard termination
Exhibit 1.0-1 Back of rack mount model, Diamond Storage Array. Left: 2 gigabit Fibre Channel. Right:
SCSI.
2
Diamond Storage Array Installation and Operation Manual
2.0 Technical Overview
The Diamond Storage Array uses Aggregated Data Transfer Technology (ADXT) to achieve the high data transfer
performance you need. ADXT
bandwidth Fibre Channel data transfer rates.
merges the performance of multiple ATA drives together to achieve sustained, full
Unlike other storage arrays which use expensive SCSI
or Fibre Channel disk drives to achieve performance,
the Diamond Series uses lower cost ATA disk drives
combined with an intelligent midplane to create a
storage array with price and performance
characteristics.
The intelligent midplane contains hardware and
software which provide the proprietary ADXT
switched data management and data movement
technology. The storage array delivers faster sustained
Exhibit 2.0-1 Data pathways and architecture for Fibre Channel operation
,
a
data transfer rates as well as impressive I/Os per
second.
The array is made up of dual SCSI or Fibre Channel
host interface cards, the intelligent midplane, a system
management card, and 12 independent disk drive sleds
containing up to 24 ATA disk drives.
ATA disk drives
ATA disk drives were known originally as Integrated
Drive Electronics (IDE), a low end disk interface. The
original IDE interface was low performance, single
threaded (no simultaneous I/O requesting), contained
minimal error detection and was unsuitable for
3
ATTO Technology Inc.
computer applications requiring high performance and
high reliability.
As IDE was refined and acquired important
capabilities, its name was changed to ATA, Advanced
Technology Attachment.
• UltraDMA transfer protocol similar to high
performance SCSI disk protocol operating at 66
MB/sec.
• Double-clocking of data transfers, doubling disk
data transfer rates
• CRC (Cyclic Redundancy Check) code allowing
full error detection and data reliability
• Multi-threaded I/O support
• Overlapped Command Support: allows
commands to be simultaneously active on
multiple drives on the same ATA bus.
• Command Queuing which allows simultaneous
multiple read/write commands to be sent to each
drive, reducing command overhead and
allowing the drive to service commands in the
most efficient manner: similar to the SCSI
feature Tagged Command Queuing.
• Faster drive speeds (5400/7200 RPM) with
higher media transfer rates
• A communication protocol and interface with a
fundamental lower cost structure than SCSI or
FC interfaces
ATA disk drives operate at performance and data
integrity levels similar to those that were previously
available only on SCSI or Fibre Channel disk drives.
ADXT Powers ATA to New Levels
The original notion of RAID was to build high
capacity, reliable storage subsystems using large
numbers of inexpensive disk drives. Thus its original
definition:
Over time that definition became
Independent Drives
inherent cost advantage in a
RAID system was lost.
Intelligent Midplane
heart of the Diamond Series
storage array is the
intelligent midplane with
ADXT
the data rates of individual
ATA disk drives to create
high data transfer rates. This
Redundant Array of Inexpensive Drives
Redundant Array of
and the
Disk 2
Disk 2
Disk 1
Disk 1
to sum or
The
aggregate
AIE2AIE3AIE4AIE5AIE6AIE7AIE8AIE9AIE10AIE11AIE
AIE
1
Disk 2
Disk 1
.
Disk 2
Disk 1
technology enables features such as serverless backup,
advanced error protection, metadata storage
techniques, virtualization software, thermal
management and advanced enclosure services.
The midplane contains a combination of custom
Application Specific Integrated Circuits (ASICs),
processors and proprietary embedded software
divided into three main processing sections which
handle the data being read or written to the ATA disk
drives from the Fibre Channel or SCSI host interfaces.
The Virtual Device Manager (VDM), Data Routing
Fabric (DRF) and ATA Interface Engines (AIE)
organize data streams for storage or retrieval.
Virtual Drive Manager
Data is accessed through
Virtual Drives using an implementation of the standard
SCSI protocol controlled by the Virtual Device
Manager.
Data Routing
Fabric incoming or outgoing data is
routed between the ATA Interface Engines (AIE) and
the Fibre Channel or SCSI interface by the custom
Data Routing Fabric ASIC, a high speed, low latency
transfer fabric with more than 2 GB/sec. of bandwidth
supported by up to 512 Megabytes of memory.
ATA Interface Engine (AIE)
The interface to each
pair of drives is through a custom ATA Interface
Engine ASIC. The AIE implements the typically
software-intensive ATA interface completely in
silicon. Each AIE contains a dedicated ATA protocol
processor to completely automate command and
protocol processing.
Disk 2
Disk 2
Disk 2
Disk 2
Disk 2
Disk 2
Disk 2
Disk 1
Disk 1
Disk 1
Disk 1
Disk 1
Disk 1
Disk 1
DATA ROUTING FABRIC
text
VIRTUAL DEVICE MANAGER
Disk 2
Disk 1
12
FIBRECHANNEL
INTERFACE 1
FIBRECHANNEL
INTERFACE 2
Diamond Storage Array Installation and Operation Manual
4
3.0 Installation Instructions
If you are familiar with the Diamond Storage Array, Fibre Channel, SCSI and RAID configurations, you
may set up and configure the array using these instructions. You will find details, illustrations and other
guidance for more involved operations and special cases in the rest of this manual.
CAUTIONCAUTION
Before configuring the Diamond Storage
Array, ensure that any computer data to be
stored on the array is properly backed up
and verified. The Manufacturer is not
responsible for the loss of any data stored
on the array under any circumstances and
any special, incidental, or consequential
damages that may result thereof.
Three major steps are required to set up the
Diamond Storage Array
1Physically set up Diamond Storage Array
2Connect to Ethernet
3Configure the drives
Step 1: physical setup
1Make sure the Diamond Storage Array is
mounted properly and has adequate air flow
around it.
2Insert the appropriate connector into the
interface card in the back of the Diamond
Storage Array.
3Connect the Fibre Channel or SCSI cable from
your host computer system to the connector. To
use ExpressNAV browser-based management
interface and configure your Diamond Storage
Array, you must connect to the Ethernet port.
4To set up the Ethernet connection: connect a
cross-over cable (for a direct connection to a
PC) or regular network cable from a network
device to the RJ45 Ethernet port on the
Ethernet management card on the front of the
Diamond Storage Array.
Step 2a: set up the Ethernet connection
The Diamond Storage Array supports service
operations over the RS-232 serial port using
standard terminal emulation software available
with most systems.
1Connect a DB-9 null modem serial cable
between the port marked RS-232 on the front of
the Diamond Storage Array and the computer’s
serial port. The cable must be no longer than
three meters.
2Make sure the power switches on the power
supplies on the rear of the unit are in the Standby position.
3Plug in the power cords to the back of the unit,
then into an appropriate power source (100-240
VAC).
4Reboot your host computer system.
5Press the Stand-by power switch for each
power supply on the Diamond Storage Array to
the ON position.
6Upon successful power up and diagnostics, the
unit displays the POST (power up self test)
information.
The Diamond is now in Command Line Interface
mode. You may modify the setup of the Diamond
Storage Array using the CLI (refer to
via ASCII-based Commands
on page 27), but the
CLI: Interface
easiest method to configure the array is by using
ATTO ExpressNAV software, a graphical user
management interface accessed through a
standard Internet browser. Refer to
ExpressNAV: Browser-based Interface
ATTO
on page 23.
Step 2b: connect to Ethernet
If the Diamond Storage Array is attached to a
DHCP server
1At the Ready prompt after POST (refer to Step
6 above), type set IPDHCP enabled
2Type SaveConfiguration Restart
3At the Ready prompt after POST (see above),
type get IPAddress
4Enter this address into your browser.
5The ATTO ExpressNAV screen appears. Log in
using the Telnet defaults:
Username: Telnet
Password: Diamond
6Follow the screens to find information about the
array or to configure the array from the factory-
5
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
default settings. The Diamond Storage Array
may be set up in a JBOD, RAID Level 0, 5 or 10
configuration with or without Hot Spare sleds
If the Diamond Storage Array is not attached to
a DHCP server and you wish to change the
defaults
1At the Ready prompt after POST (see above),
type set IPAddress [desired IP address].
2Type set IPSubnetMask [desired IP Subnet
Mask]
3Type set IPGateway [desired IP Gateway]
4Type SaveConfiguration Restart to save the
configuration and restart the Diamond Storage
Array
5After the powerup and POST complete, type
the IP address from step 1 above into your
browser.
6The ATTO ExpressNAV screen appears. After
logging in (refer to Step 2b: connect to Ethernet, Step 5 on page 5), follow the screens
to find information about the array or to
configure the array from the factory-default
settings. The Diamond Storage Array may be
set up in a JBOD, RAID Level 0, 1, 5 or 10
configuration with or without Hot Spare sleds.
Step 3: configure drives
The simplest way to set up configurations is to use
the ExpressNAV interface. Refer to
ExpressNAV: Browser-based Interface
ATTO
on page 23 for
more information on the interface. After logging
in, follow the screens to find information about
the array or to configure the array from the
factory-default settings.
You may the CLI to set up RAID, partitions and
zones.
Note
All arrays using RAID level 10 and Hot Spare
sled options must be fully populated.
6
Quick start
3.1 Components
The Diamond Storage Array has been designed to be easy to use, maintain and upgrade. It features a
durable steel outer case and modular components in either a floor or a rack mount model.
Immediately upon receipt, check the shipping
carton for damage from mishandling. Contact us
at once via the means that is easiest for you (refer
to
Warranty
on page xvi) if the carton has been
mishandled or displays any signs of damage.
The front of the Diamond Storage Array provides
access to the management card and disk drive
sleds. The rear of the unit holds the host interface
cards, power supplies and blower assemblies.
CAUTIONCAUTION
All modular components must be replaced
by qualified personnel only.
Floor model
The management system card is at the top front of
the case. At its center is a DB-9 serial RS-232
port, a connection for setup, monitoring and
upgrade of the unit from any computer system
with an RS-232 interface. The optional 10/100
BaseT Ethernet management services card
enables Telnet-based monitoring and
management. It
also provides
the ability to
update the
firmware in the
array via FTP.
LEDs to the
port’s right
indicate fault,
unit ready, host
interface cards
A and B
installation
status, and the
power status for
each power
supply.
Below the management system card are
individual disk drive sleds which also have LEDs
for each drive’s status. Each sled contains two
hard drives. Up to 24 hard drives may be installed
on the 12 sleds. Empty bays should be covered by
blank faceplates or empty sleds. Access is
provided by loosening two screws and gently
pulling on the sled handle.
On the rear of the unit are blowers which support
hard drive, cabinet and power supply cooling. The
blowers are held in by removable screws. Correct
operation is displayed by a LED at the top of each
panel.
The power supplies for the array, also in the rear
of the unit, are accessible by loosening two screws
and pulling on the power supply module handle.
The power standby on/off switch is at the top of
each module. A yellow LED indicates
and a green LED indicates
on
. The power cord
caution
socket is at the bottom of each power supply.
Between the power supplies and blower
assemblies are two slots that hold the Host
Interface cards. The HIC is the connection point
into the array and is available in three options: 1Gigabit Fibre Channel, 2-Gigabit Fibre Channel,
or Ultra 160 SCSI. Host Interface cards have
and
on-line
or
fault
and
activity
LED indicators,
fault
depending on the model.
SCSI
The SCSI card faceplate has a rotary
binary-coded hex switch to set the SCSI ID of the
array. The SCSI card also has an in channel, to
connect via cable to the unit’s communication
source, and an out channel, available for daisychaining arrays together or to complete
termination using an external LVD terminator.
Rack mount
The system management card is at the left front of
the case. At its center is a DB-9 serial RS-232 port
7
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
which allows a connection
for setup, monitoring and
upgrade of the unit from
System Management Card
any computer system with
an RS-232 interface. The
optional 10/100 BaseT
Ethernet management
services card enables
Telnet-based monitoring
and management. It also
provides the ability to
update the firmware in the
Disk drive sleds (12)
array via FTP. LEDs
farthest to the left indicate
Power supplies
fault, unit ready, Host
Interface cards A and B
installation status, and the
power status for each
power supply.
To the right of the system
management card are
individual disk drive sleds
which also have LED
indicators for each drive’s
status. Each sled contains two hard drives. Up to
24 hard drives may be installed on the 12 sleds.
Empty bays should be covered by blank
faceplates or filled with empty sleds to promote
effective cooling. Access is provided by
loosening two screws and pulling on the sled
handle.
In the rear of the unit are the blower assemblies
which support hard drive, cabinet and power
supply cooling. Correct operation is displayed by
a LED at the top of each panel. The blowers are
held in place by removable screws.
The power supplies for the array are accessible by
loosening two screws and pulling on the power
supply module handle. The power standby on/off
switch is at the top of each module. A yellow LED
indicates
caution
and a green LED indicates on.
Intelligent midplane (internal)
Host interface cards
Blower assemblies
The power cord socket is at the bottom of each
power supply.
Between the power supplies and blower
assemblies are two slots that hold the Host
Interface cards. The HIC is the connection point
into the array and is available in 2-Gigabit Fibre
Channel, or Ultra 160 SCSI. Host Interface cards
fault
and
have
on-line
or
fault
and
activity
LED
indicators, depending on the model.
SCSI
The SCSI card faceplate has a rotary
binary-coded hex switch which allows you to set
the SCSI ID of the array. The SCSI card also has
an in channel, to connect by a cable to the unit’s
communication source, and an out channel,
available for daisy-chaining arrays together or to
complete termination using an external LVD
terminator.
8
Components
3.2 Physical Setup
The Diamond Storage Array is shipped completely assembled with two 120 VAC power cords for use in
the United States and Canada.
Immediately upon receipt, check the shipping
carton for damage from mishandling. Contact us
at once by the means easiest for you (refer to
Warranty
on page xvi) if the carton has been
mishandled or displays any signs of damage.
Floor Model
The Diamond Storage Array is heavy (about 92
pounds fully loaded) and requires two people to
lift and carry it safely. Place the array on a level
surface and make sure there is adequate space in
the front and back of the unit for proper cooling
and airflow. Continue with the general
instructions.
Rack Mount
The array fully loaded is heavy (about 86 pounds).
The unit should be handled with care and requires
two people to lift, carry and/or install it safely.
The array can be mounted via several different
methods in a 19” rack with 3U (5.25”) of vertical
space required. Air flow should not be restricted
in any way.
Installation in a rack may create a differential
between the room ambient temperature and the
internal ambient temperature in the rack. While
the maximum internal operating temperature of
the array is 47°C,
you should not
run the system at
the maximum
temperature for
extended
periods.
Therefore,
ensure that the
room ambient
temperature is
kept below 30°C
for best operation.
Each side of the rack mount array chassis has
three pairs of mounting holes. One is located near
the front of the rack, one near the unit’s center of
gravity, and one near the rear of the rack. The
holes accommodate 10/32 screws but the screws
can protrude no farther than .375 inches into the
rack.
Spaced rail pairs in some rackmount cabinets.
You can mount the array using two sets of rail
pairs spaced to accommodate the overall length of
the unit (approximately 23 inches). Mount using
the rack mount brackets on the front and rear of
the unit fastened to the rail pairs using 10/32 pan
head screws with lock and flat washers.
Rack mount cabinets with stationary shelf or tray
system.
The shelf or tray must be able to support
at least 125 pounds. The shelf or tray must be
installed and secured to the rack before installing
the array. Secure the front of the array to the rack
with 10/32 screws, locks and flat washers.
Sliding shelf or tray type systems
should never
be used under any circumstances.
Two point open rack system.
The rack must be
strong enough to support the array properly.
Mounting brackets should be moved to the
9
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
centermost mounting holes and secured using
10/32 screws.
CAUTIONCAUTION
Do not mount multiple arrays on a two-rail
rack or mount the array above the midpoint
of a two-rail rack system. Do not mount the
array on any kind of rail-type system. The
array is too heavy and does not have the
proper hole pattern for rails.
Note
Insure the array has adequate air flow.
General Instructions
1Insert the proper connector into the Host
Interface Card in the back of the array. (refer to
Power Sw itch Posit ions
On Stand-by
2Connect the cable (Fibre Channel or SCSI)
from your host system to the Host Interface
Card connector on the back of the array. The
cable you use depends upon your application,
the environment and distance.
Connecting a Fibre Channel
on page 11 for Fibre
Array
Channel and
SCSI Array
SCSI).
Connecting a
on page 13 for
3Make sure the power switches on the power
supplies on the rear of the unit are in the standby position. Plug in the power cords to the back
of the unit, then into an appropriate power
source (100-240 VAC). The power source must
be connected to a protective earth ground and
comply with local electrical codes. Improper
grounding may result in an electrical shock or
damage to the unit.
4Press the stand-by power switch for each
power supply to the ON position. When the
green power LED on the back of the unit is lit,
the power supply is fully operational and
delivering power to the system. The power LED
on the front of the array lights while the
firmware executes.
When the power is turned on, the LEDs on the
front of the array flash twice. Drives spin up in
groups of three every one to two seconds. The
individual LEDs blink. After all available drives
have spun up, the individual drive LEDs stay lit.
When all available drives are operational, the
ready LED on the top front panel of the
management card remains lit.
5Reboot your computer
6Determine the best configuration for your needs
(i.e. JBOD, RAID, etc.) and refer to the rest of
this manual for configuration information.
Exhibit 3.2-1 Back side of a rack mount array.
10
Physical setup
3.2.1 Connecting a Fibre Channel Array
The Diamond Storage Array supports up to two Fibre Channel Host Interface Cards (HIC). Physical
connections and CLI commands contribute to the Fibre Channel topology.
The cable you use depends
upon your application, the
environment and the
distance required for your
2 Gb HIC uses
2 SFPs to connect up
to 2 FC cables
storage area network.
To comply with FCC and
CE EMI for the 2-gigabit
Host Interface Card, use fiber optic cables.
Exhibit 3.2-1 Fibre Channel cable options
Cable lengthCable typeCable size
<10 metersUnequalized copper
>10 <30 metersEqualized copper
Up to 175 metersmulti mode fiber optic62.5
Up to 500 metersmulti mode fiber optic50 micron
Up to 10
kilometers
single mode fiber optic 9 micron
micron
The Diamond Storage Array may have two Fibre
Channel Host Interface Cards (HIC). In 2 gigabit
Fibre Channel arrays, each HIC is connected by a
Fibre Channel cable via a SFP (small form factor
pluggable) module into a point-to-point or loop
Fibre Channel topology.
Install the SFP according to the manufacturer’s
instructions.
Each HIC has two ports and an on-board hub.
Each port has an SFP module to connect to Fibre
Channel. Each HIC is independent of the other, so
that one may be connected into a point-to-point
topology and the other into a loop. However, if
one port in a HIC is connected into a point-topoint topology, the other port cannot be used.
Autoconfiguration
The array automatically determines which HICs
are installed and if they are in loop or point-topoint topologies.
If you wish to see how the unit has been set up, go
to the CLI commands and type
Info
or go to the
Status page of the ExpressNAV interface (refer to
ATTO ExpressNAV: Browser-based Interface
on page
25. The return displays the data rate and
connection mode for each HIC (FC 0 and FC 1).
Manual configuration
You may manually configure the array using CLI
commands (refer to
Commands
Channel
(refer to
on page 35.) or access the
page of the ExpressNAV interface
ATTO ExpressNAV: Browser-based Interface
Fibre Channel Configuration
Fibre
on page 25).
•
FCConnMode
topology for both HICs on an array. Options are
loop only (loop), point-to-point only (ptp), loop
preferred (loop-ptp) or point-to-point preferred
(ptp-loop).
Loop connects to either an FC-AL arbitrated loop or
a fabric loop port (FL_Port) on a switch.
Point-to-point (ptp) connects to a direct fabric
connection, such as an F port on a switch.
Loop-ptp allows the array to determine what kind of
connection to use, but tries to connect in loop mode
first, then point-to-point mode.
Ptp-loop allows the card to determine what kind of
connection to use, but tries to connect in point-topoint mode first, then loop mode.
•
FcDataRate
at which both HICs on a Diamond operate.
Choices are 1 gigabit, 2 gigabit and
autodetection.
specifies the Fibre Channel
specifies the Fibre Channel data rate
One of the advantages of using loop topology for
Fibre Channel connections is that it allows arrays
to be daisy-chained together.
11
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
Exhibit 3.2-2 Possible 2 gigabit Fibre Channel physical connections depending on which Fibre Channel connection
mode has been selected.
point-to-
point mode
loop
mode
Diamond Storage Array ADiamond Storage Array B
no
connection
possible
loop
mode
daisy-chain
loop
mode
loop
mode
loop
mode
loop
mode
12
Fibre Channel connections
3.2.2 Connecting a SCSI Array
The SCSI Diamond Storage Array uses a VHDCI connector and SCSI cables to connect to a host. It
automatically detects the type of Host Interface Card it is using without any intervention.
To connect the SCSI Diamond Storage Array
1Insert a SCSI VHDCI connector into the Host
Interface Card in the back of the array. If the
SCSI array is the last device on the bus, you
must attach a VHDCI terminator to one
connector of the SCSI Host Interface Card or
connect a cable between the second connector
and the next device on the SCSI bus.
Exhibit 3.2-1 SCSI cable options.
Bus speed, MB/sec.
max.
SCSI-1586-8
Fast SCSI1086-8
Fast Wide SCSI20166-16
Wide Ultra SCSI40163-4
Wide Ultra SCSI40161.5-8
Wide Ultra 2 SCSI8016-1216
Ultra 3 or Ultra160
SCSI
16016-1216
Bus width,
2The SCSI Host Interface Card has a rotary
binary-coded hex switch which allows you to
set the SCSI ID of the HIC. Be sure the selected
ID is different from all other SCSI devices on
the bus.
Note
If slower devices are connected on the same
SCSI bus as the Ultra 160 array, the bus
communicates at the rate of the slowest
device.
Max. bus lengths,
bits
meters
Single-endedLVD
Max. device
support
Exhibit 3.2-2 SCSI interface cards: left without terminators attached; right with a terminator attached.
13
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
14
SCSI connections
3.3 Determining Drive and Sled Designations
g
The Diamond Storage Array has been designed with 12 sleds, each holding two drives. The easiest way
to configure an array is to use all the drives on all the sleds. The firmware uses a numbering system to
determine which drives and sleds it is affecting.
All sled slots should be filled contiguously,
starting with the first slot next to the management
card.
When configuring a array with fewer than 12
drive sleds, you must consider several factors:
RAID level, number of physical drives/sleds and
the end configuration you are trying to achieve.
Review the information about each configuration
to determine how each configuration would be
affected by using fewer sleds.
Numbering conventions
The Diamond Storage Array with firmware
version 3.1 and newer uses a unique numbering
convention to orient its drives and sleds to the
controlling firmware. Older versions do not use
this convention. (refer to
Updating Firmware
on
page 75 for information about updating the array
firmware.)
Physical
refers to the physical drives in the array,
the hardware that actually exists in a physical
sense.
Logical (or virtual)
refers to what the host
operating system recognizes as an entity. Two
physical drives may be seen as one logical drive
by the operating system.
Logical disks do not always map one-to-one with
physical disks.
In RAID configurations, for
example, several physical disk drives (or portions
of several physical drives) are grouped into a
logical disk, call a RAID Group or a Logical Unit
1 2 3 4 5 6 7 8 9 10 11 12
Disk 2
Disk 2
Disk 2
LUN 13
LUN 14
LUN 15
Disk 2
LUN 16
Rack Mount Drive Sleds
LUN 17
Disk 2
LUN 16
LUN 17
Disk 2
LUN 18
Disk 2
LUN 19
LUN 20
LUN 19
(LUN). Each RAID group is broken into logical
blocks of 512 bytes each, numbered 0 through n
(the Logical Block Number or LBN). A 100 GB
LUN has approximately 200,000,000 logical
blocks.
A RAID group is also referred to as a Virtual
Drive.
A LUN is associated with a RAID group or
Virtual Drive unless you are using partitions. If
you have partitions, a LUN is associated with
each partition. A RAID Group or Virtual Drive
may then have multiple partitions or LUNs.
Physical numbering
The Diamond RAID Storage Array contains
• Up to 24 physical disk drives
• Two drives mounted on 12
physical drive sleds
• Sleds are numbered 1-12,
starting at the top (floor units)
or the left (rackmount units).
• Each sled is connected to its
own internal ATA bus with
two disk drives, numbered 1
and 2.
• Two green LEDs, labeled
Drive 1 and Drive 2, indicate
activity for the two drives.
They remain solidly lighted
when there is no activity.
Disk 2
Disk 2
Disk 2
Disk 2
LUN 20
LUN 21
LUN 22
LUN 23
Disk 2
LUN 24
Mana
Drive 1
LUN 1
Drive 1
LUN 2
Drive 1
LUN 3
Drive 1
LUN 4
Drive 1
LUN 5
Drive 1
LUN 6
Drive 1
LUN 7
Drive 1
LUN 8
Drive 1
LUN 9
Drive 1
LUN 10
Drive 1LUN 6
ement car d
Drive 2
LUN 16
Drive 2
LUN 17
LUN 18
Drive 2
LUN 19
Drive 2
LUN 20
Drive 2
LUN 21
LUN 22
Drive 2
LUN 13
Drive 2
LUN 14
Drive 2
LUN 15
Drive 2
LUN 20
LUN 19Drive 1LUN 12
Drive 2
LUN 17
1 2 3 4 5 6 7 8 9 10 11 12
Floor Model Drive Sleds
Management Card
15
Disk 1
LUN 1
Disk 1
LUN 2
Disk 1
LUN 3
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
Disk 1
LUN 4
Disk 1
LUN 5
Disk 1
LUN 6
Disk 1
LUN 7
Disk 1
LUN 8
Disk 1
LUN 9
Disk 1
LUN 10
Disk 1
LUN 11
Disk 1
LUN 12
Drive 1
LUN 11
Drive 1
LUN 12
Drive 2
LUN 23
Drive 2
LUN 24
Logical Numbering
Logical numbering depends on the RAID
configuration of the storage array. Current
firmware includes RAID Level 0, RAID Level 5,
RAID Level 1, RAID Level 10, Zoning and
Partitioning capability and hot spare sleds. The
default configuration is QuickRAID0 0 or JBOD
(Just a Bunch of Disks), in a single zone.
Examples
JBOD
mode uses 24 LUNs. Each LUN is
equivalent to one physical drive. The array can
operate with several sleds missing, but the empty
sled(s) are treated as offline and cannot be
configured.
RAID Level 0 (QuickRAID0 [n])
With a fully
populated array, RAID 0 may be configured as 1,
2, 3, 4, 6, or 12 LUNs. As RAID0 1, all 24
physical drives are configured as a single stripe
group or LUN. You may also configure two LUNs
of 12 drives each, three LUNs of eight drives
each, four LUNs of six drives each, six LUNs of
four drives each and 12 LUNS of two drives each.
(See Exhibit 5.2-3 in
RAID Level 0
on page 52). The
command assumes there are 24 drives available to
configure the specified number of LUNS.
RAID Level 1 (QuickRAID1)
has no options: the
array can be configured into six groups, LUNs 16, with each LUN containing two physical sleds;
each sled in the LUN would be a mirror image of
the other sled in the LUN.
RAID Level 10 (QuickRAID10)
The
QuickRAID10 command first creates six
mirrored groups, then stripes them into groups of
one, two or three RAID 10 groups. Each group is
a LUN. Issuing QuickRAID10 2, the 12 physical
sleds are configured as six mirrored pairs, then
striped into two LUNs.
RAID Level 5 (QuickRAID5)
with a fully
populated Array may be configured as 1, 2, 3 or 4
LUNs. As QuickRAID5 1, all 24 physical drives
are configured as a single RAID 5 LUN. You may
also configure two LUNs of 12 drives each, three
LUNs of eight drives each, or four LUNs of six
drives each. RAID Level 5 parity reduces the
usable capacity of each LUN by the equivalent of
one drive sled.
Unique serial number for each LUN
Each LUN in a system has a unique 24-character
serial number which is updated when a system
configuration changes. It is stored in a Device
Association Table on each drive.
When a drive or sled is replaced in QuickRAID
configurations that allow for hot swap, a new
serial number is computed and is stamped onto all
member drives of the RAID group. The CLI
command
LUNInfo
the ExpressNAV interface
or the
Logical Units page of
displays the serial
number for each drive.
Exhibit 3.3-1 Format of the 24 characters of the unique serial number for each LUN
Character Placement Valid values
0-19Any ASCII char Drive Serial Number
201 for RAID1
A for RAID10
0 for RAID0
5 for RAID5
X for JBOD
21A – ZLUN Configuration Iteration Character: starts at A and increments every time
22A – ZGeneration Character: starts at A and increments every time a new
230Reserved for future use
Example of unique serial number for a JBOD configuration:
1231231231231231231XAC0
RAID Configuration Character
a member drive is replaced in a RAID Group. Reverts back to A any time the
Generation Character is updated
configuration is stamped on the system. When it reaches Z, rolls over to A.
16
Drive, sled numbering
4.0 Accessing the Array
Communicate with the Diamond Storage Array through the Fibre Channel link, the RS-232 port or the
Ethernet port using Command Line Interface commands or ATTO ExpressNAV,
management console
.
an integrated user
You may configure and tune the Diamond Storage
Array for different environments and
applications, update the firmware, monitor
internal power and temperature status, report on
hardware diagnostics and log failures.
Three avenues are available:
• In-band SCSI over Fibre Channel and over SCSI
• RS-232 port
• Telnet and SNMP over Ethernet
The following chapters describe how to access the
array and use the Command Line Interface or
ATTO ExpressNAV, an integrated user
management console.
Command Line Interface
The CLI provides access to the array through
ASCII command lines.
An initial display, after powering up the unit or
restarting the firmware, contains the information
in Exhibit 4.0-1. Once the initial display is
complete, with the word
Ready
, you are in the
CLI mode.
Help
Type
to display a list of all commands
available.
ATTO ExpressNAV
ATTO ExpressNAV is an integrated configuration
tool accessible through an Ethernet connection.
Platform independent, ExpressNAV contains all
the current capabilities of the CLI in a userfriendly GUI console. A menu on each page
provides access to information and configuration
operations.
Refer to
ATTO ExpressNAV: Browser-based Interface
on page 25 for more information on the program.
In-band SCSI over Fibre Channel
In-band SCSI commands (
) may be issued to the array to manage
Buffer
Write Buffer
and
Read
configuration via two mechanisms:
• In-band CLI over SCSI, where ASCII CLI
commands, may be issued via
CLI commands except
•
ID/value
SCSI CDB (command descriptor block) to select
the buffer ID of the configuration parameters to
be affected, and the new value of the parameter.
Most configuration options are available.
, where the application program uses a
Zmodem
Write Buffer
are supported.
. All
RS-232 port
The array provides remote service operations over
the RS-232 serial port using standard terminal
emulation software available with most systems.
Set the following serial parameters in your
terminal program:
• Bits per second: 115200
• Data Bits: 8
• Parity: None
• Stop Bits: 1
• Flow Control: None.
• Terminal type: ASCII
• Echo: on.
Ethernet port
The 10/100 BaseT Ethernet port provides Telnetor SNMP-based monitoring and management.
The default IP address is 10.0.0.1; the default
subnet mask is 255.255.0.0. To change the
defaults, first be configure the array for the
network using the RS-232 port to establish the
correct IP address. The management port provides
TCP/IP-based monitoring and management.
17
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
SNMP
SNMP, or Simple Network Management
Protocol, is an application layer protocol that
allows network devices to exchange management
information. Through a combination of standard
and custom MIBs (Management Information
indications to an SNMP server, allowing the array
to be managed with other devices in a complex
system through a common interface.
Use CLI commands to configure up to six unique
Trap addresses. A trap is a way for the array to tell
the SNMP server that something has happened.
Base), the array provides status and error
Exhibit 4.0-1 The POST information displays after boot of the Diamond Storage Array.
Firmware version 5.40 release date Mar 30 2004, 10:43:06 Build 021G
Power-On Self-Test Completion Status: GOOD
128 Megabytes of RAM Installed.
Interface Port 0 is not installed.
Interface Port 1 is 1.0624 Gb/s Fibre Channel.
Interface 0 World Wide Name = 20 00 00 10 86 10 02 DC
Interface 1 World Wide Name = 20 00 00 10 86 10 02 DC
Diamond Array Serial Number = "MIDP100197"
Diamond Array Name = " "
System Serial Number = ""
Active Configuration = ATTO
DiamondClass = (V)86
Internal Temperature = 23 C [5 - 47]
ErrorLog Contents: NO ERRORS
For help, type HELP.
Ready.
18
4.1 In-band CLI Over Fibre Channel
In-band Command Line Interface (CLI) configures and manages the Diamond Storage Array using
SCSI-based CLI commands over a Fibre Channel port connection.
In-band CLI allows a programmer to configure
the Diamond Storage Array while it is moving
data. Using a programmer’s interface, CLI
commands as described previously in this manual
may be implemented.
In-band CLI is implemented as part of LUN 0. It
uses a different LUN than the array, and reports as
a Storage Enclosure Services (SES) device
(device type 0x0D).
LUN 0 is visible on all Fibre ports but is actually
a single unit. The default value for LUN 0 is 0x00.
LUN 0 must be reserved for each Write
Buffer/Read Buffer pair, using the SCSI Reserve
command to insure integrity of the in-band CLI
session.
1An initiator (host) sends a SCSI Reserve
command to LUN 0.
• If LUN 0 is not reserved by another initiator,
LUN 0 is now reserved and available to begin
a new CLI session.
• If the array configuration is reserved by a
different CLI session (i.e. serial or Telnet),
the in-band session does not allow
modifications of the array configuration. If
you try, the results buffer of LUN 0 returns:
Process X has the configuration
reserved.
ID of this session = Y
Ready.
2The initiator issues a SCSI Write Buffer
command to LUN 0. A
must be accompanied by an ASCII buffer
representing the CLI command string such as
set DiamondName Omega1
LUN 0 executes the command line and create
feedback in the form of ASCII characters into a
buffer. This buffer is 8 KB and circular. Retrieve
the results by issuing a
before issuing another
Write Buffe
Read Buffer
Write Buffer
r command
command
command.
3A subsequent
the new command line and overwrite the
previous results in the buffer with new results.
4LUN 0 can be released by issuing a SCSI
Release
Write/Read Buffer pair, or multiple Write/Read
Buffer pairs.
Initiator (Host)Diamond Storage Array
Reserve LUN 0return: “ok”
Write Buffer
LUN 0
bid ‘AA’
“get Temperature”
Read Buffer
LUN 0
bid ‘AA’
Release LUN 0return: “ok”
Write Buffer
command to the LUN after each
command executes
executes the CLI command,
stores output in buffer
return:”Temperature=28C\r\n\
Ready.\r\n\0”
I/O details
The buffer sent to the Services LUN during the
data out phase of a Write Buffer command must
be:
• ASCII data
• maximum 80 bytes length
• terminated with either a carriage return
character (0x0D), line feed character (0x0A) or
NULL character (0x00)
• Characters following the first carriage return
character, line feed character or NULL character
are ignored.
The buffer retrieved from the Services LUN
during the data-in phase of a
Read Buffer
command:
• ASCII data
• 8 KBytes (8192 bytes) in length
• terminated with a NULL character (0x00)
• Characters following the NULL character are
meaningless.
19
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
Loading...
+ 87 hidden pages
You need points to download manuals.
1 point = 1 manual.
You can buy points or you can get point for every manual you upload.