Three major steps are required to set up the Diamond Storage Array
Step 1: physical setup
Step 2a: set up the Ethernet connection
Step 2b: connect to Ethernet
If the Diamond Storage Array is attached to a DHCP server
If the Diamond Storage Array is not attached to a DHCP server and you wish to change the
defaults
Step 3: set up RAID configuration
Status
Ethernet
SNMP
Serial Port
Fibre Channel
Storage Management
RAID
Clear Data
Logical Units
Rebuild
Configuration
Advanced
To use the advanced page CLI commands
Restart
Help
5.5 Command Line Interface Use and Guidance .....................29
5.5.1 Summary of CLI Commands .................................31
5.5.2 General Use Commands ........................................33
FirmwareRestart
Help
RestoreConfiguration
SaveConfiguration
SystemSN
VerboseMode
Configuring a fully-populated array
To set up one RAID Level 5 group with one Hot Spare sled
To set up two RAID Level 5 groups with two Hot Spare sleds
Configuring a partially-populated array
To set up one RAID Level 5 group with one Hot Spare sled
Removing RAID groups
To remove a management card
To remove a disk drive sled (Exhibit 7-2)
To remove a Host Interface Card
To remove the power supply (Exhibit 7-4),
To remove a blower assembly (Exhibit 7-4),
7.1 Hot Swap Operating Instructions .......................................71
Disk drives
Power supplies
Blower assemblies
To replace a blower assembly
7.2 Optional Hot Spare Sled ......................................................73
To set up RAID Level 1 with Hot Spare sleds
To set up RAID Level 10 with Hot Spare sleds
To set up one RAID Level 5 group with one Hot Spare sled
To set up two RAID Level 5 groups with Hot Spare sleds
Updating firmware using the RS-232 serial port
To update firmware
Updating firmware using the optional Ethernet card
10.0 System Monitoring and Reporting ..............................................79
RS-232 monitoring port and CLI
Ethernet monitoring port and CLI
Power On Self Test (POST)
Ready LED
Audible alarm
Thermal monitoring
Power supply monitoring
System fault LED and error codes
Disk drive activity and disk fault LEDs
Windows 2000 special instructions
Error messages
Specific situations and suggestions
If a drive fails to respond
If a power supply fails
To determine if the problem exists with the HIC
If you can’t access the CLI via Ethernet
If LUNs do not display on the host
Manufacturer limited warranty
Contact ATTO Technology, Inc.
Preface
This guide will take the technology-savvy user through the installation and maintenance of the Diamond Storage
Array.
The Diamond Storage Array was designed to meet
your need for large amounts of easily accessible
storage using proprietary Aggregated Data Transfer
Technology (ADXT
multiple, high performance ATA disk drives with the
sustained data transfer rates required by sophisticated
computer users.
Your comments help us improve and update our
products. Contact us
J
) to merge the power of
Disclaimer
ATTO Technology, Inc.
155 CrossPoint Parkway
Amherst, New York 14068
(716) 691-1999 • voice
(716) 691-9353 • fax
http://www.attotech.com/diamond
ATTO Technology can also be reached via e-mail at
the following addresses
Sales Support
Technical Support
:sls@attotech.com
: techsupp@attotech.com
Although reasonable efforts have been made to assure
the accuracy of the information contained herein, this
publication could include technical inaccuracies or
typographical errors. Manufacturer expressly
disclaims liability for any error in this information and
for damages, whether direct, indirect, special,
exemplary, consequential or otherwise, that may result
from such error including but not limited to loss of
profits resulting from the use or misuse of the manual
or information contained therein (even if
Manufacturer has been advised of the possibility of
such damages). Any questions or comments regarding
this document or its contents should be addressed to
Manufacturer.
Manufacturer provides this publication as is, without
warranty of any kind, either express or implied,
including, but not limited to, the implied warranties for
merchantability or fitness for a particular purpose.
Information in this publication is subject to change
without notice and does not represent a commitment
on the part of Manufacturer. Changes may be made
periodically to the information herein; these changes
will be incorporated in new editions of the publication.
Manufacturer reserves the right to make
improvements and/or changes at any time in
product(s) and/or program(s) described in this
publication.
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
1.0 Diamond Storage Array Product Overview
The Diamond Storage Array offers up to 24 ATA disk drives in a rack mount or floor model
configuration.
The Diamond Storage Array is ideally suited for
data intensive applications such as web hosting, email servers, on-line transaction processing,
digital video editing, medical imaging and digital
audio editing. Virtually any high performance
computing system with a growing need for
storage capacity can use the power of the
Diamond Storage Array.
With the cost effective approach of using ATA
disk drives, you can add more storage capacity as
your needs grow without the budget-busting costs
of other disk storage technologies. You can also
improve the performance and capacity of the
Diamond Storage Array cabinet as technology
progresses by simply replacing disk drive sleds
and Host Interface Cards.
The Diamond Storage Array is operating system
independent and supports all popular computer
hardware platforms and network environments
with a 2-gigabit Fibre Channel interface or an
Ultra160 SCSI interface.
The Diamond Storage Array is a fully populated,
turnkey solution with drives pre-installed. It is
fully supported by highly trained customer
service and engineering staffs.
The Diamond Storage Array uses Aggregated
Data Transfer Technology (ADXT
J
) to merge
the performance of multiple ATA drives to
achieve sustained, full bandwidth data transfer
rates. ADXT provides end users with the power
and sophisticated data control needed to take
moderately priced ATA disk drives, combine
them in a disk storage array, and power them to
the performance levels of SCSI or Fibre Channel
disk arrays.
Diamond Storage Array features
• Up to 7.2 Terabytes initial configuration
(expandable with future drive technology)
• 24 ATA disk drive capacity
• Aggregated Data Transfer Technology
(ADXTJ) for high performance/scalability
• Ultra ATA 100 megahertz (Mhz)
1
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
• Staggered drive spin-up to reduce peak power
demand
• Tagged command queuing to process up to 255
simultaneous data requests
• RS-232 management for local management
control; Ethernet option available for setup
connection only
• ATTO ExpressNAV
for management and information
• Two power supplies capable of 85-264 V (rated
100-240V AC) operation (340 watts each)
• UL, TUV and CE marked and compliant
• Internal thermal and power management
• Redundant hot swappable power supplies with
integrated thermal and power management
• Hot spare sleds: replace degraded sleds with
spares on the fly using software
• Floor model cabinet or 19” 3U rack mount
Exhibit 1.0-1 Back of rack mount model, Diamond Storage Array. Left: Fibre Channel. Right: SCSI.
TM
browser-based interface
Fibre Channel model
• 2 gigabit Fibre Channel Port (single or dual
channel)
• SFP-based Fibre Channel interface supports
long wave and short wave optical cables
• Built-in hub for daisy-chaining
• Up to 9,500 I/Os per second per port
• Up to 180 MB/sec. sustained Fibre Channel
transfer rates per interface
SCSI model
• Ultra 160 SCSI bus.
• Dual stacked VHDCI connectors for daisychaining and termination
• SCSI Target ID selection switch
• Support for single-ended and LVD SCSI
• No onboard termination
Product overview
2
2.0 Diamond Storage Array Technical Overview
The Diamond Storage Array uses Aggregated Data Transfer Technology (ADXT) to achieve the high
data transfer performance you need. ADXT
achieve sustained, full bandwidth Fibre Channel data transfer rates.
merges the performance of multiple ATA drives together to
Unlike other storage arrays which use expensive
SCSI or Fibre Channel disk drives to achieve
performance, the Diamond Storage Array uses
lower cost ATA disk drives combined with an
intelligent midplane to create a storage array with
exceptional price and performance
characteristics.
The intelligent midplane contains hardware and
software which provide the proprietary ADXT
Exhibit 2.0-1 Data pathways and architecture for Fibre Channel operation
,
switched data management and data movement
technology, to create a higher performance
storage solution.
The Diamond Storage Array is made up of dual
SCSI or Fibre Channel Host Interface Cards, the
intelligent midplane, a system management card,
and 12 independent disk drive sleds containing up
to 24 ATA disk drives.
a
ADXT Powers ATA to New Levels
The original notion of RAID was to build high
capacity, reliable storage subsystems using large
numbers of inexpensive disk drives. Thus its
original definition:
3
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
Redundant Array of
Inexpensive Drives
became
and the inherent cost advantage in a RAID system
was lost. The Diamond Series provides a high
capacity, high performance and highly reliable
Redundant Array of Independent Drives
. Over time that definition
disk array that uses the merged power of ATA disk
drives.
enclosure management services, serverless
backup, data replication and LUN mapping.
Intelligent Midplane
The heart of the Diamond
Series disk storage array is the intelligent
midplane with ADXT
to sum or
aggregate
the
data rates of individual ATA disk drives to create
high data transfer rates. This technology enables
features such as serverless backup, advanced error
protection, metadata storage techniques,
virtualization software, thermal management and
advanced enclosure services.
The midplane contains a combination of custom
Application Specific Integrated Circuits (ASICs),
processors and proprietary embedded software. It
is divided into three main processing sections
which handle the data being read or written to the
ATA disk drives from the Fibre Channel or SCSI
host interfaces. The Virtual Device Manager
(VDM), Data Routing Fabric (DRF) and ATA
Interface Engines (AIE) organize data streams for
storage or retrieval.
Virtual Device Manager
Data is accessed through
virtual devices using an implementation of the
standard SCSI protocol controlled by the Virtual
Device Manager. The VDM adds capabilities
such as RAID, data management services,
Data Routing Fabric
Incoming or outgoing data is
routed between the ATA Interface Engines (AIE)
and the Fibre Channel or SCSI interface by the
custom Data Routing Fabric ASIC, a high speed,
low latency transfer fabric with more than 2
GB/sec. of bandwidth supported by up to 512
Megabytes of memory.
ATA Interface Engine (AIE)
The interface to each
pair of drives is through a custom ATA Interface
Engine ASIC. The AIE implements the typically
software-intensive ATA interface completely in
silicon. Each AIE contains a dedicated ATA
protocol processor to completely automate
command and protocol processing. Automated
command overlapping and queuing maximizes
the performance of multi-threaded I/O. The AIE
transfers data using double-clocking technology
at 66 megabytes per second. The Diamond
Storage Array contains 12 AIEs for a parallel
transfer rate of 792 megabytes per second. In
addition to guaranteeing data transfer integrity
with automatic CRC checking, the command and
status transfers are validated using a patent
pending technology unique to the AIE.
Disk 2
Disk 2
Disk 2
Disk 2
Disk 2
Disk 2
Disk 2
Disk 2
Disk 2
Disk 1
Disk 1
Disk 1
Disk 1
Disk 1
Disk 1
Disk 1
Disk 1
Disk 1
AIE2AIE3AIE4AIE5AIE6AIE7AIE8AIE9AIE10AIE11AIE
AIE
1
DATA ROUTING FABRIC
text
VIRTUAL DEVICE MANAGER
FIBRECHANNEL
INTERFACE 1
FIBRECHANNEL
INTERFACE 2
Disk 2
Disk 1
Disk 2
Disk 1
Disk 2
Disk 1
12
4
3.0 Installation Instructions
If you are familiar with the Diamond Storage Array, Fibre Channel, SCSI and RAID configurations, you
may set up and configure the Diamond Storage Array using these instructions. Find details, illustrations
and other guidance for more involved operations and special cases in the rest of this manual.
CAUTIONCAUTION
Before configuring the Diamond Storage
Array, ensure that any computer data to be
stored on the array is properly backed up
and verified. The operation and storage of
computer data on any disk storage array
should be accompanied by a regular
program of computer data backups and
verification for the data stored on such an
array.
The Manufacturer is not responsible for the
loss of any data stored on the Diamond
Storage Array under any circumstances
and any special, incidental, or
consequential damages that may result
thereof.
Three major steps are required to set up the
Diamond Storage Array
1Physically set up Diamond Storage Array
2Connect to Ethernet
3Configure the drives
Step 1: physical setup
1Make sure the Diamond Storage Array is
mounted properly and has adequate air flow
around it.
2Insert the appropriate connector into the
interface card in the back of the Diamond
Storage Array.
3Connect the Fibre Channel or SCSI cable from
your host computer system to the connector. To
use ExpressNAV browser-based management
interface and configure your Diamond Storage
Array, you must connect to the Ethernet port.
4To set up the Ethernet connection: connect a
cross-over cable (for a direct connection to a
PC) or regular network cable from a network
device to the RJ45 Ethernet port on the
Ethernet management card on the front of the
Diamond Storage Array.
Step 2a: set up the Ethernet connection
The Diamond Storage Array supports service
operations over the RS-232 serial port using
standard terminal emulation software available
with most systems.
1Connect a DB-9 null modem serial cable
between the port marked RS-232 on the front of
the Diamond Storage Array and the computer’s
serial port. The cable must be no longer than
three meters.
2Make sure the power switches on the power
supplies on the rear of the unit are in the Standby position.
3Plug in the power cords to the back of the unit,
then into an appropriate power source (100-240
VAC).
4Reboot your host computer system.
5Press the Stand-by power switch for each
power supply on the Diamond Storage Array to
the ON position.
6Upon successful power up and diagnostics, the
unit displays the POST (power up self test)
information.
The Diamond is now in Command Line Interface
mode. You may modify the setup of the Diamond
Storage Array using the CLI (refer to
via ASCII-based Commands
on page 27), but the
CLI: Interface
easiest method to configure the array is by using
ATTO ExpressNAV software, a graphical user
management interface accessed through a
standard Internet browser. Refer to
ExpressNAV: Browser-based Interface
ATTO
on page 23.
Step 2b: connect to Ethernet
If the Diamond Storage Array is attached to a
DHCP server
1At the Ready prompt after POST (refer to Step
6 above), type set IPDHCP enabled
2Type SaveConfiguration Restart
3At the Ready prompt after POST (see above),
type get IPAddress
5
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
4Enter this address into your browser.
5The ATTO ExpressNAV screen appears. Log in
using the Telnet defaults:
Username: Telnet
Password: Diamond
6Follow the screens to find information about the
array or to configure the array from the factorydefault settings. The Diamond Storage Array
may be set up in a JBOD, RAID Level 0, 5 or 10
configuration with or without Hot Spare sleds
If the Diamond Storage Array is not attached to
a DHCP server and you wish to change the
defaults
1At the Ready prompt after POST (see above),
type set IPAddress [desired IP address].
2Type set IPSubnetMask [desired IP Subnet
Mask]
3Type set IPGateway [desired IP Gateway]
4Type SaveConfiguration Restart to save the
configuration and restart the Diamond Storage
Array
5After the powerup and POST complete, type
the IP address from step 1 above into your
browser.
6The ATTO ExpressNAV screen appears. After
logging in (refer to Step 2b: connect to Ethernet, Step 5 on page 6), follow the screens
to find information about the array or to
configure the array from the factory-default
settings. The Diamond Storage Array may be
set up in a JBOD, RAID Level 0, 1, 5 or 10
configuration with or without Hot Spare sleds.
Step 3: set up RAID configuration
The simplest way to set up configurations is to use
the ExpressNAV interface. Refer to
ExpressNAV: Browser-based Interface
ATTO
on page 23 for
more information on the interface. After logging
in (refer to
page 6
Step 2b: connect to Ethernet
, Step 5 on
), follow the screens to find information
about the array or to configure the array from the
factory-default settings
You may the CLI, the
.
QuickRAID
commands to
set up RAID and hot spare sleds, if required.
• Each Hot Spare sled configuration requires
a certain number of Hot Spare sleds. These
sleds, once designated as Hot Spares, are
not available for other use.
• RAID Level 10: 10 drive sleds and two Hot
Spare sleds (sleds 11 and 12).
Note
All arrays using RAID level 10 and Hot Spare
sled options must be fully populated.
• RAID Level 5: 10 drive sleds and two Hot
Spare sleds (sleds 11 and 12)
• RAID Level 5: five drive sleds (sled positions
1 through 5) and one Hot Spare sled (sled
12).
Installation
6
3.1 Components
The Diamond Storage Array has been designed to be easy to use, maintain and upgrade. It features a
durable steel outer case and modular components in either a floor or a rack mount model.
Immediately upon receipt, check the shipping
carton for damage from mishandling. Contact us
at once via the means that is easiest for you (refer
to
Contact ATTO Technology, Inc.
on page viii) if the
carton has been mishandled or displays any signs
of damage.
The front of the Diamond Storage Array provides
access to the management card and disk drive
sleds. The rear of the unit holds the Host Interface
Cards, power supplies and blower assemblies.
CAUTIONCAUTION
All modular components must be replaced
by qualified personnel only.
Floor model
The management system card is at the top front of
the case. At its center is a DB-9 serial RS-232
port, a connection for setup, monitoring and
upgrade of the unit from any computer system
with an RS-232 interface. The optional 10/100
BaseT Ethernet management services card
enables Telnet-based monitoring and
management. It also provides the ability to update
the firmware in the array via FTP.
LEDs to the port’s
right indicate fault,
unit ready, Host
Interface Cards A
and B installation
status, and the power
status for each power
supply.
sled contains two hard drives. Up to 24 hard
drives may be installed on the 12 sleds. Empty
bays should be covered by blank faceplates or
empty sleds. Access is provided by loosening two
screws and gently pulling on the sled handle.
On the rear of the unit are blowers which support
hard drive, cabinet and power supply cooling. The
blowers are held in by removable screws. Correct
operation is displayed by a LED at the top of each
panel.
The power supplies for the array, also in the rear
of the unit, are accessible by loosening two screws
and pulling on the power supply module handle.
The power standby on/off switch is at the top of
each module. A yellow LED indicates
and a green LED indicates
on
. The power cord
caution
socket is at the bottom of each power supply.
Between the power supplies and blower
assemblies are two slots that hold the Host
Interface Cards. The HIC is the connection point
into the array and is available in three options: 1Gigabit Fibre Channel, 2-Gigabit Fibre Channel,
or Ultra 160 SCSI. Host Interface Cards have
and
on-line
or
fault
and
activity
LED indicators,
fault
depending on the model.
SCSI
The SCSI card faceplate has a rotary
binary-coded hex switch to set the SCSI ID of the
array. The SCSI card also has an in channel, to
connect via cable to the unit’s communication
source, and an out channel, available for daisychaining arrays together or to complete
termination using an external LVD terminator.
Below the
management system
card are individual
disk drive sleds
which also have
LEDs for each
drive’s status. Each
7
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
Rack mount
The system management card is at the left front of
the case. At its center is a DB-9 serial RS-232 port
which allows a connection for setup, monitoring
and upgrade of the unit from any computer system
with an RS-232 interface. The optional 10/100
monitoring and
management. It also
provides the ability to
update the firmware in the
array via FTP. LEDs
farthest to the left indicate
fault, unit ready, Host
Interface Cards A and B
installation status, and the
Disk drive sleds (12)
power status for each
power supply.
Power supplies
To the right of the system
management card are
individual disk drive sleds
which also have LED
indicators for each drive’s
status. Each sled contains
two hard drives. Up to 24
hard drives may be
installed on the 12 sleds.
Empty bays should be
covered by blank faceplates or filled with empty
sleds. Access is provided by loosening two screws
and pulling on the sled handle.
In the rear of the unit are the blower assemblies
which support hard drive, cabinet and power
supply cooling. Correct operation is displayed by
a LED at the top of each panel. The blowers are
held in place by removable screws.
The power supplies for the array are accessible by
loosening two screws and pulling on the power
supply module handle. The power standby on/off
switch is at the top of each module. A yellow LED
indicates
caution
and a green LED indicates on.
The power cord socket is at the bottom of each
power supply.
Intelligent midplane (internal)
Host interface cards
Blower assemblies
Between the power supplies and blower
assemblies are two slots that hold the Host
Interface Cards. The HIC is the connection point
into the array and is available in a 2-Gigabit Fibre
Channel or Ultra 160 SCSI. Host Interface Cards
fault
and
have
on-line
or
fault
and
activity
LED
indicators, depending on the model.
SCSI
The SCSI card faceplate has a rotary
binary-coded hex switch which allows you to set
the SCSI ID of the array. The SCSI card also has
an in channel, to connect by a cable to the unit’s
communication source, and an out channel,
available for daisy-chaining arrays together or to
complete termination using an external LVD
terminator.
Components
8
3.2 Physical Set Up
The Diamond Storage Array is shipped completely assembled with two 120 VAC power cords for use in
the United States and Canada.
Immediately upon receipt, check the shipping
carton for damage from mishandling. Contact us
at once by the means easiest for you (refer to
Contact ATTO Technology, Inc.
on page viii) if the
carton has been mishandled or displays any signs
of damage.
Floor model
The Diamond Storage Array is heavy (about 92
pounds fully loaded) and requires two people to
lift and carry it safely. Place the array on a level
surface and make sure there is adequate space in
the front and back of the unit for proper cooling
and airflow. Continue with the general
instructions.
Rack mount
The array fully loaded is heavy (about 86 pounds).
The unit should be handled with care and requires
two people to lift, carry and/or install it safely.
The Diamond Array can be mounted via several
different methods in a 19” rack with 3U (5.25”) of
vertical space required. Air flow should not be
restricted in any way.
Installation in a rack may create a differential
between the room ambient temperature and the
internal ambient temperature in the rack. While
the maximum internal operating temperature of
the array is 47°C, you should not run the system
at the maximum temperature for extended
periods. Therefore, ensure that the room ambient
temperature is kept below 30°C for best
operation.
Spaced rail pairs in some rackmount cabinets.
You can mount the array using two sets of rail
pairs spaced to accommodate the overall length of
the unit (approximately 23 inches). Mount using
the rack mount brackets on the front and rear of
the unit fastened to the rail pairs using 10/32 pan
head screws with lock and flat washers.
Rack mount cabinets with stationary shelf or tray
system.
The shelf or tray must be able to support
at least 125 pounds. The shelf or tray must be
installed and secured to the rack before installing
the array. Secure the front of the array to the rack
with 10/32 screws, locks and flat washers.
Sliding shelf or tray type systems
should never
be used under any circumstances.
Two point open rack system.
The rack must be
strong enough to support the Diamond Array
properly. Mounting brackets should be moved to
the centermost mounting holes and secured using
10/32 screws.
CAUTIONCAUTION
Do not mount multiple arrays on a two-rail
rack or mount the array above the midpoint
of a two-rail rack system. Do not mount the
array on any kind of rail-type system. The
array is too heavy and does not have the
proper hole pattern for rails.
Insure the array has adequate air flow and
continue with the general instructions.
Each side of the rack mount array chassis has
three pairs of mounting holes. One is located near
the front of the rack, one near the unit’s center of
gravity, and one near the rear of the rack. The
holes will accommodate 10/32 screws but the
screws can protrude no farther than .375 inches
into the rack.
9
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
General instructions
1Insert the proper connector into the Host
Interface Card in the back of the array. (Refer to
Connecting a Fibre Channel Array
Fibre Channel and
page 13 for SCSI).
2Connect the cable (Fibre Channel or SCSI)
from your host system to the Host Interface
Card connector on the back of the array. The
cable you use depends upon your application,
the environment and distance.
3Make sure the power switches on the power
supplies on the rear of the unit are in the standby position. Plug in the power cords to the back
of the unit, then into an appropriate power
source (100-240 VAC). The power source must
be connected to a protective earth ground and
comply with local electrical codes. Improper
grounding may result in an electrical shock or
damage to the unit.
Exhibit 3.2-1 Back side of a rack mount Fibre Channel Diamond Storage Array.
Connecting a SCSI Array
on page 11 for
on
4Press the stand-by power switch for each
power supply to the ON position. When the
green power LED on the back of the unit is lit,
the power supply is fully operational and
delivering power to the system. The power LED
on the front of the array will light once the
firmware begins to execute.
When the power is turned on, the LEDs on the
front of the array will flash twice. Drives will spin
up in groups of three, about every one to two
seconds. The individual LEDs will blink. After all
available drives have spun up, the individual
drive LEDs will stay lit. When all available
drives are operational, the ready LED on the
top front panel of the management card will
remain lit.
5Reboot your computer
6Determine the best configuration for your needs
(i.e. JBOD, RAID, etc.) and refer to the rest of
this manual for more detailed explanations and
configuration information.
10
Installation
3.2.1 Connecting a Fibre Channel Array
The Diamond Storage Array supports up to two Fibre Channel Host Interface Cards (HIC). Physical
connections and CLI commands contribute to the Fibre Channel topology.
The cable you use depends
upon your application, the
environment and the
distance required for your
2 Gb HIC uses
2 SFPs to connect up
to 2 FC cables
storage area network.
To comply with FCC and
CE EMI for the 2-gigabit
Host Interface Card, use fiber optic cables.
Exhibit 3.2-2 Fibre Channel cable options
Cable lengthCable typeCable size
<10 metersUnequalized copper
>10 <30 metersEqualized copper
Up to 175 metersmulti mode fiber optic62.5 micron
Up to 500 metersmulti mode fiber optic50 micron
Up to 10 kilometerssingle mode fiber optic9 micron
The Diamond Storage Array may have two Fibre
Channel Host Interface Cards (HIC). In 2 gigabit
Fibre Channel arrays, each HIC is connected by a
Fibre Channel cable via a SFP (small form factor
pluggable) module into a point-to-point or loop
Fibre Channel topology.
Install the SFP according to the manufacturer’s
instructions.
Each HIC has two ports and an on-board hub.
Each port has an SFP module to connect to Fibre
Channel. Each HIC is independent of the other, so
that one may be connected into a point-to-point
topology and the other into a loop. However, if
one port in a HIC is connected into a point-topoint topology, the other port cannot be used.
One of the advantages of using loop topology for
Fibre Channel connections is that it allows arrays
to be daisy-chained together.
Autoconfiguration
The array automatically determines which HICs
are installed and if they are in loop or point-topoint topologies.
If you wish to see how the unit has been set up, go
to the CLI commands and type
Info or go to the
Status page of the ExpressNAV interface. The
return displays the data rate and connection mode
for each HIC (FC 0 and FC 1).
Manual configuration
You may manually configure the array using CLI
commands (refer to
Guidance
Channel
(refer to
on page 29.) or access the
page of the ExpressNAV interface
ATTO ExpressNAV: Browser-based Interface
Command Line Interface Use and
Fibre
on page 25).
•
FCConnMode
topology for both HICs on an array. Options are
loop only (loop), point-to-point only (ptp), loop
preferred (loop-ptp) or point-to-point preferred
(ptp-loop).
Loop connects to either an FC-AL arbitrated loop or
a fabric loop port (FL_Port) on a switch.
Point-to-point (ptp) connects to a direct fabric
connection, such as an F port on a switch.
Loop-ptp allows the array to determine what kind of
connection to use, but tries to connect in loop mode
first, then point-to-point mode.
Ptp-loop allows the card to determine what kind of
connection to use, but tries to connect in point-topoint mode first, then loop mode.
• FcDataRate specifies the Fibre Channel data
rate at which both HICs on a Diamond operate.
Choices are 1 gigabit, 2 gigabit and
autodetection.
specifies the Fibre Channel
11
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
Exhibit 3.2-3 Possible 2 gigabit Fibre Channel physical connections depending on which Fibre Channel connection
mode has been selected.
point-to-
point mode
loop
mode
Diamond Storage Array ADiamond Storage Array B
no
connection
possible
loop
mode
daisy -chain
loop
mode
loop
mode
loop
mode
loop
mode
12
Fibre Channel connections
3.2.2 Connecting a SCSI Array
The SCSI Diamond Storage Array uses a VHDCI connector and SCSI cables to connect to a host. It will
automatically detect the type of Host Interface Card it is using without any intervention.
To connect a SCSI array
1Insert a SCSI VHDCI connector into the Host
Interface Card in the back of the array.
2Connect the cable from your host system to one
of the VHDCI connectors on the Host Interface
Card connector on the back of the array.
3If the SCSI array is the last device on the bus,
you must attach a VHDCI terminator to one
connector of the SCSI Host Interface Card or
Exhibit 3.2-3 SCSI cable options.
Bus speed, MB/sec.
max.
SCSI-1586-8
Fast SCSI1086-8
Fast Wide SCSI20166-16
Wide Ultra SCSI40163-4
Wide Ultra SCSI40161.5-8
Wide Ultra 2 SCSI8016-1216
Ultra 3 or Ultra160 SCSI16016-1216
Bus width,
bits
connect a cable between the second connector
and the next device on the SCSI bus.
4The SCSI Host Interface Card has a rotary
binary-coded hex switch which allows you to
set the SCSI ID of the HIC. Be sure the selected
ID is different from all other SCSI devices on
the bus.
Note
If slower devices are connected on the same
SCSI bus as the Ultra 160 array, the bus will
communicate at the rate of the slowest device.
Max. bus lengths,
meters
Single-endedLVD
Max. device
support
Exhibit 3.2-4 SCSI interface cards: left without terminators attached; right with a terminator attached.
13
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
14
SCSI connections
4.0 Determining Drive and Sled Designations
g
The Diamond Storage Array has been designed with 12 sleds, each holding two drives. The easiest way
to configure a array is to use all the drives on all the sleds. The firmware uses a numbering system to
determine which drives and sleds it is affecting.
We recommend that all sled slots are filled
contiguously, starting with the first slot next to the
management card.
When configuring a Diamond Array with fewer
than 12 drive sleds, you must consider several
factors: RAID level, number of physical
drives/sleds and the end configuration you are
trying to achieve. Review the information about
each configuration to determine how each
configuration would be affected by using fewer
sleds.
Numbering conventions
The Diamond Storage Array with firmware
version 3.1 and newer uses a unique numbering
convention to orient its drives and sleds to the
controlling firmware. Older versions do not use
this convention. (Refer to
Updating Firmware
on
page 77 for information about updating the array
firmware.)
Physical
refers to the physical drives in the array,
the hardware that actually exists in a physical
sense.
Logical (or virtual)
refers to what the host
operating system recognizes as an entity. Two
physical drives may be seen as one logical drive
by the operating system.
Logical disks do not always map one-to-one with
physical disks.
In RAID configurations, for
example, several physical disk drives (or portions
of several physical drives) are grouped into a
1 2 3 4 5 6 7 8 9 10 11 12
Disk 2
Disk 2
Disk 2
LUN 13
LUN 14
LUN 15
Disk 2
LUN 16
Rack Mount Drive Sleds
LUN 17
Disk 2
LUN 16
LUN 17
Disk 2
LUN 18
Disk 2
LUN 19
LUN 20
LUN 19
logical disk or Logical Unit (LUN). Each LUN is
broken into logical blocks of 512 bytes each,
numbered 0 through n (the Logical Block Number
or LBN). A 100 GB LUN has approximately
200,000,000 logical blocks.
Physical numbering
The Diamond RAID Storage Array contains
• Up to 24 physical disk drives
• Two drives mounted on 12 physical drive sleds
• Sleds are numbered 1-12, starting at the top
(floor units) or the left (rackmount units).
• Each sled is connected to its own internal ATA
bus with two disk drives, numbered 1 and 2.
• Two green LEDs, labeled Drive 1 and Drive 2,
indicate activity for the two drives. They remain
solidly lighted when there is no activity.
Logical numbering
Logical numbering depends on
the RAID configuration of the
storage array. S Class firmware
includes RAID Level 0, RAID
Level 1, RAID Level 10 and
RAID Level 5 capability. A
variety of configurations may
be selected using the
QuickRAID0 command. The
default configuration is
QuickRAID0 0 or JBOD (Just
a Bunch of Disks).
Disk 2
Disk 2
Disk 2
Disk 2
LUN 20
LUN 21
LUN 22
LUN 23
Disk 2
LUN 24
Mana
Drive 1
LUN 1
Drive 1
LUN 2
Drive 1
LUN 3
Drive 1
LUN 4
Drive 1
LUN 5
Drive 1
LUN 6
Drive 1
LUN 7
Drive 1
LUN 8
Drive 1
LUN 9
Drive 1
LUN 10
Drive 1LUN 6
ement car d
Drive 2
LUN 16
Drive 2
LUN 17
LUN 18
Drive 2
LUN 19
Drive 2
LUN 20
Drive 2
LUN 21
LUN 22
Drive 2
LUN 13
Drive 2
LUN 14
Drive 2
LUN 15
Drive 2
LUN 20
LUN 19Drive 1LUN 12
Drive 2
LUN 17
1 2 3 4 5 6 7 8 9 10 11 12
Floor Model Drive Sleds
Drive 2
Drive 1
LUN 11
Drive 1
LUN 12
LUN 23
Drive 2
LUN 24
Management Card
15
Disk 1
LUN 1
Disk 1
LUN 2
Disk 1
LUN 3
Disk 1
LUN 4
Disk 1
LUN 5
Disk 1
LUN 6
Disk 1
LUN 7
Disk 1
LUN 8
Disk 1
LUN 9
Disk 1
LUN 10
Disk 1
LUN 11
Disk 1
LUN 12
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
Examples
JBOD
mode uses 24 LUNs. Each LUN is
equivalent to one physical drive. The array will
operate with several sleds missing, but the empty
sled(s) will be treated as offline and will not be
configurable.
RAID Level 0 (
QuickRAID0 [n]
)
With a fully
populated array, RAID 0 may be configured as 1,
2, 3, 4, 6, or 12 LUNs. As
QuickRAID0 1
, all 24
physical drives are configured as a single stripe
group or LUN. You may also configure two LUNs
of 12 drives each, three LUNs of eight drives
each, four LUNs of six drives each, six LUNs of
four drives each and 12 LUNS of two drives each.
(See Exhibit 6.2-3 in
RAID Level 0
on page 53). The
command assumes there are 24 drives available to
configure the specified number of LUNS.
configured as six mirrored pairs, then striped into
two LUNs.
RAID Level 5 (
QuickRAID5
)
with a fully
populated Array may be configured as 1, 2, 3 or 4
LUNs. As
QuickRAID5 1
, all 24 physical drives are
configured as a single RAID 5 LUN. You may
also configure two LUNs of 12 drives each, three
LUNs of eight drives each, or four LUNs of six
drives each. RAID Level 5 parity will reduce the
usable capacity of each LUN by the equivalent of
one drive sled.
Unique serial number for each LUN
Each LUN in a system has a unique 24-character
serial number which is updated when a system
configuration changes. It is stored in a Device
Association Table on each drive.
RAID Level 1 (
array will be configured into six groups, LUNs 16, with each LUN containing two physical sleds;
each sled in the LUN will be a mirror image of the
other sled in the LUN.
RAID Level 10 (
command first creates six mirrored groups, then
stripes them into groups of one, two or three
RAID 10 groups. Each group is a LUN. Issuing
QuickRAID10 2
Character
Placement
0-19Any ASCII
201 for Raid1
21A – ZLUN Configuration Iteration Character: starts at A and increments every time a member drive is
22A – ZGeneration Character: starts at A and increments every time a new configuration is stamped on the
230Reserved for future use
QuickRAID1
QuickRAID10
)
has no options: the
)
The
QuickRAID10
, the 12 physical sleds are
Valid values
char
A for Raid10
0 for Raid0
5 for Raid5
X for JBOD
Example of unique serial number for a JBOD configuration:
Drive Serial Number
RAID Configuration Character
replaced in a RAID Group. Reverts back to A any time the Generation Character is updated
system. When it reaches Z, rolls over to
When a drive or sled is replaced in
QuickRAID
configurations that allow for hot swap, a new
serial number will be computed and will in turn be
stamped onto all member drives of the RAID
group. The CLI command
ExpressNAV interface page
LUNInfo
or the
Logical Units will
display the serial number for each drive.
Format of the 24 characters is defined in the chart
below:
1231231231231231231XAC0
A
.
16
Numbering conventions
5.0 Accessing the Array
Communicate with the Diamond Storage Array through the Fibre Channel link, the RS-232 port or the
Ethernet port using Command Line Interface commands.
You may configure and tune the Diamond Storage
Array for different environments and
applications, update the firmware, monitor
internal power and temperature status, report on
hardware diagnostics and log failures.
Three avenues are available
• In-band SCSI over Fibre Channel and over SCSI
• RS-232 port
• Telnet and SNMP over Ethernet
The rest of this manual describes how to access
the array and use the Command Line Interface or
the ExpressNAV interface.
Command Line Interface
The Command Line Interface provides access to
the array through ASCII command lines.
An initial display, after powering up the unit or
restarting the firmware, will contain the
information in Exhibit 5.0-1. Once the initial
display is complete, with the word
Ready
,
you are
in the Command Line Interface mode.
• In-band CLI over SCSI, where ASCII CLI
commands, may be issued via
CLI commands except
• ID/value, where the application program uses a
SCSI CDB (command descriptor block) to select
the buffer ID of the configuration parameters to
be affected, and the new value of the parameter.
Most configuration options are available.
Zmodem
Write Buffer.
are supported.
All
RS-232 port
The array provides remote service operations over
the RS-232 serial port using standard terminal
emulation software available with most systems.
You will need to set the following serial
parameters in your terminal program
• Bits per second: 115200
• Data Bits: 8
• Parity: None
• Stop Bits: 1
• Flow Control: None.
• Terminal type: ASCII
• Echo: on
Help
Type
available.
ATTO ExpressNAV
ATTO ExpressNAV is an integrated configuration
tool accessible through an Ethernet connection.
Platform independent, ExpressNAV contains all
the current capabilities of the CLI in a userfriendly GUI console. A menu on each page
provides access to information and configuration
operations.
Refer to
on page 25 for more information.
to display a list of all commands
ATTO ExpressNAV: Browser-based Interface
Ethernet port
The 10/100 BaseT Ethernet port provides Telnetor SNMP-based monitoring and management.
The default IP address is 10.0.0.1; the default
subnet mask is 255.255.255.0. These parameters
can be changed using the RS-232 port or Telnet.
The array should first be configured for the
network using the RS-232 port to establish the
correct IP address.
SNMP, or Simple Network Management
Protocol, is an application layer protocol that
allows network devices to exchange management
information. Through a combination of standard
In-band SCSI over Fibre Channel
In-band SCSI commands (
Buffer
) may be issued to the array Fibre Channel
Write Buffer
and
Read
port to manage configuration via two mechanisms
17
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
and custom MIBs (Management Information
Base), the array provides status and error
indications to an SNMP server, allowing the array
to be managed with other devices in a complex
system through a common interface. You use CLI
commands or the
SNMP
page of the
ExpressNAV interface to configure up to six
unique Trap addresses. A trap is a way for the
array to tell the SNMP server that something has
happened.
SNMP
SNMP, or Simple Network Management
Protocol, is an application layer protocol that
allows network devices to exchange management
Exhibit 5.0-1 An example of the POST information on screen after boot
Firmware version 5.40 release date Mar 30 2004, 10:43:06 Build 021G
Power-On Self-Test Completion Status: GOOD
64 Megabytes of RAM Installed.
Interface Port 0 is not installed.
Interface Port 1 is 1.0624 Gb/s Fibre Channel.
information. Through a combination of standard
and custom MIBs, the array provides status and
error indications to an SNMP server, allowing the
array to be managed with other devices in a
complex system through a common interface.
You use CLI commands or the
ExpressNAV interface to configure up to six
unique Trap addresses. A trap is a way for the
array to tell the SNMP server that something has
happened.
SNMP
page of the
Interface 0 World Wide Name = 20 00 00 10 86 10 02 DC
Interface 1 World Wide Name = 20 00 00 10 86 10 02 DC
Diamond Array Serial Number = "MIDP100197"
Diamond Array Name = " "
System Serial Number = ""
Active Configuration = ATTO
DiamondClass = (S)83
Internal Temperature = 26 C [5 - 47]
ErrorLog Contents: NO ERRORS
For help, type HELP.
Ready.
18
Interfaces
5.1 In-band CLI Over the Fibre Channel
In-band Command Line Interface (CLI) configures and manages the Diamond Storage Array using
SCSI-based CLI commands over a Fibre Channel port connection.
In-band CLI allows a programmer to configure
the Diamond Storage Array while it is moving
data. Using a programmer’s interface, CLI
commands as described previously in this manual
may be implemented.
In-band CLI is implemented as part of LUN 0. It
uses a different LUN than the array, and reports as
a Storage Enclosure Services (SES) device
(device type 0x0D).
LUN 0 is visible on all Fibre ports but is actually
a single unit. The default value for LUN 0 is 0x00.
LUN 0 must be reserved for each Write
Buffer/Read Buffer pair, using the SCSI Reserve
command to insure integrity of the in-band CLI
session.
1An initiator (host) sends a SCSI Reserve
command to LUN 0.
• If LUN 0 is not reserved by another initiator,
LUN 0 is now reserved and ready to begin a
new CLI session.
• If the array configuration is reserved by a
different CLI session (i.e. serial or Telnet),
the in-band session will not be allowed to
modify the array configuration. If you try, the
results buffer of LUN 0 will return
Process X has the configuration
reserved.
ID of this session = Y
Ready.
2The initiator issues a SCSI Write Buffer
command to LUN 0. A
must be accompanied by an ASCII buffer
representing the CLI command string such as
set DiamondName Omega1
3LUN 0 will execute the command line and
create feedback in the form of ASCII characters
into a buffer. This buffer is 8 KB and circular.
Retrieve the results by issuing a Read Buffer
command before issuing another Write Buffer
command.
Write Buffer
command
4A subsequent Write Buffer command will
execute the new command line and overwrite
the previous results in the buffer with new
results.
5LUN 0 can be released by issuing a SCSI
Release command to the LUN after each
Write/Read Buffer
Buffer pairs.
Initiator (Host)Diamond Array
Reserve LUN 0return: “ok”
Write Buffer
LUN 0
bid ‘AA’
“get Temperature”
Read Buffer
LUN 0
bid ‘AA’
Release LUN 0return: “ok”
pair, or multiple
executes the CLI command,
stores output in buffer
return:
”Temperature=28C\r\n\Ready
.\r\n\0”
Write/Read
I/O details
The buffer sent to the Services LUN during the
data out phase of a Write Buffer command must
be
• ASCII data
• maximum 80 bytes length
• terminated with either a carriage return
character (0x0D), line feed character (0x0A) or
NULL character (0x00)
• Characters following the first carriage return
character, line feed character or NULL character
are ignored.
The buffer retrieved from the Services LUN
during the data-in phase of a Read Buffer
command will be
• ASCII data
• 8 KBytes (8192 bytes) in length
• terminated with a NULL character (0x00)
• Characters following the NULL character are
meaningless.
19
ATTO Technology Inc. Diamond Storage Array Installation and Operation Manual
Loading...
+ 75 hidden pages
You need points to download manuals.
1 point = 1 manual.
You can buy points or you can get point for every manual you upload.