Professional Documents
Culture Documents
Student Guide
Third-party software, including font technology, is copyrighted and licensed from Sun suppliers.
Sun, Sun Microsystems, the Sun logo, Sun StorEdge, Sun StorageTek, Solaris, and Java are trademarks or registered trademarks of Sun
Microsystems, Inc. in the U.S. and other countries.
All SPARC trademarks are used under license and are trademarks or registered trademarks of SPARC International, Inc. in the United
States and other countries. Products bearing SPARC trademarks are based upon an architecture developed by Sun Microsystems, Inc.
UNIX is a registered trademark in the U.S. and other countries, exclusively licensed through X/Open Company, Ltd.
Export Laws. Products, Services, and technical data delivered by Sun may be subject to U.S. export controls or the trade laws of other
countries. You will comply with all such laws and obtain all licenses to export, re-export, or import as may be required after delivery to
You. You will not export or re-export to entities on the most current U.S. export exclusions lists or to any country subject to U.S. embargo
or terrorist controls as specified in the U.S. export laws. You will not use or provide Products, Services, or technical data for nuclear, missile,
or chemical biological weaponry end uses.
DOCUMENTATION IS PROVIDED “AS IS” AND ALL EXPRESS OR IMPLIED CONDITIONS, REPRESENTATIONS, AND
WARRANTIES, INCLUDING ANY IMPLIED WARRANTY OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE
OR NON-INFRINGEMENT, ARE DISCLAIMED, EXCEPT TO THE EXTENT THAT SUCH DISCLAIMERS ARE HELD TO BE
LEGALLY INVALID.
THIS MANUAL IS DESIGNED TO SUPPORT AN INSTRUCTOR-LED TRAINING (ILT) COURSE AND IS INTENDED TO BE
USED FOR REFERENCE PURPOSES IN CONJUNCTION WITH THE ILT COURSE. THE MANUAL IS NOT A STANDALONE
TRAINING TOOL. USE OF THE MANUAL FOR SELF-STUDY WITHOUT CLASS ATTENDANCE IS NOT RECOMMENDED.
Ce produit ou document est protégé par un copyright et distribué avec des licences qui en restreignent l’utilisation, la copie, la distribution,
et la décompilation. Aucune partie de ce produit ou document ne peut être reproduite sous aucune forme, par quelque moyen que ce soit,
sans l’autorisation préalable et écrite de Sun et de ses bailleurs de licence, s’il y en a.
Le logiciel détenu par des tiers, et qui comprend la technologie relative aux polices de caractères, est protégé par un copyright et licencié
par des fournisseurs de Sun.
Sun, Sun Microsystems, le logo Sun, Sun StorEdge, Sun StorageTek, Solaris, et Java sont des marques de fabrique ou des marques déposées
de Sun Microsystems, Inc. aux Etats-Unis et dans d’autres pays.
Toutes les marques SPARC, utilisées sous licence, sont des marques déposées ou enregistrées de SPARC International, Inc. aux Etats-Unis
et dans d’autres pays. Les produits portant les marques SPARC sont basés sur une architecture développée par Sun Microsystems, Inc.
UNIX est une marque enregistree aux Etats-Unis et dans d’autres pays et licenciée exclusivement par X/Open Company Ltd.
Législation en matière dexportations. Les Produits, Services et données techniques livrés par Sun peuvent être soumis aux contrôles
américains sur les exportations, ou à la législation commerciale dautres pays. Nous nous conformerons à lensemble de ces textes et nous
obtiendrons toutes licences dexportation, de ré-exportation ou dimportation susceptibles dêtre requises après livraison à Vous. Vous
nexporterez, ni ne ré-exporterez en aucun cas à des entités figurant sur les listes américaines dinterdiction dexportation les plus courantes,
ni vers un quelconque pays soumis à embargo par les Etats-Unis, ou à des contrôles anti-terroristes, comme prévu par la législation
américaine en matière dexportations. Vous nutiliserez, ni ne fournirez les Produits, Services ou données techniques pour aucune utilisation
finale liée aux armes nucléaires, chimiques ou biologiques ou aux missiles.
LA DOCUMENTATION EST FOURNIE “EN L’ETAT” ET TOUTES AUTRES CONDITIONS, DECLARATIONS ET GARANTIES
EXPRESSES OU TACITES SONT FORMELLEMENT EXCLUES, DANS LA MESURE AUTORISEE PAR LA LOI APPLICABLE, Y
COMPRIS NOTAMMENT TOUTE GARANTIE IMPLICITE RELATIVE A LA QUALITE MARCHANDE, A L’APTITUDE A UNE
UTILISATION PARTICULIERE OU A L’ABSENCE DE CONTREFAÇON.
CE MANUEL DE RÉFÉRENCE DOIT ÊTRE UTILISÉ DANS LE CADRE D’UN COURS DE FORMATION DIRIGÉ PAR UN
INSTRUCTEUR (ILT). IL NE S’AGIT PAS D’UN OUTIL DE FORMATION INDÉPENDANT. NOUS VOUS DÉCONSEILLONS DE
L’UTILISER DANS LE CADRE D’UNE AUTO-FORMATION.
Course Goals
Upon completion of this course, you should be able to:
● Describe the features, functions and terminology of the Sun
StorageTek 6540 array
● Describe the customer benefits and requirements to migrate to or use
the Sun StorageTek 6540 array
● Describe the architecture of the Sun StorageTek 6540 array
● Install the Sun StorageTek 6540 array hardware
● Install the management software (Common Array Manager)
● Configure the 6540 array using CAM
● Attach production hosts to the Sun StorageTek 6540 array
● Configure and use Snapshots on the Sun StorageTek 6540 array
● Configure and use Volume Copies on the Sun StorageTek 6540 array
● Configure and use Replication Set on the Sun StorageTek 6540 array
● Diagnose problems using available tools
Objectives
Upon completion of this module, you should be able to:
● Describe the Sun StorageTek 6540 key features
● Identify the hardware components of the 6540 controller enclosure
● Describe the functionality of the 6540
● Interpret LEDs for proper parts replacement
The 6998 controller used in the 6540 storage system is the most
sophisticated and highest-performing controller to date from SUN
StorageTek for the 6000 mid-range disk product line. Its sixth-generation
XBB architecture boasts our fastest cache memory, 4 Gbps Fibre Channel
host and drive interfaces, high-speed busses, and multiple processing
elements to optimize resource utilization.
The 6998 controller's high-speed XOR engine generates RAID parity with
no performance penalty, enabling this compute-intensive task to be
handled efficiently and effortlessly. A separate 2.4 Ghz Xeon processor
focuses on data movement control, allowing setup and control
instructions to be processed and dispatched independent of data.
The 6540 storage system has eight 4 Gbps FC-AL host or FC-SW SAN
connections and eight 4 Gbps FC-AL drive expansion connections.
The SUN StorageTek 6140 and 6540 storage systems run similar firmware.
This unique implementation creates a lower total cost of ownership and
higher return on investment by enabling seamless data and model
migration, common features and functionality, centralized management, a
consistent interface and reduced training and support costs. Additionally,
the 6140 storage system can be upgraded to a high-performance 6540 HPC
storage system. And in each instance, all configuration and user data
remains intact on the drives.
The Sun StorageTek 6540 storage system is modular and rack mountable,
and scalable from a single controller tray (CRM=Controller RAID
Module) plus one expansion tray (CEM=Controller Expansion Module) to
a maximum of 13 additional expansion trays.
● 4GB, 8GB and 16GB cache options are available (2GB / 4GB / 8GB
per controller respectively).
● 4 drive loops per controller that can support either 2Gb/s or 4Gb/s
drive enclosures.
● All components are hot-swappable
● RoHS compliant
Sun Sun
Sun
StorageTek StorageTek
StorageTek
6140 6540
6140 Lite
Controller
CPU Processor 667 Mhz 667 Mhz 2.4 Ghz
Xscale w/ Xscale w/ Xeon
XOR XOR dedicated
XOR
Host Ports 1/2/4 Gb/s 1/2/4 Gb/s 1/2/4 Gb/s
2 per ctlr 4 per ctlr 4 per ctlr
Expansion Ports 2 per ctlr 2 per ctlr 4 per ctlr
Controller Cache 1 GB per ctlr 2 GB per ctlr 2/4/8 GB per
ctlr
Ethernet Ports 2 per ctlr 2 per ctlr 2 per ctlr
Controller 3992 3994 6998
Expansion Tray
IOM FC FC FC
# of Disk Drives per 16 16 16
Tray
Sun Sun
Sun
StorageTek StorageTek
StorageTek
6140 6540
6140 Lite
Disk Types 2/4 Gb/s: FC, 2/4 Gb/s: FC, 2/4 Gb/s: FC,
SATA II SATA II SATA
# Expansion Trays 3 6 14
Maximum Disks 64 112 224
Disk Types 2/4 Gb/s: FC, 2/4 Gb/s: FC, 2/4 Gb/s: FC,
SATA II SATA II SATAII
Configuration
Maximum Hosts 512 512 512
(256 (256
redundant) redundant)
Maximum Volumes 1024 1024 1024
Performance Targets
Burst I/O rate -
Cache Read 120,100 120,100 575,000
Sustained I/O rate -
Disk Read 30,235 44,000 85,000
Sustained I/O rate -
Disk Write 5,789 9,000 22,000
Sustained 750 MBps 990 MBps 1,600 MBps
throughput - Disk
Read
Sustained 698 MBps 850 MBps 1,300 MBps
throughput -
Disk Writes
Hardware Overview
This section describes the main components of the Sun StorageTek 6540
controller tray (CRM). The CSMII is covered in another module.
Controller Tray
Figure 1-2 shows a block diagram for the Sun StorageTek 6540. The blocks
represent placement of controllers, power-fan canisters and removable
mid-plane canister.
The Sun StorageTek 6540 controller enclosure has five main canisters:
● two Power-Fan canisters
● one Interconnect canister (removable mid-plane)
● two controller canisters
There are also two battery FRU’s (Field Replaceable Units) within the
Interconnect-Battery canister, bringing the total number of FRU’s for the
6540 controller enclosure to seven.
The enclosure does not have a mid-plane but instead has been designed
such that all the canisters interconnect with one another.
- Fans -packs
2 battery packs - Fans
Fans 2 battery Fans- 2 Battery charger
host interface card - 2 battery chargers
2 Battery Chargers
- Thermal sensor
- Audible alarm
Audible alarm
- LEDs 2 Battery Chargers
- Thermal sensor
Thermal Sensor LED’s Thermal Sensor
Front
Power-Fan Canister
Figure 1-5 6540 Power Fan Canister and Battery Canister LEDs
● fan failure
● defective circuitry in the power supply
● blocked air vent
● failure in other devices installed in the cabinet
If the internal temperature rises above 70°C (158°F) one or both power
supplies will automatically shut down, and the storage management
software will report the exception. Critical event notifications will also be
issued if event monitoring is enabled and event notification is configured.
In the figure above, note the black connector when looking at the back of
the canister - this connector connects to one of the controllers. The Power-
Fan canister on the right has the connector at the top and therefore
connects to controller A. The Power-Fan canister on the left is upside
down and has the connector on the bottom and therefore connects to
controller B.
Information about the condition of the power supplies, fans and battery
charger is conveyed by indicator lights on the front of each Power-Fan
canister. You must remove the front cover of the 6540 enclosure to see the
indicator lights.
Interconnect-Battery Canister
● front bezel LED’s - the LED’s that are displayed through the front
cover are located on the Interconnect-Battery canister.
The Power, Service Action Required, and Locate lights are general
indicators for the entire command enclosure, not specifically for the
Interconnect-Battery canister. The Service Action Required light turns on
if a fault condition is detected in any component in the controller
enclosure. The Power, Service Action Required, and Locate lights shine
through the front cover.
The above figure shows the Interconnect-Battery canister with the access
cover removed. For clarity, the picture shows only one battery pack, there
would normally be two. The battery pack is mounted to a sheet metal
bracket. You can see the flange at the end of the bracket closest to the
access - grasp the flange to remove the battery pack. When replacing the
battery pack, the battery pack must be pushed firmly into the
interconnect-battery canister to ensure it completely engages with the
connectors at the back of the Interconnect-Battery canister.
Cache
Memory
Controller A (top)
Voltage
Regulator
Interconnect Canister
Left Power/Fan Canister Right Power/Fan Canister
Battery Packs
Front
Figure 1-9 6540 as seen from the top, showing the power distribution
The 6540 enclosure does not have a midplane (sometimes also referred to
as a backplane) that can be found in all pre-sixth generation SUN
StorageTek 6140 and 6130 products.
This diagram shows how the canisters are interconnected, and also gives
an overview of how the power distribution and battery system work.
The power from the left Power-Fan canister is distributed via controller B,
and power from the right Power-Fan canister is distributed via controller
A. Both controllers must be in place in order to provide redundant power
to each controller.
Data cache
● Optional 2, 4, or 8 GB of cache per controller
Faster memory
Host Channels
● Four independent 4 Gbps FC channels per controller (8 independent
ports per dual-controller system)
● Auto-negotiate to 1, 2, and 4 Gbps speeds
Drive channels
● Two 4 Gbps FC loop switches per controller
● Total of 8 drive loops per system
● Run at 2 Gbps and 4 Gbps
● Auto-detect drive side speed
● Can support both 2 Gbps and 4 Gbps drive enclosures behind the
same controller on different drive channels.
The 6540 command enclosure has two 6998 controllers. Both controllers
are identical. The controllers install from the rear of the command
enclosure. The top controller is controller A and the bottom controller is
controller B. All connections to the hosts and the drives in the storage
controller are through the controller canisters.
The host side connections support fibre-optic connections. The drive side
connections support either copper or fibre-optic connections.
Each 6540 controller canister provides the following connections and LED
indicators which are described in detail in the following sections:
● Four 4 Gbps Host Interface Ports
● Four 4 Gbps Disk expansion ports
● Dual 10/100 Base-T Ethernet Ports With EEPROM
● Serial Port Connector
● Seven segment display
● Controller service indicators
● AC or DC power (DC power connector present, but DC power not
currently implemented)
The 6540 storage system has eight 4 Gbps FC-AL host or FC-SW SAN
connections.
● The host side connections perform link speed negotiation on each
host channel port (also referred to as auto-negotiation) for 4, 2, or 1
Gbps FC host connectivity speeds resulting in minimal or no impact
on the existing storage network. Link speed negotiation for a given
host channel is limited to link speeds supported by the Small Form-
factor Pluggable (SFP) transceiver on that channel. The controllers
will enter into auto-negotiation at these points in time:
● Controller boot-up sequence
If the auto-negotiation process fails, the controllers will consider the link
to be down until negotiation is again attempted at one of these points in
time. For a 4-Gb controller, the supported link speeds are 1, 2, and 4Gbps.
Auto-Negotiation
The Fibre Channel host interface performs link speed negotiation on each
host channel Fibre Channel port. This process, referred to as auto-
negotiation, means that it will interact with the host or switch to
determine the fastest compatible speed between the controller and the
other device. The fastest compatible speed will become the operating
speed of the link. If the device on the other end of the link is a fixed speed
device or is not capable of negotiating, the controller will automatically
detect the operating speed of the other device and set its link speed
accordingly.
Each 6540 controller canister has two dual-ported drive channels. Each
6540 controller has two drive channels, each channel consists of two drive
loops, each drive loop has an external connection - so, each 6540 controller
has 4 drive side port connections.
The drive channels can operate at 2 Gbps or 4 Gbps. The drive channels
perform link speed detection (which is different than link speed
negotiation) - the controller will automatically match the link speed of the
attached drive enclosures. Drive channels can operate at different link
speeds, but both ports of a single channel must run at the same speed.
Two LEDs indicate the speed of the channel of the disk drive ports, as
shown in the figure below.
4 2
P1 Ch 2 (Ctrl B) P2
Ch 2 (Ctrl A)
The fibre channel port by-pass indicator has two settings: on and off.
Figure 1-19 shows the indicator.
Each Drive Channel has two independent drive loops, represented by the
two ports per Drive Channel.
Host and drive side cabling will be covered after a hardware overview of
the CSMII drive enclosure.
The 6540 has two RJ-45 ports per controller canister. Ethernet port 1 must
be for Management Host while port 2 reserved for future use. Do not use
this port for management of the trays.
To access the serial port, use a RS232 DB9 null modem serial cable. This
port is used to access the Service Serial Interface used for viewing or
setting a static IP address for the controllers. This interface can also clear
the storage system password.
Figure 1-21shows the RS232 DB9 null modem cable for serial port access.
Seven-Segment Display
Each digit of the numeric display has a decimal point, and is rotated 180
degrees relative to the other digit. With this orientation, the display looks
the same regardless of controller orientation. The numeric display as
shown in Figure 1-24 shows the tray identification (Tray ID) or a
diagnostic error code
The heartbeat is the small decimal on the lower right hand corner of the
1st digit - when the heartbeat is blinking the number displayed is the Tray
ID. The diagnostic light is the small decimal in the upper left hand corner
of the 2nd digit - when the diagnostic light is blinking the number
displayed is a diagnostic code.
Value Description
-- Boot FW is booting up
FF Boot Diagnostic executing
88 This controller/IOM is being held in reset by the
other controller/IOM
AA ESM-A application is booting
bb ESM-B application is booting
L0 Mismatched IOM types
L2 Persistent memory errors
L3 Persistent hardware errors
L9 Over temperature
H0 SOC (Fibre Channel Interface) Failure
H1 SFP Speed mismatch (2 Gb SFP installed when
operating at 4 Gb)
H2 Invalid/incomplete configuration
H3 Maximum reboot attempts exceeded
H4 Cannot communicate with the other IOM
H5 Mid-plane harness failure
H6 Firmware failure
H7 Current enclosure Fibre Channel rate different than
rate switch
H8 SFP(s) present in currently unsupported slot (2A or
2B)
● If no data is in cache and all cache data has been written to disk.
OFF.
● Data is in cache. ON (GREEN)
4) Explain how tray IDs are set. How can you change them?
1 2
5b) Which port should be used for normal
operation? __________________
8) If one drive port on one channel is set at 4 Gbps link speed and the
other is set at 2 Gbps what will be the speed for both ports?
10) Where can you find the “heart beat” of the controller?
11) What is the default controller tray ID that is set by the controller
firmware?
1) Identify the module, shown above. The module is the 6998 controller
canister module.
Using the letters, identify the parts of the component shown above
2a) A Host side ports
2b) B Ethernet ports
2c) C Controller Service Indicators (Service action allowed,
Service Action Required, Data in Cache)
2d) D 7 segment display for tray ID and fault identification
2e) E Drive side ports
2f) F Serial port
4) Explain how tray IDs are set. How can you change them?
Tray IDS are soft set by the controller to avoid tray ID conflicts. You can
change them through CAM or through the SCSS command line.
8) If one drive port on one channel is set at 4 Gbps link speed and the
other is set at 2 Gbps what will be the speed for both ports?
The port will interact with the host HBA or switch to determine the
fastest compatible speed between the controller and the other device.
10) Where can you find the “heart beat” of the controller?
On the lower right hand corner of the left box of the 7-segment display
11) What is the default controller tray ID that is set by the controller
firmware?
85
Objectives
Upon completion of this module, you should be able to:
● Provide an overview of the Sun StorageTek 6140 and its associated
management software
● Describe the hardware for the controller and expansion trays
● Describe the overall architecture of the controller and SBODs
The Sun StorageTek 6140 storage system is modular, rack mountable and
scalable from a single controller tray (CRM) to a maximum of six
additional expansion trays (CEM).
The Sun StorageTek 6140 storage system offers these new features:
● New technology:
● End-to-end 4-Gbit/sec FC
● Mix FC & SATA in tray
● More connectivity using 8 host ports (4 per controller)
● More density:
● 16 drives per tray
● 112 drives in 7 trays (1 controller tray and 6 expansion trays)
● More performance:
● 4 Gbyte cache (2 Gbyte per controller)
● 120K IOPs, 1500 Mbytes
● More Serviceability:
● Battery is a separate FRU
● RS232 interface
● On IOM
● Removable drive cage
Sun
Sun
StorEdge™ StorageTek
StorageTek
6130 6140
6140 Lite
Controller
CPU Processor 600 Mhz 667 Mhz 667 Mhz
Xscale w/ Xscale w/ Xscale w/
XOR XOR XOR
Host Ports 1/2 Gb 1/2/4 Gb 1/2/4 Gb
2 per ctlr 2 per ctlr 4 per ctlr
Expansion Ports 1 per ctlr 2 per ctlr 2 per ctlr
Controller Cache 1 GB per ctlr 1 GB per ctlr 2 GB per ctlr
Ethernet Ports 1 per ctlr 2 per ctlr 2 per ctlr
Controller 2882 3992 3994
Expansion Tray
IOM FC or SATA FC FC
# of Disk Drives per 14 16 16
Tray
Disk Types 1/2 Gb: FC, 2/4 Gb: FC, 2/4 Gb: FC,
SATA SATA II SATA II
# Expansion Trays 7 3 6
Maximum Disks 112 64 112
Sun
Sun
StorEdge™ StorageTek
StorageTek
6130 6140
6140 Lite
Disk Types 1/2 Gb: FC, 2/4 Gb: FC, 2/4 Gb: FC,
SATA SATA II SATA II
Configuration
Maximum Hosts 256 512 512
(256 (256
redundant) redundant)
Maximum Volumes 1024 1024 1024
Performance Targets
Burst I/O rate -
Cache Read 77,000 120,100 120,100
Sustained I/O rate -
Disk Read 25,000 30,235 44,000
Sustained I/O rate -
Disk Write 5,000 5,789 9,000
Burst throughput -
Cache Read 800 MBps 1,270 MBps 1,500 MBps
Sustained
throughput - Disk
Read 400 MBps 750 MBps 990 MBps
Sustained
throughput -
Disk Writes 300 MBps 698 MBps 850 MBps
✓ IOps or MBps?
The difference between the controller tray and the CSM2 are the controller
canisters and the Input/Output Modules (IOMs).
Figure 2-1 shows a block diagram for the Sun StorageTek 6140. The blocks
represent placement of drives, drive cage, power-fan canisters and either
the controller canister or IOM.
2 power-fan canister
2 controllers or IOMs
16 drives
Note – The preferred management tool for Sun Storage products is CAM.
However, you should be aware of the other tools and be familiar with
how to use them.
Hardware Overview
This section describes the main components of the Sun StorageTek 6140
controller tray (CRM) and the expansion tray CEM).
Controller Tray
The controller tray contains up to 16 drives, two controller canisters, two
power-fan canisters and a removable drive cage.
The front of the controller tray has a molded frame that contains global
lights and the Link Rate switch.
Figure 2-2 shows the front view of the Sun StorageTek 6140.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
The SATA II drives utilize the same canister as the FC drives, but since
SATA II drives are single-ported, an additional SATA II Interface Card
(SIC) is added to the rear of the canister. The card provides a fibre channel
connector and simulates a dual-port configuration, 3 Gbyte to 4 Gbyte
buffering and SATA II to FC protocol translation. The SATA II drive
negotiates between 2 Gbit and 4 Gbit based on the setting to the Link Rate
Switch on the tray.
The drives are removed by gently lifting on the lower portion of the
handle, which releases the handle.
DACstore
DACstore Benefits
DACstore exists on every drive and can be read by all Engenio controllers.
Therefore, when an entire virtual disk is moved from one storage system
to a new storage system, the data remains intact and can be read by the
controllers in the new storage system.
Sundry Drive
A Sundry drive stores the summary information about all of the drives in
the storage system. The controllers assign a minimum of three sundry
drives. Sundry drives are designated to hold certain global information
regarding the state of the storage system. This information resides within
the DACstore area on the sundry drives.
The field replaceable drive cage holds sixteen 3.5 inch drives. The mid-
plane is located on the back of the cage as shown in Figure 2-4.
1 2 3
Each component in a tray has LEDs that indicate functionality for that
individual component. Global LEDs indicate functionality for the entire
tray. Global LEDs are shown in Figure 2-6.
The Link Rate Switch shown in Figure 2-8 enables you to select the data
transfer rate between the IOMs, drives and controllers. Setting the Link
Rate switch determines the speed of the back end drive channel.
Caution – Change the Link Rate switch only when there is no power to
the CRM or CEM tray.
The Sun StorageTek 6140 lite with two host ports (3992 controller) is 4
Gbit capable, front and back. The 3992 auto-negotiates 1 Gbit, 2 Gbit and 4
Gbit speeds on the host side. With dual controllers, there are a total of 4
host ports per storage system. The 3992 controller has 1 Gbyte of cache
memory. The two expansion ports support 2 Gbit or 4 Gbit speeds
selected by the Link Rate Switch.
The Sun StorageTek 6140 with 4 host ports is 4 Gbit capable, front and
back. The 3994 controller auto-negotiates 1 Gbit, 2 Gbit and 4 Gbit speeds.
With dual controllers, there are a total of 8 host ports per storage system.
Each 3994 controller has 2 Gbyte of cache memory. The two expansion
ports support 2 Gbit or 4 Gbit speeds selected by the Link Rate Switch.
Caution – Never insert a 3992 controller and a 3994 controller into the
same unit. This will cause the storage system to become inoperable.
FC host ports
The fibre channel host port LEDs indicate the speed of the ports, as shown
in Figure 2-12.
1 4 2
Ch 1
The ports:
● Support speeds of 1/2/4 Gbit/sec using Agilent DX4+
● Auto-negotiate for speed
Host port LEDs - Two LEDs indicate the speed of the port
Auto-Negotiation
The Fibre Channel host interface performs link speed negotiation on each
host channel Fibre Channel port. This process, referred to as auto-
negotiation, means that it will interact with the host or switch to
determine the fastest compatible speed between the controller and the
other device. The fastest compatible speed will become the operating
speed of the link. If the device on the other end of the link is a fixed speed
device or is not capable of negotiating, the controller will automatically
detect the operating speed of the other device and set its link speed
accordingly.
1 2
Ethernet port 1 must be for Management Host while port 2 reserved for
future use. Do not use this port for management of the trays.
To access the serial port, use a RS232 DB9-MINI DIN 6 with a null modem
serial cable. This port is used to access the Service Serial Interface used for
viewing or setting a static IP address for the controllers. This interface can
also clear the storage system password.
Figure 2-15 shows the RS232 DB9-MINI DIN 6. Use with a null modem
cable for serial port access.
Two LEDs indicate the speed of the channel of the disk drive ports, as
shown in Figure 2-16.
4 2
P1 Ch 2 (Ctrl B) P2
Ch 2 (Ctrl A)
● When both LEDs are OFF, there is no Ethernet connection or the link
is down.
● With the first LED in the OFF position and the right LED in the ON
position, the port is at 2 Gbit/sec.
● When both LEDs are in the ON position, the port is at 4 Gbit/sec.
The fibre channel port by-pass indicator has two settings: on and off.
Figure 2-17 shows the indicator.
Seven-Segment Display
The heartbeat is the small decimal on the lower right hand corner of the
1st digit. The diagnostic light is the small decimal in the upper left hand
corner of the 2nd digit.
The expansion tray IDs are automatically set during power-on to avoid
conflicts with existing expansion tray IDs. Values on each display will be
shown as if the digits had the same orientation.
It is possible, however, that one controller may display the tray ID, while
the other controller displays a diagnostic code.
● Power on behavior - The Diagnostic Light, the Heartbeat Light, and
all 7 segments of both digits will be on if a power-on or reset occurs.
The tray ID display may be used to temporarily display diagnostic
codes after each power cycle or reset. The Diagnostic Light will
remain on until the tray ID is displayed. After diagnostics are
completed, the current tray ID will be displayed.
● Diagnostic behavior - Diagnostic codes in the form of Lx or Hx,
where x is a hexadecimal digit, are used to indicate state information.
In general, these codes are displayed only when the canister is in a
non-operational state. The canister may be non-operational due to a
configuration problem (such as mismatched IOM and/or controller
types), or it may be non-operational due to hardware faults. If the
controller/IOM is non-operational due to system configuration, the
controller/IOM Fault Light will be off. If the controller/IOM is non-
operational due to a hardware fault, the controller/IOM Fault Light
will be on.
Table 2-2 Numeric Display Diagnostic Codes
Value Description
-- Boot FW is booting up
FF Boot Diagnostic executing
88 This controller/IOM is being held in reset by the
other controller/IOM
Value Description
● If no data is in cache and all cache data has been written to disk
OFF.
● Data is in cache. ON (GREEN)
Battery
Figure 2-23 shows the controller with batter, which is to the right of the
controller.
Each day, the storage system controllers determine the age of each battery
package in the storage system by comparing the current date to the
installation date. If a battery package has reached its expiration age, cache
battery failure event notification will occur. The storage system can be
configured to generate cache battery near expiration event notification
prior to reaching the expiration age.
The controller module has a removable battery canister. The Lithium Ion
battery will need to be replaced every three years. It will hold data in
cache for up to 72 hours.
● They are able to operate in ranges from 90 VAC to 264 VAC (50
Hz to 60 Hz) or if the DC supply is selected, they will operate in
the range from –36VDC to –72VDC.T
● Two integrated +12V blower fans.
● If one blower fails, the second blower will automatically
increase to maximum speed to maintain cooling until a
replacement power supply is available.
● Blower speed control will be monitored and controlled by a
microcontroller and thermal sensor within the power supply.
Controller Architecture
Figure 2-27 illustrates the architecture for the controller.
___________________________________
3c) What are the function of the Leds to the far left and far right?
6) Explain how tray IDs are set. How can you change them?
1 2
8b) Which port should be used for normal
operation? __________________
Exercise Solutions
Task - Complete the Following
1) Identify the module, shown above. The module is the 3994 controller
module.
Using the letters, identify the parts of the component shown above
2a) A Host side ports
2b) B Ethernet ports
2c) C Service action allowed
2d) D 7 segment display for tray ID and fault identification
2e) E Drive side ports
2f) F Serial port
3c) What are the function of the Leds to the far left and far right?
The sundry drive contains information about the entire system. Whereas
all the other drives just contain their own information in the DACstore.
6) Explain how tray IDs are set. How can you change them?
Tray IDS are soft set by the controller to avoid tray ID conflicts. You can
change them through the SANtricity GUI or through the command line.
The 6140 lite controller has two host ports. The 6140 controller has four
host ports.
8b) Which port should be used for normal operation? Ethernet port 1.
Objectives
Upon completion of this module, you should be able to:
● Describe the Sun StorageTek CSMII Expansion Tray key features
● Identify the hardware components of the CSMII Expansion Tray
● Describe the functionality of the CSMII Expansion Tray
● Interpret LEDs for proper parts replacement
The 4 Gbps ready CSMII Expansion tray offers a new 16-bay disk
enclosure for attachment to selected mid-range 6000 storage systems, with
up to 4.8 terabytes (TB) physical capacity per expansion unit using sixteen
300 GB FC disk drives. The CSMII supports the current 2 Gbps FC drives,
and the intermix of 4 Gbps FC drives and SATA II drives, all within the
same enclosure. The CSMII contains redundant (AC) power and cooling
modules, and IOM interfaces.
Hardware Overview
This section describes the main components of the CSMII expansion tray.
Power/cooling
FC/SATA Drives
IOM
The front of the controller tray has a molded frame that contains global
lights and the Link Rate switch.
Figure 3-3 shows the front view of the Sun StorageTek CSMII.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
The SATA II drives utilize the same canister as the FC drives, but have a
SATA II Interface Card (SIC) added to the rear of the canister. The SIC
card serves three purposes:
4. provides redundant paths to the disk. SATA II drives are single-
ported so the SIC card acts as a multiplexer. and effectively simulates
a dual-ported disk
5. provides SATA II to FC protocol translation thereby enabling a
SATAII disk to function within an FC expansion tray.
6. provides speed-matching. The SIC card negotiates between 2 Gbps
and 4 Gbps based on the setting of the Link Rate Switch on the
expansion tray. SATAII drives run at 3 Gbps, the SIC card does the 3
Gbps to 4 Gbps buffering so the SATAII drive effectively runs at 4
Gbps speed (and similarly can run at 2 Gbps speed).
SATADisk
Drive
SATA
Connector
“Interposer
Card”
Intelligent
Controller
FC
Connector
The drives are removed by gently lifting on the lower portion of the
handle, which releases the handle.
DACstore
DACstore is a region on each drive that is reserved for the use of the
storage system controller. One can think of it as storage system
configuration metadata stored on each drive.
The DACstore area is created when the drives are introduced to the
controller. Each drive contains a DACstore area that is used to store
information about the drive’s state or status, volume state or status, and
other information needed by the controller. The DACstore region extends
512 Mbytes from the last sector of the disk.
DACstore Benefits
DACstore exists on every drive and can be read by all 6140 / 6540
controllers. Therefore, when an entire virtual disk is moved from one
storage system to a new storage system, the data remains intact and can
be read by the controllers in the new storage system.
Sundry Drive
A Sundry drive stores the summary information about all of the drives in
the storage system. The controllers assign a minimum of three sundry
drives. Sundry drives are designated to hold certain global information
regarding the state of the storage system. This information resides within
the DACstore area on the sundry drives.
The field replaceable drive cage holds sixteen 3.5 inch drives. The mid-
plane is located on the back of the cage as shown in Figure 3-5.
1 2 3
Each component in a tray has LEDs that indicate functionality for that
individual component. Global LEDs indicate functionality for the entire
tray. Global LEDs are shown in Figure 3-7.
The Link Rate Switch shown in Figure 3-8 enables you to select the data
transfer rate between the IOM’s, drives and controllers. Setting the Link
Rate switch determines the speed of the back end drive channel.
Caution – Change the Link Rate switch only when there is no power to
the CSMII tray.
At the back of the expansion tray, the IO modules (IOM) and the power-
fan canisters on the top are inverted 180 degrees from the canisters on the
bottom. In a fully configured system, the field replaceable canisters are
fully redundant. If one component fails, its counterpart can maintain
operations until the failed component is replaced.
The IOM has 4 drive ports. However, only two are available to use. Do
not use the drive ports (2A and 2B) nearest the seven-segment display.
These are reserved for future functionality. The IOM is 2 Gbit or 4 Gbit,
determined by the switch on the front side of the expansion tray.
Only disk drive ports 1A and 1B should be used. Two LEDs indicate the
speed of the channel of the disk drive ports, as shown in Figure 3-12.
1A 1B
The fibre channel port by-pass indicator has two settings: on and off.
Figure 3-13 shows the indicator.
The Seven Segment Display and Service Indicators shown in Figure 3-14
have the same function and definition as already described for the 6540
controller module - please refer to the 6540 module for a description of
these indicators.
The CSMII tray has two removable power-fan canisters. Each power-fan
canister contains one power supply and two fans. The four fans pull air
through the canister from front to back across the drives. The fans provide
redundant cooling, which means that if one of the fans in either fan
housing fails, the remaining fans continue to provide sufficient cooling to
operate the system. Cooling is improved by using side cooling for the
IOM’s.
● The two power supplies are tied to a common power bus on the
mid-plane using active current share between the redundant
pair.
● The power supplies have power-factor correction and support
wide-ranging AC or DC input.
● They are able to operate in ranges from 90 VAC to 264 VAC (50
Hz to 60 Hz) or if the DC supply is selected, they will operate in
the range from –36VDC to –72VDC.T
● Two integrated +12V blower fans.
● If one blower fails, the second blower will automatically
increase to maximum speed to maintain cooling until a
replacement power supply is available.
● Blower speed control will be monitored and controlled by a
microcontroller and thermal sensor within the power supply.
a. The power cord is plugged in, the power switch is on and the
power supply is not correctly connected to the mid-plane.
b. Power cord is plugged in, the power switch is on, the power
supply is correctly seated in the mid-plane, and a power supply
or blower fault condition exists.
Normal status OFF. Problem status ON (AMBER).
Architecture Overview
The following section shows the architecture for the CSMII expansion tray
which is a switched bunch of disks (SBODs).
Knowledge Check
A B C
D E
D E
1) Identify the module, shown above. The module is the CSMII IOM
module.
Using the letters, identify the parts of the component shown above
A Drive expansion ports
B IOM Service Indicators
C 7 segment display for tray ID and fault identification
D Serial port
E Reserved ports
The sundry drive contains information about the entire system. Whereas
all the other drives just contain their own information in the DACstore.
Objectives
Upon completion of this module, you should be able to:
● List the basic steps for installing the Sun StorageTek 6540
● Describe proper cabling techniques and methodologies
● List the basic steps of hot-adding CSMII drive trays to a 6540
● Perform the proper power sequence for the 6540 storage system
● Describe procedure to set static IP addresses for the 6540
The first three tasks will not be covered in this section. Utilize the
instructions for unpacking and physically installing the hardware that
ships with the product to complete the first three installation tasks.
1. Unpack the hardware according to the directions in the unpacking
guide that should be attached to the outside of the shipping carton.
2. Install the cabinet, controller tray and expansion trays by following
the directions in the hardware installation guide.
3. Attach the power cables.
4. Attach Ethernet cables - one to each controller.
5. Cable the controller and expansion trays.
6. Check the link rate switch.
7. Turn on the power.
8. Set the controllers’ IP addresses.
9. Use the hardware compatibility matrix to evaluate system set-up.
10. Attach the host interface cables.
Standard 19” cabinets can be customized for maximum flexibility and can
contain a combination of twelve enclosures. Always start loading the
cabinet from the bottom up. Always push the cabinet from the front.
Cabling Procedures
The following section highlights proper cabling methods for the controller
and expansion trays, keeping in mind how to cable for redundancy.
Cable Types
Fiber-optic cables and small form-factor pluggable (SFP) transceivers are
used for connections to the host. If the system will be cabled with fiber-
optic cables, you must install active SFPs into each port where a cable will
be connected before plugging in the cable.
Note – Host connections require the use of fiber-optic cables but either
copper or fiber-optic cables can be used to connect expansion trays.
The choice between optical fiber and electrical (or "copper") transmission
for a particular system is made based on a number of trade-offs. Optical
fiber is generally chosen for systems with higher bandwidths, spanning
longer distances, than copper cabling can provide. The main benefits of
fiber are its exceptionally low loss, allowing long distances between
amplifiers or repeaters; and its inherently high data-carrying capacity,
such that thousands of electrical links would be required to replace a
single high bandwidth fiber.
Typically copper cables are used for short distances, such as inter-
connecting drive enclosures. Fibre cables are used for long distances, such
as connecting the storage system directly to servers or to a FC switch.
Fiber-Optic Connection
Fiber-optic cable
Copper Connection
Passive SFP transceiver
(integrated with cable)
Copper cable
Cable Considerations
Caution – Fibre-optic cables are fragile. Do not bend, twist, fold, pinch, or
step on the fiber-optic cables. Doing so can degrade performance or cause
loss of data connectivity.
If both redundant drive loops are cabled in the same direction, then a loss
of power or communication to one drive enclosure can result in loss of
access to the remaining drive enclosures.
Figure 4-4 If a drive enclosure fails, the remaining drive enclosure can
be accessed with Drive Channel 3.
When attaching expansion trays, create a cabling topology that uses the
redundant paths to eliminate inter-tray connections as a potential single
point of failure. To ensure that the loss of an expansion tray itself does not
effect access to other trays, cable one drive channel from controller A of
the 6540 top-down, and one drive channel from controller B bottom-up.
Thus, the loss of a single tray will not prohibit access to trays on the other
side of the failure from being accessed by the other path.
Figure 3-4 shows full redundancy cabling on the drive channel side. Each
drive tray is cabled to both controllers - i.e. from each drive tray, one IOM
is cabled to Controller A, and the other IOM is cabled to Controller B.
Drive Channel 1 from Controller A is cabled top down. The redundant
loop, Drive Channel 3 from Controller B is cabled bottom up. Even if a
whole drive tray fails, the connection to all other drive trays is not lost.
Each 6540 controller has two drive channels, and dual expansion ports for
each drive channel. Splitting the trays between drive channels, or between
each ports of a single channel further isolates the effect of a tray failure by
half.
Figure 4-5 6540 best practice for creating redundant drive side loops
Balancing drive trays also provides some additional tray loss protection if
Virtual Disks are properly configured across enclosures.
Note – This is only a high level overview of the procedure, please refer to
the appropriate user documentation for details.
1. Install the the new drive enclosure in the rack but do not apply power
to it yet.
Figure 4-7 Hot Add step 1 - install new drive enclosure in the rack but
do not apply power to it yet
2. Add the new enclosure to the top-down loop (in this example Drive
Channel 1)
5. Re-cable the bottom-up loop to include the new enclosure (in this
example Drive Loop Channel 3)
Stack 2 Stack 3
Stack 1 Stack 4
Figure 4-10 A fully configured 6540 with 14 CSMII expansion enclosures
Cabling Summary
● Have two fibre channel drive loops to each drive enclosure for
redundancy. One drive loop from Controller A to the left side IOM of
a drive enclosure. The redundant drive loop from Controller B to the
right side IOM of a drive enclosure.
● Have drive loops travel in opposite directions across all of the drive
enclosures on those loops for robustness in case of a drive enclosure
failure
● Use all of the drive side channels (drive loops) available for
improved performance
● From the controller enclosure, cable to the 1B port of the drive
enclosure IOM.
● Use odd numbered drive channels as a redundant pair, and even
numbered drive channels as a redundant pair.
Each controller also has dual expansion ports. Splitting the trays between
each port or channel further isolates the effect of a tray failure by half.
The first path is created by cabling the expansion trays sequentially from
Controller A. For example, Controller A is connected to expansion tray 1
through Port 1B, which is connected to expansion tray 2 through Port 1B,
which is connected to expansion tray 3 through Port 1B, which is
connected to expansion tray 4 through Port 1B.
The alternate path is created by cabling the drive modules in the reverse
order from Controller B. For example, Controller B is connected to
expansion tray 4, which is connected to expansion tray 3, which is
connected to expansion tray 2, which is connected to expansion tray 1.
In the event that expansion tray 2 fails, expansion tray 3 and 4 are still
accessible through the alternate path. While identical cabling topologies
are simpler, a single point of failure exists. If a expansion tray fails, all
expansion trays beyond the failure are no longer accessible. This topology
is vulnerable to loss of access to data due to an expansion tray failure.
A B
Expansion Trays
Controller Tray
Figure 4-11 Redundant Cabling With One Controller Tray and Four
Expansion Trays
✓ above?
Expansion Tray
Controller Tray
Expansion Tray
Expansion Tray
Controller Tray
Expansion Tray
Expansion Tray
Expansion Tray
Controller Tray
Expansion Tray
Expansion Tray
Expansion Tray
Expansion Tray
Controller Tray
Expansion Tray
Expansion Tray
Expansion Tray
Expansion Tray
Expansion Tray
Controller Tray
Expansion Tray
Expansion Tray
Expansion Tray
Expansion Tray
Expansion Tray
Expansion Tray
Controller Tray
Before powering on the system, check to see if the Link Rate switch is set
to the appropriate data transfer rate. If the Link Rate switch is not set to
the correct data transfer rate, move the switch to the correct position.
● 4 Gbps to the left
● 2 Gbps to the right
Since the switch is recessed, you will need to use a small tool to slide the
switch to the proper position.
Before powering on the system, check to see the link rate switch is set the
appropriate data transfer rate.
Note – All expansion trays and the controller tray must be set to operate
at the same data transfer rate.
The controllers read the storage system configuration from the DACstore
on the drives, therefore all drives need to have power to them before the
controllers are turned on. The first thing a controller does is issue a Drive
Spin Up command to each drive. After all drives are spun up, the
controller goes out and reads the DACstore information from each drive.
Power-on Procedure
Note – To diagnose and correct the fault, you may need help from the
storage management software. The use of storage manager to recover
from faults will be covered in a later section.
Power-off Procedure
1. Stop all I/O activity to each tray you are going to power off.
Note – Always wait until the Cache Active light on the back of the
controller tray turns off and all drive active lights stop blinking before
turning off the power.
2. Check the lights on the back of the controller and expansion trays.
a. If one or more fault lights are on, do not continue with the
power-off procedure until you have corrected the fault.
3. Turn off the power switches on each fan-power canister in the
controller tray.
4. Turn off the power switches on each fan-power canister in each
expansion tray.
Note – Power on: first expansion trays then controller tray. Power off: first
the controller tray then the expansion trays.
Note – Each controller must have its own IP address. The default IP
address for controller port A1 is 192.168.128.101. The default IP address
for controller port B1 is 192.168.128.102.
There are several ways to change the controller’s default IP address to the
desired static IP addresses.
1. Connect the controller tray directly to a management host using a
cross-over Ethernet cable and change the IP address using the
management software CAM.
2. Connect the controller tray to a management host using an Ethernet
hub and change the IP address using the management software
CAM.
3. Connect the controller tray on an existing subnet and change the IP
address using the management software CAM.
4. Utilize the Serial Port Service Interface through the serial port
This interface:
● Displays network parameters
● Sets network parameters
● Clears the storage system password.
To connect to the 6540 serial port, use the null-modem cable. This should
be supplied with the controller tray.
Figures Figure 4-21 and Figure 4-22 show sample screens of the Service
Interface Main Menu and the Ethernet Port Configuration screen.
If you answer “Y” to configure using DHCP, the system tries for 20
seconds to connect to the DHCP server. If no DHCP server is found, the
system cycles back to the main menu.
Always utilize the hardware compatibility matrix to verify that all SAN
components are certified with the Sun StorageTek 6540. The components
include data host, OS version, Host Bus Adaptors and switches. Always
verify firmware levels and bios settings for new systems or firmware
upgrades.
Note – Refer to the SUN web for the Interoperability Matrix. As of 3/2007
the URL for the Interoperability Matrix is:
https://extranet.stortek.com/interop/interop
Caution – If you will be using Remote Replication, for the 6998 controller,
do not use host port 4 on both controller A and Controller B. When
Remote Replication is activated, host port 4 on each controller is reserved
for Replication and any data host connected will be logged out.
Before you connect data hosts directly to the system, check that the
following prerequisites have been met:
● Fiber-optic cables of the appropriate length are available to connect
the array host ports to the data host HBAs.
● Redundant connections from the host to each controller module are
available.
● Certified failover software is enabled on the host.
Before you connect data hosts, check that the following prerequisites have
been met:
● The FC switches are installed and configured as described in the
vendor’s installation documentation.
● Redundant switches inherently provide two distinct
connections to the storage system.
● Interface cables are connected and routed between the host bust
adapters (HBA’s), switches and installation site.
● Fiber-optic cables of adequate length are available to connect the
array to the FC switch.
CH1 CH2
4 3 2 1
1 2 3 4
CH3 CH4
Knowledge Check
the diagram below, design a cabling scheme for the Sun StorageTek
Objectives
Upon completion of this module, you should be able to:
● Describe the Sun StorageTek Common Array Manager (CAM)
● Explain the function of the main components of CAM
● List when and where to install each component of CAM
● Explain the function of firmware and NVSRAM files
● Describe the out-of-band method of management used by CAM
● Describe logging into and navigating within CAM
SMI-S Overview
Management today is a myriad of different software packages by different
vendors that are not coordinated with each other. Furthermore, many of
these applications are deficient in the necessary functionality, security, and
dependability needed to ensure greater business efficiency. Incompatible
Application Program Interfaces (API) for storage management spread
throughout today's multi-vendor SANs. The Storage Management
Initiative Specification (SMI-S) assists administrators to gather and
examine data from dissimilar vendors' products, and puts it in a common
format. This lets storage managers manipulate all devices on the SAN
from a centralized application.
There are two current versions of SMI-S, version 1.01 and version 1.0.2.
The Common Array Manager uses version 1.0.2. This will allow for the
most up to date support for SAN infrastructures. SMI-S gives
heterogeneous vendor support, functionally rich, dependable, and secure
monitoring/control of mission essential resources. This interface sweeps
away the deficiencies related with legacy management.
Software Components
The management software is delivered on compact disk (CD) or can be
dowloaded from the Sun website. The management software consists of a
number of components:
● Sun StorageTek Management Host Software (CAM)
● Sun StorageTek Data Host Software
● Sun StorageTek Remote Management Host Software
Data host software controls the data path between the data host and the
array. The data host software contains tools that manage the data path
I/O connections between the data host and the array. This includes
drivers and utilities that enable hosts to connect to, monitor, and transfer
data in a storage area network (SAN).
The type of data host software you need depends on your operating
system. You must obtain the data host software from the Sun Download
Center or other source.
The data host software in the past has consisted of the following:
● Sun StorageTek SAN Foundation Software - Used to manage the data
paths between data hosts and the storage arrays. This software
includes drivers and utilities that enable data hosts to connect to,
monitor, and transfer data in a Storage Area Network (SAN).
● Sun StorageTek Traffic Manager software - Provides multipathing
functionality and the ability to reliably communicate with the
storage array.
The Remote Management Host software contains only the CLI. This can
be installed on remote Solaris and non-Solaris systems allowing users to
manage the storage array remotely.
Use of the Remote Management Host Software still requires the use of a
CAM managment host. The CLI simply communicates with the CAM
management host to perform desired tasks.
When installing a new version of CAM, the CAM software will confirm
that the firmware on the controller is compatible with the version of
CAM. All versions of CAM are backwards compatible, which allows
higher levels of CAM to manage storage systems running on lower levels
of firmware. If the detected array is not at a baseline firmware level, the
firmware can be upgraded during the array registration process, or at a
later time. Consult the CAM installation and support guide and the
compatibility matrix for more information. Firmware resides on each
controller, each IOM and each disk drive.
Note – Full array management requires that both controllers are accessible
via Ethernet. If only one controller is accessible then only a subset of the
array management functions will be available.
Note – Multiple users can be logged into the CAM management server
concurrently.
CAM Server
Fabric
Controllers Controllers
Ethernet Connections Ethernet Connections
Firmware/NVSRAM Firmware/NVSRAM
FC I/O Path
Management Host - used to manage the storage system. This can be any
host that has a network connection to the storage system and has the
CAM Management Host Software installed.
Data Host - used to read and write data to the storage system. This can be
any host that has a FC connection to the storage system and has the CAM
Data Host Software installed.
Hosts that have both network and FC connection to the storage system
can act as both Management and Data hosts.
Before you start the installation, ensure the following requirements are
met:
● The root password of the management host is available (for running
the installation script). Note that the root password is required for
the initial login to the Sun Java Web Console after the software is
installed.
● The following amount of space will be required for the installation:
● 555 Mbytes on Solaris
● 660 Mbytes on Linux
● 530 Mbytes on Windows
Note – Review the release notes for the most up to date list of supported
operating systems.
The installation wizard provides two choices for installation: typical and
custom. In a typical installation, the management host software, the data
host software and the Sun StorageTek Configuration Service packages are
installed. If the custom installation is selected the user can choose the
packages to be installed. Select the custom option only if you want to
specify the installation of only the management host software or only the
data host software, or if you also want to install the Remote Management
Host software.
Note – During the software installation, the progress indicator may reflect
0% for a considerable portion of the installation process. This is the
expected progress indication when the “typical” installation process is
selected.
Most screens are broken into three sections: the banner, the navigation tree
and the content area.
● Log Out - Logs the current user out and then displays the Sun Java
Web Console login page
● Help - Opens the online help system
The quick status display on the left of the banner provides the current
user’s role and the server name. The display on the right provides the
number of current users logged in, the date and time of the array was
refreshed (by the refresh button), and current alarms.
Note – Not all of the navigation aides in Table 5-1 are available in every
content screen.
Icon/Indicator Description
Icon/Indicator Description
Note – Unless otherwise specified, all steps detailed are for the Sun
StorageTek Configuration Service interface.
Configure IP Addressing
To configure the IP address for each controller's Ethernet port, an IP
connection between the controller trays and a management host must
already have been established using the controllers default IP addresses. It
is important that both controllers are configured with an IP address to
ensure proper function. The controller's Ethernet ports can be configured
with either a dynamic or a static IP address.
The Sun StorageTek 6x40 array has the following default internal IP
addresses for the first port:
● Controller A: 192.168.128.101
● Controller B: 192.168.128.102
3. From the Sun Java Web Console page, click Sun StorageTek
Configuration Service.
4. Click the Register button.
6. Verify that the array(s) have been added to the Array Summary
page.
If the array is not displayed, check the hardware connections and
ensure that the array can be contacted using the ping command.
7. Select the array for which you want to modify the IP addresses.
8. Click Administration.
9. The General Setup page will be displayed. Enter the array name and
default host type and then click OK.
10. Click Administration > Controllers.
11. For Controller A's Ethernet ports select Specify Network
Configuration and then enter the IP address, gateway address, and
subnet, then Click OK.
12. Repeat steps 10 and 11 for controller B’s Ethernet ports.
Note – An error message indicating that contact has been lost with the
array may be displayed as a result of the changed IP address. This is
expected due to the change.
Then log in to the Common Array manager as root using the root
password for the managment host (on Solaris this is typically login root,
and password root, on Windows this is typically login Administrator and
password is whatever was specified.). Then select Sun StorageTek
Configuration Service from the Storage section of the Sun Java Web
Console page.
Naming an Array
The storage array will come with a default name which you should
change to a unique name to simplify identification. The Array Name can
be changed on the Administration Details page.
If an array is moved from one management host to another, the user will
need to provide the password when registering the array that was moved.
Note – The password can be unique between different arrays on the same
management host. However, if a single array is being managed by more
than one management host, the password must be the same on each
management host.
To be eligible for privileges to the CAM interface users must have valid
Solaris or Windows user accounts and have access to the management
host. The users can then log in to the CAM interface using their Solaris or
Windows user names and passwords.
If multiple users are logged in to the array with Storage privileges there is
a risk of one user’s changes overwriting those of another user’s. For this
reason Storage administrators should develop procedures to manage this
risk.
Objectives
Upon completion of this module, you should be able to:
● List the configuration components of CAM
● List the functions available in CAM
● Describe the parameters that affect a volume
Host Group
Based on a
Storage Profile
Other
vDisk and volumes
Storage Profiles
A storage profile consists of a set of attributes that are applied to a storage
pool. Each disk or Virtual Disk must meet the attributes defined by the
storage profile to be a member of a storage pool. The use of storage
profiles simplifies configuration by configuring the basic attributes that
have been optimized for a specific application or data type. Prior to
configuring an array it is important to review the available storage
profiles and ensure a profile exists that matches the users targeted
application and performance needs - if not, the user can create a new
storage profile.
Read-
RAID Segment Drive Number of
Name Ahead
Level Size Type Drives
Mode
Read-
RAID Segment Drive Number of
Name Ahead
Level Size Type Drives
Mode
To view the Storage Profile Summary screen select Profiles from the
navigation pane.
Note – The last profile listed is Test. The Test profile is a custom profile
and is selectable by clicking on the check box to the left of the profile
name.
Name is the unique identifier for the storage profile. The profile name can
be up to 32 characters.
Description is a typed description of the profile. This parameter is
optional.
RAID Level can be 0, 1, 3, 5 or 10. This is the RAID level that will be
configured across all disks within a virtual disk.
Segment size is the amount of data, in kilobytes (KB) that the controller
writes on a single drive in a Volume before writing data on the next drive.
Data blocks store 512 bytes of data and are the smallest units of storage.
The size of a segment determines how many blocks it contains. For
example, an 8 KB segment holds 16 data blocks. A 64 KB segment holds
128 data blocks.
Read ahead allows the controller, while it is reading and copying host-
requested data blocks from disk into the cache, to copy additional data
blocks into the cache. This increases the chance that a future request for
data could be fulfilled from the cache.
Setting this value to Disabled will turn off read ahead. Setting this value
to Enabled tells the controllers to determine the most optimal multiplier
value.
Number of Disks can be set to a value of between 1 and 30, or to the value
Variable. This parameter specifies the number of disks to be grouped
together in a virtual disk. For example, if you create a storage pool with a
profile that has the number of disks parameter set to a number, all virtual
disks that are part of that storage pool must have the same number of
disks. If the number of disks parameter is set to the Variable value you are
prompted for the number of disks when storage is added to the pool.
Note – Tray loss protection is achieved when all the drives that comprise
the Virtual Disk are located in different expansion trays.
Disk Type specifies the drive type to be used for the volume. It can be set
to FC, SATA or Any. Mixing drive types (SATA or Fibre Channel) within a
single virtual disk is not permitted.
If disk drives available have different capacities and/or different speeds,
the overall capacity of the Virtual Disk will be based on the smallest
capacity drive and the slowest drive.
Storage Pools
An array can be divided into storage pools. Each pool is associated with a
profile and acts as a container for volumes or physical storage devices that
meet the storage profile. This allows users to optimize each storage pool
to the type of application that it will be used with.
Note – Removing a storage pool destroys all stored data in the pool and
deletes all volumes that are members of the pool. The data can be restored
from backup after new storage pools are added, but it is far easier to
avoid the difficulty in the first place.
Volumes
A volume is a “container” into which applications, databases, and file
systems can store data. A volume is created from a Virtual Disk that is
part of a storage pool. The creation of a volume is comparable to
partitioning a disk drive, in that a volume is a part of a Virtual Disk.
Volume Parameters
Cache Settings
Write caching allows write operations from the host to be stored in cache
memory. Unwritten volume data in cache is written to disk, or flushed,
automatically every 10 seconds.
Virtual Disks
During the configuration of a volume, the Common Array Manager
creates a Virtual Disk automatically.
Virtual disks are created and removed indirectly through the process of
creating or deleting volumes or snapshots. A Virtual Disk is the RAID set
which contains the specified number of disks and is created based on the
RAID level assigned in the storage profile. The disk drives that participate
in the virtual disk must all be of the same type, either Serial Attached
Technology Advancement (SATA) or Fibre Channel (FC).
Array Name
Note – The storage management software does not check for duplicate
names. Verify that the name chosen is not already in use by another
system.
The host type defines how the controllers in the storage array will work
with the particular operating system on the data hosts that are connected
to it when volumes are accessed. The host type depicts an operating
system (Windows 2000, for example) or variant of an operating system
(Windows 2000 running in a clustered environment). Generally, you will
use this option only if all hosts connected to the storage array have the
same operating system (homogeneous host environment).
If you are in an environment where there are attached hosts with different
operating systems (heterogeneous host environment), you will define the
individual host types as part of creating Storage Domains.
Hot Spares
Your storage system volume remains online and accessible while you are
replacing the failed drive, since the hot spare drive is automatically
substituting for the failed drive.
Reconstruction is the process of reading data from the remaining
drives and the parity drive. This data is processed through an XOR
operation to recreate the missing data. This data is written to the hot
spare.
Copy-back is the process of copying the data from the GHS drive to
the drive that has replaced the failed drive. The time to reconstruct
the GHS drive varies and depends on the activity of the storage
system, the size of the failed volume and the speed of the drives.
A hot spare drive is not dedicated to a specific volume group but
instead is global which means that is can be used for any failed drive
in the storage system with the same or smaller capacity. Hot spare
drives are only available for a RAID level 1, 3, or 5 volume group.
When creating a global hot spare, keep the following in mind:
Select a drive with a capacity equal to or larger than the total
capacity of the drive you want to cover with the hot spare.
Generally, you should not assign a drive as a hot spare unless its
capacity is equal to or greater than the capacity of the largest drive in
the storage system.
The maximum number hot spare drives per system is 15.
There are cache settings that can be set at the storage system level that are
in effect for all volumes in the storage system.
Cache start and stop percentages - The start value (percentage) indicates
when unwritten cache data should be written to disk (flushed). The stop
value (percentage) indicates when a cache flush should stop
When the cache holds the specified start percentage of unwritten data, a
flush is triggered. When the cache flushes down to the specified stop
percentage, the flush is stopped. For example, you can specify that the
controller start flushing the cache when the cache reaches 80% full and
stop flushing the cache when the cache reaches 16% full.
Note – Unwritten writes are written to disk every 10 seconds. This is not
affected by the cache settings. For best performance, keep the start and
stop values equal.
The cache block size indicates the cache block size used by the controller
in managing the cache. For the 6540 the default cache block size is set to
16 KB and cannot be modified.
This parameter is applied to the entire storage system. The Cache Block
Size is for all volumes in the storage system. For redundant controller
configurations, this includes all volumes owned by both controllers
within the storage system.
● 4 KB (a good choice for file system or database application use)
● 16 KB (a good choice for applications that generate sequential I/O,
such as multimedia)
Disk Scrubbing
When enabled, it runs on all volumes in the storage system that are
● optimal
● have no modification operations in progress
The disk scrubbing Interval specifies the number of days over which the
media scan should be run on the eligible Volumes. The controller uses the
duration period, in conjunction with its knowledge of which Volumes
must be scanned to determine a constant rate at which to perform media
scan activities. This rate is maintained regardless of host I/O activity.
By default this parameter is not enabled. Additional disk scrubbing options exist
for individual volumes.
The Failover Alert Delay specifies at the time that a critical event be
logged when a Volume is transferred to a non-preferred controller. A
value of 0 will create a log entry immediately.
System Time
The System Time option synchronizes the storage system controller clocks
with the storage management station. This option ensures that event
timestamps written by controllers to the Event Log match event
timestamps written to host log files. Controllers remain available during
synchronization. You also have the option to manually set the date and
time.
Manage Passwords
Note – If you have forgotten the password, contact your customer support
representative.
Knowledge Check
1. You can mix drive types (SATA and Fibre Channel) in a single tray.
True False
9. What can happen if you do not set your controller clocks to match
your management station?
10. What part of the storage system takes advantage of the cache block
size? What does it do with it?
Storage Domains
Objectives
Upon completion of this module, you should be able to:
● Explain the benefits of Storage Domains
● Define Storage Domains terminology
● Describe the functionality of Storage Domains
● Calculate Storage Domain usage
Figure 7-1 Partitioning one physical storage system into several virtual
storage systems.
Storage
H Port1
Partition 1 B LUN 0 Volume 0
A Port2
LUN 1
Volume 1
Host
H Port1
B Volume 2
A Port2 LUN 0
Storage LUN 5 Volume 3
Partition 2
H Port1 Storage
B
A Port2
System
Host Group
number during Volume creation (else the Volume has a State of Free
waiting to be assigned a LUN number).
A Default Storage Domain exists to include the following:
● All host groups and hosts that do not have a volume
explicitly mapped to it.
● All volumes that have a default volume-to-LUN mapping
assigned.
● All automatically detected initiators.
Any volumes within the default storage domain can be accessed by all
hosts and host groups within that storage domain.
Creating an explicit volume-to-LUN mapping for any host or host group
and volume within the default storage domain causes the management
software to remove the specified host or host group and volume from the
default storage domain and create a new separate storage domain.
Host Group: a label for one or most hosts that need to share access to a
Volume.
Characteristics: Define a Host Group only if you have two or more
hosts that will share access to the same Volumes.
Host: a label for a Host that contains one or more FC ports that are
connected to the storage system
Characteristics: A host is a computer that is attached to the storage
system and accesses various Volumes on the storage system
through its host ports (host adapters).
Storage System
Note – A host group or host can access Volumes with either default
mappings or specific mappings, but not both.
● When using the Storage Domains wizard, there is only one Volume
to LUN mapping allowed. If more than one server needs to access a
single Volume, a host group should be used. All servers in a host
group can access all the Volumes in that domain.
● Server clusters need to use host groups, so that all the servers can
share access to the same Volumes. But, servers in a host group do not
necessarily need to run clustering software. Keep in mind, however,
that without file sharing software, all servers in a host group can
access the same Volumes, which can lead to data integrity issues.
● A given host can be part of a host group and have its own individual
mappings.
● Each host has its own LUN address space within a domain. Meaning
the same LUN number can be used in multiple Volume to LUN
mappings, just not in the same domain.
LUN 1
50 Volume
GB Marketing
If we look at what the host sees, it has a LUN number that maps to a
Volume on the storage system. Each host can only see the Volume in its
domain. For instance, host A has two LUNs that map to two Volumes
(shown in red). It has no idea that there is additional capacity on the
storage system. The same is true for host B, which sees only its two blue
Volumes, and host group C, which sees only its three green Volumes.
Unmapped Volumes can be assigned to any of the domains.
Port WWN
50 Volume ID
GB
Figure 7-5 The two Volumes mapped to Host A are on different Virtual
Disks, and could even be on different drive technologies (A1
on FC and A3 on SATA)
It doesn’t matter where the Volume resides within the storage system.
Storage Domains’s volume-to-LUN mapping implementation creates
valuable flexibility for the storage administrator, as any available Volume
can be mapped to any attached server. So, while the individual server sees
a virtual storage system that consists of only their mapped LUNs, the
physical Volumes can be intermixed throughout the storage system within
one or more RAID Virtual Disks.
The previous diagram showed Host A had two red Volumes that
comprised its domain. On this diagram, you can see those two Volumes
reside on two different RAID Virtual Disks. Volume A, which is Host A’s
LUN 0, is in Virtual Disk A1, and volume G, which is Host A’s LUN1, is in
Virtual Disk A3.
Now, the Volume(s) will be ready for use by the host (or host groups).
Knowledge Check
True or False
1. A storage domain is created when a host group or a single host is
associated with a Volume-to-LUN mapping.
True False
2. A host group or host can access volumes with default mappings and
specific mappings.
True False
3. You can not use the same LUN number in more than one Volume-to-
LUN mapping.
True False
True False
Multiple Choice
1. After defining the first specific Volume-to-LUN mapping for a host,
a) Host ports must be defined
b) the host type can no longer be changed
c) The LUN number can not be used by other hosts in the
topology
d) The host and host ports move out of the Default host group
2. In a heterogeneous environment,
a) Each host type must be set to the appropriate operating
system during host port definition
b) Volumes can have more than one Volume-to-LUN number
c) Hosts with different operating systems can share volumes
d) A host can access volumes with either default mappings or
specific Volume-to-LUN mappings.
Customer Scenario
Mr. Customer has the 3 servers and one storage system (6540). The servers:
W2003 (two single ported HBA's), Linux (one dual ported HBA) and AIX
(one single ported HBA). The W2003 server will be running the Exchange
application and the Exchange Administrator has requested 2 'drives' - one
for the database, the other for a log file. The Linux server will be used for
software development and will require disk space for source code and
development tools (2 volumes). The AIX server will be running the
engineering document database and will require 1 volume. The Finance
Dept. has requested a 'disk' for storing employee expense statements. The
application to access the employee expense statement will run on all the
servers.
First draw a diagram showing the servers and the storage, so you and the
customer have the same understanding of the requested configuration.
List the Hosts that will be created under each Host Group:
How many partitions would you need for the configuration below?_______
Objectives
Upon completion of this module, you will be able to:
● Explain the data presented by the CAM built-in Performance
Monitor
● List the factors that influence storage system performance
● Explain how cache parameters impact performance
● List the Dynamic functions and explain how they impact
performance
This section explains the 40% of hardware set-up that consists of fine
tuning the configuration of the storage system.
Stack 2 Stack 3
Performance Tip- Preferred Ownership for Volume
Stack 1 Group and Volumes: Stack 4
A - Stacks 1 & 3
B – Stacks 2 & 4
Cabling
Cabling the drive trays as above ensures redundancy and utilizes all
available backend channels. By utilizing all channels the maximum
aggregate bandwidth can be obtained.
Performance Monitor
The statistics displayed in each page is described in the CAM on-line help.
Fine Tuning
The following describes how some of the data fields can be used to
analyze the performance of the storage array.
Total IOPS - this data field is useful for monitoring the I/O activity to a
specific controller and a specific volume. This field helps you identify
possible I/O "hot spots."
If I/O rate is slow on a volume, try increasing the number of drives in the
virtual disk by using the DCE option.
If you notice the workload across the storage subsystem (Total IOPS
statistic on the Performance Monitoring page) continues to increase over
time while application performance decreases, this might indicate the
need to add additional storage arrays to your enterprise so that you can
continue to meet application needs at an acceptable performance level.
Because of the dependency on I/O size and transmission media, the only
technique you can use to improve transfer rates is to improve the I/O
request rate. Use host operating system utilities to gather I/O size data so
you understand the maximum transfer rates possible. Then use tuning
options available in the storage management software to optimize the I/O
Request Rate so you can reach the maximum possible transfer rate.
Average IOPS - Factors that affect I/Os per second include access pattern
(random or sequential), I/O size, RAID level, segment size, and number of
drives in the virtual disks or storage subsystem. The higher the cache hit
rate, the higher I/O rates will be.
Performance improvements caused by changing the segment size can be
seen in the I/Os per second statistics for a volume. Experiment to
determine the optimal segment size or use the file system or database block
size.
Higher write I/O rates are experienced with write caching enabled
compared to disabled. In deciding whether to enable write caching for an
individual volume, consider the current and maximum I/Os per second.
You should expect to see higher rates for sequential I/O patterns than for
random I/O patterns. Regardless of your I/O pattern, it is recommended
that write caching be enabled to maximize I/O rate and shorten
application response time.
If you notice that the Total IOPS or Average IOPS is not as expected then
a factor might be host-side file fragmentation. Minimize disk accesses by
defragmenting your files. Each access of the drive to read or write a file
results in movement of the read/write heads. Make sure the files on your
volume are defragmented. When the files are defragmented, the data
blocks making up the files are contiguous so the read/write heads do not
have to travel all over the disk to retrieve the separate parts of the file.
Fragmented files are detrimental to the performance of a volume with
sequential I/O access patterns.
Polling Interval
The frequency that the performance data is obtained from the storage
array is controller by the Polling Interval. Each time the polling interval
elapses, the Performance Monitor re-queries the storage array for
performance statistics.
If you are monitoring the array via CAM, update the statistics frequently
by selecting a short polling interval, for example, 3 or 5 seconds.
If you are saving results to a file to look at later via SSCS, choose a slightly
longer interval, for example, 30 to 60 seconds, to decrease the system
overhead and the performance impact.
Raid 0
Advantages: Good for reads, small I/Os, many concurrent I/Os and
random I/Os.
Important:
● You cannot cancel this operation after it begins.
● Your data remains available during this operation.
● The virtual disk must be in an Optimal state before you can perform
this operation.
If you do not have enough capacity in the virtual disk to convert to the
new RAID level, you will receive an error message and the operation will
not continue. If you have unassigned drives, add additional drives to the
virtual disk. Then, retry the operation.
RAID 1 - Minimum of two drives then after that you must have an even
number of drives in the virtual disk. If you do not have an even number
of drives and you have some remaining unassigned drives, use the Virtual
Disk>>Add Free Capacity option to add additional drives to the virtual
disk, then retry the operation.
RAID 1+ 0 - Minimum of four drives and then after that an even number
of drives.
If you do not have enough capacity in the virtual disk to convert to the
new RAID level, you will receive an error message and the operation will
not continue. If you have unassigned drives, use the Virtual Disk >> Add
Free Capacity option to add additional capacity to the virtual disk. Then,
retry the operation.
● Copyback
● Reconstruction
● Initialization
● Changing Segment Size
● Defragmentation of a virtual disk
● Adding Free Capacity to a virtual disk
● Changing RAID Level of a virtual disk
The Lowest priority rate favors system performance, but the modification
operation will take longer. The Highest priority rate favors the
modification operation, but system performance may be compromised.
A default segment size is set during volume creation, based on the virtual
disk RAID level and the volume I/O characteristics specified in the
Storage Profile in use. These two parameters should optimize the segment
size appropriately for your environment. When you create the volume,
you are also given the option of selecting a custom segment size rather
than accepting the default segment size.
Monitor your storage system and change segment size when necessary for
optimal performance based on the guidelines below.
If your typical I/O size is larger than your segment size, increase your
segment size in order to minimize the number of drives needed to satisfy
an I/O request. This technique helps even more if you have random I/O
access patterns. Using a single drive for a single request leaves other
drives available to simultaneously service other requests.
If you are using the volume in a single-user, large I/O environment such
as multimedia application storage, performance is optimized when a
single I/O request can be serviced with a single system data stripe (the
segment size multiplied by the number of drives in the virtual disk used
for I/ O). In this case, multiple disks are used for the same request, but
each disk is only accessed once.
Remember
● You cannot cancel this operation once it begins.
● Do not begin this operation unless the virtual disk is Optimal.
● The controller firmware determines the segment size transitions that
are allowed. Segment sizes that are inappropriate transitions from
the current segment size are unavailable on the menu. Allowed
transitions typically are double or half of current segment size. For
example, if the current volume segment size is 32K, a new volume
segment size of either 16K or 64K is allowed.
How long a Change Segment Size operation can take depends on many
variables, including:
● the I/O load from the hosts
● the modification priority of the volume
● the number of drives in the virtual disk
● the number of drive channels
● the processing power of the storage system controllers.
If you want this operation to complete faster, you can change the
modification priority, although this may decrease system I/O
performance. To change the priority, select a volume in the virtual disk,
then select Volume >> Change>>Modification Priority.
Before DCE
After DCE
You can add 1 or 2 drives that were either previously unassigned or newly
inserted into an existing drive tray at a time.
However, in the case of Snapshot reserve volumes since they are not
mapped to hosts, expansion is supported for all host environments.
Example:
● Step 1 - Look at space directly below volume 1 - no capacity
● Step 2 - Look at space directly above volume 1 - no capacity
● Steps 3 & 4 - Continue to move outward - capacity is found above
volume 0
● Must position capacity below Volume 1
● Volumes 0 and 1 must move up 1 GB
● Steps 5 - 1.5 GB still required
● Must postion capacity below volume 1
● Volumes 2 and 3 must move down 1.5 GB
Knowledge Check
3. What are the cache parameters that can be set for each volume?
True or False
1. Increasing the segment size will always improve performance.
True False
True False
3. The higher the modification priority is set, the faster the I/O’ are
serviced, and the modification operations complete at slower pace.
True False
True False
True False
6. The Dynamic Functions (DSS; DRM, DCE and DVE) will terminate if
the storage system is powered off.
True False
Multiple Choice
1. What is performance?
a) How well a storage system stores or retrieves data for
various host workloads.
b) The probability that a disk sub-system is available 7 x 24.
c) The maximum ratio of read operations to write operations
that a storage system can execute.
d) The number of requests that can be fulfilled simultaneously
to retrieve data.
Objectives
Upon completion of this module, you should be able to:
● Identify the available data services for the Sun StorageTek 6540 array
● List the benefits and application of Snapshot
● Explain how Snapshot is implemented
These features are ideal for data protection surrounding backup, business
continuance and disaster recovery situations. All three features require a
license and can be enabled or disabled as you choose.
Snapshot
Snapshot creates a “static” or point-in-time (PiT) image of a volume. The
Snapshot volume is created almost instantaneously and appears and
functions as a volume as shown in Figure 9-1.
Snapshot Terminology
To better understand snapshot, there are several terms which must be
defined. This list includes:
● Base volume
● Snapshot volume
● Reserve volume
● Original data blocks
Base Volume
Definition: The base volume defines volume from which the Snapshot is
taken.
Snapshot Volume
Additionally:
● Snapshots can be Disabled - stopped
● Snapshots can be Recreated at a later time
● A maximum of 4 Snapshots per Base volume can exist.
● The maximum number of Snapshots per storage system is one half
the total number of volumes supported by the controller model.
● The Snapshot is virtual, but actually consists of the Base and Reserve
Volumes.
Note – Due to the dependency on the Base volume, a Snapshot should not
be utilized for data migration or disaster recovery purposes in protecting
against catastrophic failure to the original volume or storage system
Reserve Volume
Definition: The original data blocks are data blocks that are on the base
volume at the time the Snapshot was taken.
The Snapshot Volume is comprised of Original data blocks that are still on
the Base Volume and Original data blocks that have been copied to the
Reserve Volume.
Note – Due to the dependency on the Base volume, a Snapshot should not
be utilized for data migration or disaster recovery purposes in protecting
against catastrophic failure to the original volume or storage system.
At 11 am, the host issues a read request for block A to the storage system.
The data block resides on the Base volume, and the read data comes
directly from there. Figure 9-2 shows a standard read with no snapshot.
Snapshot is Created
At 11:05, a Snapshot is created. When the Snapshot is “taken,” the
controller suspends I/O to the Base volume for a few seconds while it
creates a physical Reserve volume to store the Snapshot metadata and
copy-on-write data. The logical Snapshot volume is also created and is
immediately available for mapping. See Figure 9-3.
Notice that the Snapshot volume is identical to the Base volume at the
time the Snapshot is created. In this example with data blocks A, B, C, D,
E, F, G and H. No matter how much the Base volume changes after 11:05,
the Snapshot volume will look the same as the Base volume did at 11:05.
Snapshot will always reflect the Original data.
At 11:15, the host issues a read for data block A from the Snapshot
volume. As mentioned previously, no physical data resides in the
Snapshot volume. The Reserve volume combined with the original Base
volume creates the logical Snapshot volume. So, when data is requested
from the Snapshot volume, the disk system determines if the data is in the
Base volume or the Reserve volume. In this case, data block A resides in
the Base volume, so the read comes directly from there.
Write to Base
Now at 11:30, the host issues a write to the Base volume. Blocks Z,Y are
going to overwrite blocks A,B. As blocks A,B are needed for the Snapshot
volume, a copy-on-write occurs, copying blocks A,B into the Reserve
volume for safekeeping. Once this is done, the write of blocks Z,Y
completes to the Base volume and is acknowledged to the host. Figure 9-5
shows the Snapshot writing to base.
Re-Write to Base
New Re-write changes do not need to do anything to the Snapshot,
because the original data is already in the Reserve as shown in Figure 9-6.
At 11:45, the host issues another write to the Base volume. This time block
X is overwriting block Z. Snapshot is more correctly described as copy on
FIRST write technology. No additional copy on write operation is needed
because the original data block – block A – was moved to the Reserve
when the first write to this block took place, therefore, subsequent writes
to the block do not require any action.
In this example, block X simply overwrites block Z, and the I/O write is
acknowledged to the host.
At 12:00, the host issues a read for blocks A,B,C,D from the Snapshot
volume. When we read A at 11:15 (see Figure 9-7), block A was still in its
original location in the Base volume. Since then, however, it has been
overwritten. So now when the host issues a read for blocks A,B,C,D to the
Snapshot volume, the storage system uses the metadata map. A,B are in
the Reserve volume and the data is read directly from there. C,D are still
on the Base volume and have not been overwritten since the Snapshot
was taken, so the Snapshot (metadata map) simply points to the original
blocks still on the Base Volume, and blocks C,D are read directly from the
Base Volume.
Write to Snapshot
If a write is performed to the Snapshot, the original data is overwritten in
the Reserve, and the result is the Snapshot is now no longer a PiT of the
original data.Writes to the Snapshot are stored in the Reserve, as the
Snapshot is not a physical volume and therefore cannot store data.
In Figure 9-8, the host is overwriting data block D in the Snapshot volume
with block M. As the Snapshot volume is not a physical volume, block M
has to go somewhere. Writing to the Snapshot volume puts the data
directly into the Reserve.
Snapshot Considerations
There are several things to consider when you are creating a snapshot,
including:
● Performance
● Volume failover considerations
● Handling deleting base or snapshot respository volumes
● Maintaining the state on your volumes
Performance
Read operations from the Snapshot might be slower than reads from the
Base volume, because the metadata map in the Reserve has to be
consulted first.
Volume Failover
Ownership changes affect Base and ALL of its Snapshots. The Base
volume, Snapshot and Reserve are all owned by the same controller. The
rules that apply to the Base volume for AVT and RDAC modes, also apply
to the associated Snapshots and Repositories. All “related” volumes
change controller ownership as a group.
Base, Snapshot and Reserve volumes are all associated. Each Snapshot
requires it’s own Reserve. A Snapshot cannot exist without a Base or a
Reserve. When you delete a Base volume, all Snapshots of this volume
and associated Repositories will also be deleted. When you delete a
Snapshot, the associated Reserve will also be deleted.
When using cloning operations like Snapshot and Volume Copy, the most
important part is to ensure that the source volume is in a consistent state.
By nature, a clone is an identical copy of it’s source volume. If the source
volume is not consistent, the clone is also not consistent. An inconsistent
volume might be unusable for the purpose needed.
Open applications like databases or even mounted file systems keep files
open. Flags on the physical disks usually indicate the opened status. If we
now take a Snapshot, the Snapshot will also indicate the opened state. If
this had been a file system or database, the application would ask for a
database check or file system check, which easily takes a couple of hours.
Hosts use buffers - reserved space in the memory of the host - which act as
a kind of cache space. The data used most, such as directory structures, or
bitmap tables, is quite often kept in buffers to improve overall disk
performance. Snapshot and Volume Copy can only copy what’s physically
on disk, but not what’s stored in the hosts memory.
Snapshot OS support
Since the Snapshot is identical to the Base volume, OS’s may not support
the mapping of volumes with identical signatures/block information to
the same data host.
Use Snapshot on
Use Snapshot on
Host Environment and DIFFERENT
SAME system as
Volume Type system
Base Volume
as Base Volume
Use Snapshot on
Use Snapshot on
Host Environment and DIFFERENT
SAME system as
Volume Type system
Base Volume
as Base Volume
Managing Snapshots
The following section covers things to consider when creating and
managing a snapshot using CAM.
Creating a Snapshot
Prior to creating a snapshot with the Common Array Manager it is
important to plan the following aspects:
● The name of the snapshot reserve volume - When a snapshot is
created, it must be assigned a unique name. This simplifies
identification of the primary volume.
Each snapshot has an associated reserve volume that stores
information about the data that has changed since the snapshot was
created. It too must have a unique name making it easy to identify as
the reserve volume of the snapshot to which it corresponds.
● The capacity of the reserve volume - To determine the appropriate
capacity, calculate both the management overhead required and
percentage of change expected on the base volume.
● The warning threshold - When a snapshot volume is created, the
threshold at which the management software will generate messages
to indicate the level of space left in the reserve volume can be
specified. By default, the software generates a warning notification
when data in the reserve volume reaches 50 percent of the available
capacity. The percentage of space used can be monitored on the
Snapshot Details page for the snapshot.
● The method used to handle snapshot failures - When a snapshot
volume is created, you can determine how the management software
will respond when the reserve volume for the snapshot becomes full.
The management software can do either of the following:
● Fail the snapshot volume. In this case the snapshot becomes
invalid, but the base volume continues to operate normally.
● Fail the base volume. In this case, attempts to write new data to
the primary volume fail. This leaves the snapshot as a valid
copy of the original base volume.
● The virtual disk selection method - A snapshot can be created on a
virtual disk as long as the virtual disk has enough capacity for the
snapshot.
The following options are available:
When creating a snapshot, the size of the snapshot reserve volume that
will store snapshot data and any other data that is needed during the life
of the snapshot must be specified. When prompted to specify the size of
the snapshot reserve volume, the size is entered as a percentage of the size
of the base volume, as long as that percentage does not translate to a size
of less than 8 megabytes.
The capacity needed for the snapshot reserve volume varies, depending
on the frequency and size of I/O writes to the base volume and how long
the snapshot volume will be kept. In general, choose a large capacity for
the reserve volume if you intend to keep the snapshot volume for a long
period of time or if you anticipate heavy I/O activity, which will cause a
large percentage of data blocks to change on the base volume during the
life of the snapshot volume. Use historical performance to monitor data or
other operating system utilities to help you determine typical I/O activity
on the base volume.
If you receive a warning and determine that the snapshot reserve volume
is in danger of filling up before you have finished using the snapshot
volume, increase its capacity by navigating to the Snapshot Details page
and clicking Expand. If the snapshot reserve volume fills up before you
have finished using the snapshot, the snapshot failure handling
conditions specify the action that will be taken.
Creating a Snapshot
Figure 9-11 illustrates an overall view of creating the snapshot.
Objectives
Upon completion of this module, you should be able to:
● Explain how Volume Copy is implemented
● List the benefits and application of Volume Copy
● Explain the functions that can be performed on a Copy Pair
Source Volume
Definition: The source volume is the volume that accepts host I/O and
stores application data. When a volume copy is started, data from the
source volume is copied in its entirety to the target volume.
In order to maintain the data integrity of the point in time target, volume
copy suspends write to the source volume during the copy process.
Therefore, in order to maintain normal I/O activity and ensure data
availability, volume Copy must be used in conjunction with Snapshot -
where the Snapshot volume is the source volume for the volume Copy.
Target volume
Caution – A volume copy will overwrite all data on the target volume
and automatically make the target volume read-only to hosts. Ensure that
you no longer need the data or have backed up the data on the target
volume before starting a volume copy.
While the copy is in progress, the target volume is not available for any
I/O from a host. When the copy is complete the target volume by default
will be read-only, but can be modified by the user to be read and write
accessible.
The target volume must be the same or greater capacity as the source
volume, but can be of a different RAID level.
If you choose the base volume of an active Snapshot volume as the target
volume, you must disable all Snapshot volumes associated with the base
volume before creating a volume copy.
Copy Pair
The source volume and it’s associated target volume for a single volume
copy are know as a copy pair. The copy pair relationship links the source
and target volumes together. A copy pair can be:
● Stopped - stop the copy, but copy pair relationship is maintained
● Re-copied - re-copy the source to the target, thereby overwriting the
previous data on the target
● Removed - sever the copy pair relationship, leaving the data on the
source and target intact.
Note – A maximum of eight copy pairs can have a status of “In Progress”
at one time. If more than eight volume copies are created, they will each
have a status of “Pending” until one of the volume copies with a status of
“In Progress” completes. For example, if a ninth copy pair is defined, it
will be placed in a queue until one of the existing eight copy processes
completes, at which time the ninth copy process will begin.
Figure 10-2 Migrate Data to Larger Drives and Change RAID Level
During a volume copy, the same controller must own both the source and
target Volumes. If both Volumes do not have the same preferred controller
when the volume copy starts, the ownership of the target Volume is
automatically transferred to the preferred controller of the source Volume.
When you create a new volume copy, you will define the copy priority to
determine how much controller processing time is allocated for the
volume copy process and diverted from I/O activity.
There are five relative priority settings. The Highest priority rate supports
the volume copy at the expense of I/O activity. The Lowest priority rate
supports I/O activity at the expense of volume copy speed.
Once the volume copy starts an operation from the source volume over to
the target volume, no read or write requests to the target volume are
allowed. The volume copy goes from an idle to an active state, displaying
either an In Progress or Pending (resources not available) status. These
status conditions are displayed in the Jobs window of CAM.
State Description
State Description
State Description
State
During volume copy After volume copy
I/O
Source Target Source Target Target
write- write-
protect protect
disabled enabled
on a source or target volume after a volume copy has been created, the
modification operation must complete before the volume copy can start. If
a volume copy has a status of In Progress, modification operations will not
be allowed.
After the volume copy has completed, the target Volume automatically
becomes read-only to hosts, and write requests to the target Volume will
not be permitted.
The Volume Details page of either the source or the target volumes can be
used to re-copy a copy pair, stop a copy pair with a status of In Progress,
remove copy pairs (which removes the copy pair association information
from the storage system but leaves the data in tact on both the source and
target Volumes), change the volume copy priority, and disable the target
Volume’s Read-Only attribute.
Recopying a Volume
The Re-Copy option enables you to create a new Volume copy for a
previously defined copy pair that may have been Stopped, Failed, or has
Completed. This option can be used for creating scheduled, complete
backups of the target Volume that can then be copied to tape drive for off-
site storage.
After starting the Re-Copy option, the data on the source Volume is
copied in its entirety to the target Volume. Volume Copy does not support
the ability to resynchronize the target with only the changes that occurred
to the source after the copy was completed. The copy process is a full,
block by block replication at a given point in time. It is not mirroring
technology, which continuously updates the target.
You can also set the copy priority for the volume copy at this time. The
higher priorities will allocate storage system resources to the volume copy
at the expense of the storage system’s performance.
Re-Copy Considerations
● This option will overwrite existing data on the target Volume and
make the target Volume read-only to hosts. This option will fail all
snapshot Volumes associated with the target Volume, if any exist.
● Only one copy pair at a time can be selected to be re-copied.
● Similar to Snapshot re-create
● New full-size point-in-time copy using same source and target
● The Re-copy option is always available EXCEPT when there is
already a copy pending, a copy is already in progress (option
available when Copy Failed), target is a degraded Volume, source or
target Volume is also a secondary Enhanced Remote Mirror Volume,
offline Volume, failed Volume or missing Volume.
After the volume copy has been stopped, the Re-Copy option can be used
to create a new volume copy using the original copy pair.
Note – When the volume copy is stopped, all mapped hosts will have
write access to the source Volume, If data is written to the source Volume,
the data on the target Volume will no longer match the data on the source
Volume.
After the volume copy is removed, the target Volume can be selected as a
source Volume or target Volume for a new volume copy. Removing a
volume copy also permanently removes the Read-Only attribute for the
target Volume.
Note – If the volume copy has a status of In Progress, you must stop the
volume copy before you can remove the copy pair.
There are five relative settings ranging from Lowest to Highest. The
Highest priority rate supports the volume copy, but I/O activity may be
affected. The Lowest priority rate supports I/O activity, but the volume
copy will take longer.
Volume Permissions
Read and write requests to the target Volume will be rejected while the
volume copy has a status of In Progress, Pending, or Failed. After the
volume copy has completed, the target Volume automatically becomes
read-only to hosts. You may want to keep the Read-Only attribute enabled
in order to preserve the data on the target Volume.
Examples of when you may want to keep the Read-Only attribute enabled
include:
● If you are using the target Volume for backup purposes
● If you are copying data from one array to a larger array for greater
accessibility
● If you are planning to use the data on the target Volume to copy back
to the base Volume in case of a disabled or failed Snapshot Volume
If you decide to allow host write access to the data on the target Volume
after the volume copy is completed, use the Volume details page in CAM
to disable the Read-Only attribute for the target Volume.
Note – Some OS’s may report an error when accessing a read only device,
in which case the read only access on the Target volume must be disabled
in order to allow read and write access. Unix OS’s may allow access to a
read only device as long as it is mounted as a read only device.
Storage Partitioning
If you want hosts to have read and write access to the data on the target
volume, use the Volume Details page in CAM to disable the read-only
attribute for the target volume.
Snapshot
In order to maintain the data integrity of the point in time clone, volume
Copy suspends writes to the source during the copy process. If the
volume being copied is large, this can result in an extended period of time
without the ability to make updates or changes. Even though the source
volume does support read-only access, many operating systems still try to
write to the volume when it is in a read-only mode. If this happens, the
server can hang.
Figure 10-3 shows that copying the Snapshot creates a full PiT clone copy
while I/O continues to the Base volume.
Copying the Snapshot creates the same full point in time clone of the
desired source volume, and it does so while I/O continues to the
production volume. The process is straightforward. First a Snapshot of a
volume is created. Then volume Copy uses the Snapshot volume as its
source volume. Once the copy is complete, the Snapshot volume can be
deleted.
The volume for which the Snapshot is created is known as the base
volume and must be a standard volume in the storage system. For the
volume copy feature, the base volume of a Snapshot volume is permitted
to be selected as the source volume for a volume copy.
Note – If you choose the base volume of a Snapshot volume as your target
volume, you must disable all Snapshot volumes associated with the base
volume before you can select it as a target volume. Otherwise, the base
volume cannot be used as a target volume.
Any volume copies that are In-Progress will fail and cannot be restarted
until the primary volume is demoted back to its original secondary
volume role.
Use Target on
Use Target on
Host Environment and DIFFERENT
SAME system as
Volume Type system
Source Volume
as Source Volume
Use Target on
Use Target on
Host Environment and DIFFERENT
SAME system as
Volume Type system
Source Volume
as Source Volume
After the volume copy has completed, the target volume automatically
becomes read-only to hosts, and write requests to the target volume will
not be permitted.
Caution – Remember, a volume copy will overwrite all data on the target
volume and automatically make the target volume read-only to hosts.
After the volume copy process has finished, you can enable hosts to write
to the target volume by changing the target volume’s Read-Only attribute
on the Volume Details page.
Note – Because a target volume can have only one source volume, it can
participate in one copy pair as a target. However, a target volume can also
be a source volume for another volume copy, enabling you to make a
volume copy of a volume copy.
● Set the copy priority for the volume copy. During a volume copy, the
storage array’s resources may be diverted from processing I/O
activity to completing a volume copy, which may affect the storage
array’s overall performance.
Note – If you disable the volume copy feature, but volume copy pairs still
exist, you can still remove the copy pair, start a copy using the existing
copy pair, and change the setting of the read-only attribute for target
volumes. However, you cannot create new volume copies.
Before creating a volume copy, be sure that a suitable target volume exists
on the storage array, or create a new target volume specifically for the
volume copy. You can create a copy of a standard volume, a target
volume, or a snapshot volume.
5. Select the target volume you want from the Target Volumes list.
Select a target volume with a capacity similar to the usable capacity
of the source volume to reduce the risk of having unusable space on
the target volume after the volume copy is created.
6. Before starting the volume copy process:
a. Stop all I/O activity to the source and target volumes.
b. Unmount any file systems on the source and target volumes, if
applicable.
7. Review the specified information on the Copy Volume page. If you
are satisfied, click OK to start the volume copy.
8. A message confirms that the volume copy has successfully started.
After the volume copy process has finished:
a. Remount any file systems on the source volume and target
volume, if applicable.
b. Enable I/O activity to the source volume and target volume.
Caution – Recopying a volume copy will overwrite all data on the target
volume and automatically make the target volume read-only to hosts.
Ensure that you no longer need the data or have backed up the data on
the target volume before recopying a volume copy.
Re-Copying a Volume
After starting the Re-Copy option, the data on the source volume is
copied in its entirety to the target volume. Volume copy does not support
the ability to re-synchronize the target with only the changes that
occurred to the source after the copy was completed. The copy process is
a full, block by block replication at a given point in time. It is not
mirroring technology, which continuously updates the target.
You can also set the copy priority for the volume copy at this time. The
higher priorities will allocate storage system resources to the volume copy
at the expense of the storage system’s performance.
The Stop Copy option allows you to stop a volume copy that has a status
of In Progress, Pending, or Failed. Using this option on a volume copy
with a status of Failed clears the Critical Alarm status displayed for the
storage system in the Current Alarms of the storage management
software.
After the volume copy has been stopped, the Re-Copy option can be used
to create a new volume copy using the original copy pair.
Note – When the volume copy is stopped, all mapped hosts will have
write access to the source volume. If data is written to the source volume,
the data on the target volume will no longer match the data on the source
volume.
The Remove Copy Pairs option allows you to remove one or more volume
copies pairs. Any volume copy-related information for the source volume
and target volume is removed from the volume Properties and Storage
Array Profile dialogs.
After the volume copy is removed, the target volume can be selected as a
source volume or target volume for a new volume copy. Removing a
volume copy also permanently removes the Read-Only attribute for the
target volume.
Note – If the volume copy has a status of “In Progress”, you must stop the
volume copy before you can remove the copy pair.
The data on the source volume or target volume is not deleted when you
remove a copy pair.
The Change Copy Priority dialog allows you to set the rate at which the
volume copy completes.
Volume Permissions
Read and write requests to the target volume will be rejected while the
volume copy has a status of In Progress, Pending, or Failed. After the
volume copy has completed, the target volume automatically becomes
read-only to hosts. You may want to keep the Read-Only attribute enabled
in order to preserve the data on the target volume.Examples of when you
may want to keep the Read-Only attribute enabled include:
● If you are using the target volume for backup purposes
● If you are copying data from one virtual disk to a larger virtual disk
for greater accessibility
● If you are planning to use the data on the target volume to copy back
to the base volume in case of a disabled or failed Snapshot volume
If you decide to allow host write access to the data on the target volume
after the volume copy is completed, use the Volume Details page in CAM
to disable the Read-Only attribute for the target volume.
Objectives
Upon completion of this module, you should be able to:
● Explain how Replication is implemented in the storage manager
● Describe the benefits and applications of Replication
● Differentiate between Synchronous and Asynchronous mirroring
modes
When a disaster occurs at one site, the secondary or backup site takes over
responsibility for computer services. Therefore, users and hosts that were
previously mapped to a primary storage system can have access to a
secondary storage system. Essentially this is a good BCDR (Business
Continuity and Disaster Recovery) plan where the ability to have a robust
business continuance strategy keeps essential services operational during
and after a failure or a disaster.
Note – The terms local and remote are relative. Support for cross mirrors
between storage systems means a given system can be considered both
local and remote with both primary and secondary volumes.
Mirror
Repositories
Mirror
Repositories
Mirror Pairs or
Replication Set
V1 -> V1M
V2 -> V2M
Primary Secondary V3 -> V3M Secondary Primary
Primary Volume
The volume residing in the primary or local storage system is the primary
volume. The primary volume accepts host I/O and stores application
data. The data on a primary volume is replicated to the secondary
volume.
When a mirror pair is first created, data from the primary volume is
copied in its entirety to the secondary volume. This process is known as
full synchronization and is directed by the controller that owns the
primary volume. During a full synchronization, the primary volume
remains fully accessible for all read and write host I/O.
Secondary volume
The secondary volume has to be of equal or greater size than the primary.
RAID level and drive type do not matter. The secondary volume can also
be the base volume for a Snapshot.
Two mirror repositories are required per storage system, one for each
controller. The mirror reserve volume is 128 Mbyte (256 Mbyte total per
storage system). Unlike Snapshot reserves, mirror reserves are not
required for each mirrored pair as actual read/write data is not stored in
the mirror reserve.
The delta log and the FIFO log are kept in the mirror reserve.
The delta log is used to track changes to the primary volume that have
not yet been replicated to the secondary volume. Therefore, if an
interruption occurs to the communication between the two storage
systems, the delta log can be used to re-synchronize the data between the
secondary and primary volumes. The delta log is a bit map (maximum 1
million bits per mirror), where each bit represents a section of the primary
volume that was written by the host, but has not yet been copied to the
secondary volume. The number of blocks represented by a single bit is
computed based on the usable capacity of the primary volume. The
minimum amount of data represented by a single bit is 64K, that is 128-
512-byte blocks. For example, for a 2TB volume, each bit will represent a
data range of 2 Mbyte.
The FIFO log is used during Write Consistency mirroring mode to ensure
writes are completed in the same order on both the primary and
secondary volumes.
Replication Set
Synchronous Mirroring
A write I/O from a host must be written to both the primary and
secondary volumes before the I/O is reported as complete.
When the controller owner of the primary volume receives a write request
from a host, the controller first logs the information about the write to the
mirror reserve, then writes the data to the primary volume. The controller
then initiates a remote write operation to copy the affected data blocks
from the primary to the secondary volume.
After the host write request has been written to the primary volume and
the data has been successfully copied to the secondary volume, the
controller removes the log entry from the mirror reserve and sends an I/O
completion status back to the data host.
Asynchronous Mirroring
Host write requests are written to just the primary volume before the
controller sends an I/O completion status back to the host system
regardless of when the data was successfully copied to the secondary
storage system.The asynchronous write mode offers faster I/O
performance but does not guarantee that the copy has been successfully
completed before processing the next write request.
In Asychronous mirroring, the primary storage system does not wait for
the I/O to complete to the secondary storage system before sending an
I/O completion status to the server. Therefore, there can be multiple
outstanding I/O’s to the secondary storage system. Remote Replication
supports up to 128 outstanding I/O’s per mirror pair. After the 128th I/O
has been issued to the secondary volume, the primary volume will
suspend any new I/O’s until one of the outstanding I/O’s to the
secondary volume has completed and freed up space in the queue for
pending I/O’s.
The write consistency option is available for any primary and secondary
volumes participating in an asychronous remote replication relationship.
When the write consistency option is selected for multiple volumes on the
same storage system, the order in which data is synchronized is
preserved. Selecting the write consistency option for a single mirror pair
does not make sense because the process in which data is replicated does
not change. More than one mirror pair must have the write consistency
option selected for the replication process to change.
When multiple replication pairs exist on the same storage systems and
have been configured for Asychronous Mirroring with Write Consistency,
they are considered to be an interdependent group known as a write
consistency group. All mirror pairs in the write consistency group
maintain the same order when sending writes from the primary volume
to their corresponding secondary volume. The data on the secondary
volume cannot be considered fully synchronized until all mirror pairs in
the write consistency group are synchronized. If one mirror pair in a write
consistency group becomes unsynchronized, all of the mirrored pairs in
the write consistency group will become unsynchronized. Any write
activity to the remote site will be prevented to protect the write
consistency of the remote data set.
Role Reversal
A role reversal is the act of promoting the secondary volume to be the
primary volume within the mirrored volume pair, and or demoting the
primary volume to be the secondary volume. The role reversal process
always requires a user to initiate the process.
When the secondary volume becomes a primary volume, any hosts that
are mapped to the volume through a volume-to-LUN mapping will now
be able to read or write to the volume.
Host
Mirrored Pair
Unsynchronized State
The primary controller will also mark the mirror pair as unsynchronized
when a volume error on the secondary prevents the remote write from
completing. For example, an offline or a failed secondary volume can
cause the mirror pair to become unsynchronized. When the volume error
is corrected (the secondary volume is placed online, or recovered to an
optimal status), a synchronization (automatic or manual) is required, and
the mirror pair transitions to synchronization in progress.
Figure 11-10 on page 11-280 shows how the arrays with their associated
hosts are cabled using three switches.
Note – The same goal can be obtained using a single 16-port switch
configured with three zones.
Note – The array dedicates Fibre Channel (FC) port 4 on each controller
for use with the Sun StorageTek Data Replicator software premium
feature. Before enabling data replication on an array, you must ensure that
FC port 4 on each controller is not currently in use. If it is in use, you must
move all connections from FC port 4 to FC port 1.
Note – You must delete all replication sets before you can deactivate the
premium feature.
Before any mirror relationships can be created, volumes must exist at both
the primary and secondary sites. If a primary volume does not exist, one
will need to be created on the primary storage system. If a secondary
volume does not exist, one will need to be created on the secondary
storage system. Consider the following when creating the secondary
volume:
● The secondary volume must be of equal or greater size than the
associated primary volume.
● The RAID level and drive type of the secondary volume do not
have to be the same as the primary volume.
● Stop all I/O activity and unmount any file systems on the secondary
volume. Do this just before creating the replication set.
● Log in to the system using the storage user role.
The Create Replication Set wizard enables you to create a replication set,
either standalone or as part of the consistency group.
5. Click Replicate.
6. The Create Replication Set wizard is displayed. Follow the steps in
the wizard. The Create Replication Set wizard also allows you to
include the new replication set in the consistency group, if desired.
When creating the replication set, the system copies all data from the
primary volume to the secondary volume, overwriting any existing
data on the secondary volume. If replication is suspended, either
manually or due to a system or communication problem, and then
resumed, only the differences in data between volumes are copied.
When creating a Replication Set you have the option to select the
Synchronization Priority Level. You can choose from five different
synchronization priorities for the primary volume, ranging from
lowest to highest, that determine how much of a priority the full
synchronization will receive relative to host I/O activity and,
therefore, how much of a performance impact there will be. The
following guidelines roughly approximate the differences between
the five priorities. Note that volume size can cause these estimates to
vary widely.
Reversing Roles
If communication with the remote storage system is down, you can force
a role reversal even when there will be a resulting dual-primary or dual-
secondary condition. Use the Recovery Guru to recover from one of these
conditions after communication is restored with the remote system.
2. The Array Summary page is displayed. Click the name of the array
containing the volume in the replication set whose role you want to
reverse.
3. The Volume Summary page is displayed. Click the Replication Sets
tab.
4. The Replication Set Summary page is displayed. Click the name of
the replication set that includes the volume.
5. The Replication Set Details page is displayed. Click Role to
Secondary or Role to Primary, as appropriate.
6. A confirmation message is displayed. Click OK.
Note – Any data that is written to the primary volume will be tracked
while the replication set is suspended and will automatically be written to
the secondary volume when replication is resumed. A full
synchronization will not be required.
Note – For backup routines, use the Suspend option rather than
Removing the mirror relationship.
Ethernet Network
Site 1 Site 2
Switch 1A Switch 2A
Fabric
10 Km Max
Port Port
Controller
Controller
A1 A1
A
Port Port
Storage A2 A2 Storage
Array 1 Array 2
Port Port
Controller
Controller
B1 B1
B
Port Port
B2 B2
Switch 1B Switch 2B
Fabric
10 Km Max
Sonoran-2
Host FC Cable 4 Switches
Storage Array FC Cable Dual Site
Dedicated RVM Feature FC Cable
S2Config1.vsd
Ethernet Network
Site 1 Site 2
Switch 1 Switch 2
Fabric
10 Km Max
Port Port
Controller
Controller
A1 A1
A
Port Port
A2 A2
Port Port
Controller
Controller
B1 B1
B
B
Port Port
B2 B2
Storage Storage
Array 1 Array 2 Low-Cost
Enhanced Remote
Host FC Cable
Mirroring Campus
Storage Array FC Cable Configuration
Dedicated Enhanced Remote Mirroring Feature FC Cable
E th e r n e t N e tw o rk
S w itc h 1 S w it c h 2
P o rt P o rt
C o n t r o lle r
C o n t r o lle r
A1 A1
A
A
P o rt P o rt
A2 A2
C o n t r o lle r
P o rt C o n t r o lle r P o rt
B1 B1
B
B
P o rt P o rt
B2 B2
S to r a g e S to ra g e
A rra y 1 A rra y 2
H o s t F C C a b le
S to ra g e A rra y F C C a b le
D e d i c a t e d R V M F e a t u r e F C C a b le
S 2 C o n fig 3 .v s d
True False
2. Volume Copy volumes MUST have the same RAID level and
configuration.
True False
True False
True False
True False
True False
10. What happens if a data block on the base volume is changed more
than once after the snapshot is taken?
13. What is the maximum number of copy pairs that can be in progress
at one time?
16. What are the two logs kept in the mirror repository volume? Briefly
describe what each does.
18. What happens if there is a link interruption during the remote mirror
process?
Problem Determination
Objectives
Upon completion of this module, you should be able to:
● Utilize the tools in CAM to analyze information about a storage
system issue
● Use service advisor to determine how to solve problems
● Use the SSCS to import and export the configuration
Problem Determination
Ask yourself these questions about the Storage System Environment:
● What has or is changing?
● What does the physical look like (LEDs, cabling)?
● What is the configuration?
● Other Indicators?
● What other questions should you ask?
● What tools are available to aid in observation?
Compatibility Matrix
A table of all 3rd party hardware and software
components that should be used with a particular
level of controller firmware.
Service Advisor
● Collection of “service”
procedures
● Manually locate the one you
need ... OR ...
● Arrive via an Alarm
● Controllers
● Batteries
● Interconnect Canister
● Disk Drives
● I/O Module (IOM)
● Battery
● Power Supply
● SAS Interface Cable
X-Options
● Offline/Online Controllers
● Reset Controller
● Correcting an IOM Firmware Mismatch
● Redistribute Volumes
● Setting the Drive Channel to Optimal
● Reviving a Disk Drive
● Recovering from an Overheated Power Supply
Service Only
Support Data
Alarms
Current Alarms
● Black - Down
● Red - Major
● Yellow - Minor
● Blue - Note
Summary List
Detailed Listing
● Probable Cause
● Recommended Action
● View Aggregated Events
● Link to Service Advisor
With pictures
Controller Diagnostics
Reset
Test communications
Summary Pages
Component Summary
Events
Array Administration
Administration
● Manage Passwords
● Redistribute Volumes
● Reset Configuration
● Upgrade Firmware
● Change Array Name
● Define Default Host Type
● Define Start/Stop Cache %
● Configure Background Disk Scrubbing
● Configure Alert Fail-Over Delay
● Set Time Manually
● Synchronize Time with Time Server
● Array Health Monitoring
● Enable Health monitoring
● Configure Performance Monitoring
● Enable/Disable
● Set Polling interval
● Set Data Retention Perior.
● Add licenses
● Disable licenses
● View Activity Log
● View System Specific Alarms
Sun Connection
Health Administration
Notification
SNMP Traps
Activity Log
Alarms
Jobs
Events
Activity log
Performance monitoring
● I/O statistics
Features
● Remote CLI shared across product line
● Multi-Platform Support
● Full Feature Set
● Scriptable
● Backward compatible – Continued support for 6120, 6320, 6920
● Man pages (UNIX)
Benefits
● All processing performed on the server.
● New features installed on the server immediately available to all
clients.
● Client upgrade not necessary with server upgrade
● Performance independent of client machine
● Code sharing with GUI
Usage
Login
Example:
Built in help
help keyword
Note – The sscs command has an inactivity timer. The session terminates
if you do not issue any sscs commands for 30 minutes. You must log in
again after the timeout to issue a command.
> /temp/configfile.xml
Profile
Virtual disk
Disks
$FMS_HOME/bin/ras_admin
● The “backend” for many of the Browser User Interface (BUI) and CLI
functions
● Useful for diagnosing why an array can't be registered through the
BUI or CLI
● Offers in-band and device list method for discovering arrays
● Can also perform other functions
● List and delete alerts
● List and delete devices
● Add, list and delete email addresses for notifications
● List and display reports
● List and display topologies
$SLM_HOME/bin/csmservice
With CAM 5.1 will support stand-alone mode, the displaying of current
firmware versions and individual array component updates.
Example
$ /opt/SUNEstksm/bin/csmservice -h
CSMServices 5.1.0.10
Where:
Output Name
Save Location
● Unix
● /tmp
● Windows
● %SYSTEMDRIVE%\tmp
Methods of collecting