Troubleshooting, Recovery, and Maintenance Guide: IBM Storwize V7000
Troubleshooting, Recovery, and Maintenance Guide: IBM Storwize V7000
Version 6.3.0
GC27-2291-02
Note
Before using this information and the product it supports, read the general information in “Notices” on page 143, the
information in the “Safety and environmental notices” on page ix, as well as the information in the IBM Environmental
Notices and User Guide on the documentation DVD.
This edition applies to the IBM Storwize V7000, Version 6.3.0, and to all subsequent releases and modifications until
otherwise indicated in new editions.
| This edition replaces GC27-2291-01.
© Copyright IBM Corporation 2010, 2011.
US Government Users Restricted Rights – Use, duplication or disclosure restricted by GSA ADP Schedule Contract
with IBM Corp.
Contents
Figures . . . . . . . . . . . . . . . v | Chapter 4. Understanding the medium
| errors and bad blocks . . . . . . . . 27
Tables . . . . . . . . . . . . . . . vii
Chapter 5. Storwize V7000 user
Safety and environmental notices . . . ix interfaces for servicing your system . . 29
Sound pressure . . . . . . . . . . . . . ix Management GUI interface . . . . . . . . . 29
When to use the management GUI . . . . . 30
About this guide . . . . . . . . . . . xi Accessing the management GUI . . . . . . 30
Who should use this guide . . . . . . . . . xi Service assistant interface . . . . . . . . . . 31
| Summary of changes for GC27-2291-02 Storwize When to use the service assistant . . . . . . 31
| V7000 Troubleshooting, Recovery, and Maintenance Accessing the service assistant . . . . . . . 32
| Guide . . . . . . . . . . . . . . . . xi Cluster (system) command-line interface. . . . . 33
Summary of changes for GC27-2291-01 Storwize When to use the cluster (system) CLI . . . . . 33
V7000 Troubleshooting, Recovery, and Maintenance Accessing the cluster (system) CLI. . . . . . 33
Guide . . . . . . . . . . . . . . . . xi Service command-line interface . . . . . . . . 33
Emphasis . . . . . . . . . . . . . . . xii When to use the service CLI . . . . . . . . 33
Storwize V7000 library and related publications . . xiii Accessing the service CLI. . . . . . . . . 33
How to order IBM publications. . . . . . . . xv USB key and Initialization tool interface . . . . . 33
Sending your comments . . . . . . . . . . xvi When to use the USB key. . . . . . . . . 34
Using a USB key . . . . . . . . . . . 34
Chapter 1. Storwize V7000 hardware Using the initialization tool . . . . . . . . 34
satask.txt commands . . . . . . . . . . 35
components . . . . . . . . . . . . . 1
Components in the front of the enclosure . . . . . 2
Drives . . . . . . . . . . . . . . . 2
Chapter 6. Resolving a problem . . . . 39
Drive indicators . . . . . . . . . . . . 2 Start here: Use the management GUI recommended
Enclosure end cap indicators . . . . . . . . 4 actions . . . . . . . . . . . . . . . . 39
Components in the rear of the enclosure . . . . . 5 Problem: Storage system management IP address
Power supply unit and battery for the control unknown . . . . . . . . . . . . . . . 40
enclosure . . . . . . . . . . . . . . 6 Problem: Unable to connect to the management GUI 40
Power supply unit for the expansion enclosure . . 7 Problem: Unable to log on to the storage system
Node canister ports and indicators . . . . . . 8 management GUI . . . . . . . . . . . . 41
Expansion canister ports and indicators . . . . 15 Problem: Cannot create a clustered storage system 41
Problem: Unknown service address of a node
canister . . . . . . . . . . . . . . . . 42
Chapter 2. Best practices for Problem: Cannot connect to the service assistant . . 43
troubleshooting . . . . . . . . . . . 19 Problem: Management GUI or service assistant does
Record access information . . . . . . . . . 19 not display correctly . . . . . . . . . . . 44
Follow power management procedures . . . . . 20 Problem: Node canister location error . . . . . 44
Set up event notifications . . . . . . . . . . 20 Problem: SAS cabling not valid . . . . . . . . 44
Set up inventory reporting . . . . . . . . . 20 Problem: New expansion enclosure not detected . . 45
Back up your data . . . . . . . . . . . . 21 Problem: Control enclosure not detected . . . . . 45
Manage your spare and failed drives . . . . . . 21 Problem: Mirrored volume copies no longer
Resolve alerts in a timely manner . . . . . . . 21 identical . . . . . . . . . . . . . . . 46
Keep your software up to date . . . . . . . . 21 Problem: Code not processed from USB key . . . 46
Keep your records up to date . . . . . . . . 22 Procedure: Resetting superuser password . . . . 47
Subscribe to support notifications . . . . . . . 22 Procedure: Identifying which enclosure or canister
Know your IBM warranty and maintenance to service . . . . . . . . . . . . . . . 47
agreement details . . . . . . . . . . . . 22 Procedure: Checking the status of your system . . 48
Procedure: Getting node canister and system
Chapter 3. Understanding the Storwize information using the service assistant . . . . . 48
V7000 battery operation for the control Procedure: Getting node canister and system
enclosure . . . . . . . . . . . . . 23 information using a USB key . . . . . . . . 49
Procedure: Understanding the system status using
Maintenance discharge cycles . . . . . . . . 24
the LEDs . . . . . . . . . . . . . . . 49
2. Locate IBM Storwize V7000 Safety Notices with the user publications that were
provided with the Storwize V7000 hardware.
3. Find the matching identification number in the IBM Storwize V7000 Safety
Notices. Then review the topics concerning the safety notices to ensure that you
are in compliance.
4. Optionally, read the multilingual safety instructions on the Storwize V7000
website. Go to the Support for Storwize V7000 website at www.ibm.com/
storage/support/storwize/v7000 and click the documentation link.
Sound pressure
Attention: Depending on local conditions, the sound pressure can exceed 85
dB(A) during service operations. In such cases, wear appropriate hearing
protection.
The chapters that follow introduce you to the hardware components and to the
tools that assist you in troubleshooting and servicing the Storwize V7000, such as
the management GUI and the service assistant.
The troubleshooting procedures can help you analyze failures that occur in a
Storwize V7000 system. With these procedures, you can isolate the components
that fail.
You are also provided with step-by-step procedures to remove and replace parts.
| New information
| This topic describes the changes to this guide since the previous edition,
| GC27-2291-01. The following sections summarize the changes that have since been
| implemented from the previous version.
| Changed information
| This version includes updated navigation paths for the management GUI.
This topic describes the changes to this guide since the previous edition,
GC27-2291-00. The following sections summarize the changes that have since been
implemented from the previous version.
Changed information
Emphasis
Different typefaces are used in this guide to show emphasis.
The IBM Storwize V7000 Information Center contains all of the information that is
required to install, configure, and manage the Storwize V7000. The information
center is updated between Storwize V7000 product releases to provide the most
current documentation. The information center is available at the following
website:
publib.boulder.ibm.com/infocenter/storwize/ic/index.jsp
Unless otherwise noted, the publications in the Storwize V7000 library are
available in Adobe portable document format (PDF) from the following website:
Table 3 lists IBM publications that contain information related to the Storwize
V7000.
Table 3. Other IBM publications
Title Description Order number
IBM Storage Management This guide describes how to GC27-3909
Pack for Microsoft System install, configure, and use the
Center Operations Manager IBM Storage Management Pack publibfp.dhe.ibm.com/
User Guide for Microsoft System Center epubs/pdf/c2739092.pdf
Operations Manager (SCOM).
IBM Storage Management This publication provides GA32-0929
Console for VMware vCenter, installation, configuration, and
version 2.6.0, Installation usage instructions for the IBM publibfp.dhe.ibm.com/
Guide Storage Management Console for epubs/pdf/a3209295.pdf
VMware vCenter.
Table 4 lists websites that provide publications and other information about the
Storwize V7000 or related products or technologies.
Table 4. IBM documentation and related websites
Website Address
Support for Storwize V7000 (2076) Support for Storwize V7000 website at
www.ibm.com/storage/support/storwize/v7000
Support for IBM System Storage® www.ibm.com/storage/support/
and IBM TotalStorage products
IBM Publications Center www.ibm.com/e-business/linkweb/publications/
servlet/pbi.wss
IBM Redbooks® publications www.redbooks.ibm.com/
To view a PDF file, you need Adobe Acrobat Reader, which can be downloaded
from the Adobe website:
www.adobe.com/support/downloads/main.html
www.ibm.com/e-business/linkweb/publications/servlet/pbi.wss
To submit any comments about this book or any other Storwize V7000
documentation:
v Go to the feedback page on the website for the Storwize V7000 Information
Center at publib.boulder.ibm.com/infocenter/storwize/ic/index.jsp?topic=/
com.ibm.storwize v7000.doc/feedback.htm. There you can use the feedback page
to enter and submit comments or browse to the topic and use the feedback link
in the running footer of that page to identify the topic for which you have a
comment.
v Send your comments by email to [email protected]. Include the following
information for this publication or use suitable replacements for the publication
title and form number for the publication on which you are commenting:
– Publication title: IBM Storwize V7000 Troubleshooting, Recovery, and Maintenance
Guide
– Publication form number: GC27-2291-02
– Page, table, or illustration numbers that you are commenting on
– A detailed description of any information that should be changed
There are several model types. The main differences among the model types are
the following items:
v The number of drives that an enclosure can hold. Drives are located on the front
of the enclosure. An enclosure can hold up to 12 3.5-inch drives or up to 24
2.5-inch drives.
v Whether the model is a control enclosure or an expansion enclosure.
Control enclosures contain the main processing units that control the whole
system. They are where external systems, such as host application servers, other
storage systems, and management workstations are connected through the
Ethernet ports or Fibre Channel ports. Control enclosures can also be connected
to expansion enclosures through the serial-attached SCSI (SAS) ports.
Expansion enclosures contain additional storage capacity. Expansion enclosures
connect either to control enclosures or to other expansion enclosures through the
SAS ports.
v If the control enclosure has either 1 Gbps Ethernet capability or 10 Gbps
Ethernet capability.
The machine type and model (MTM) are shown on these labels that are located on
the front and the rear of each enclosure:
v The left end cap on the front of the enclosure. The label also indicates if the
enclosure is a control enclosure or an expansion enclosure.
v The rear of the left enclosure flange.
Note: The labels also show the enclosure serial number. You must know the serial
number when you contact IBM support.
Because of the differences between the enclosures, you must be able to distinguish
between the control enclosures and the expansion enclosures when you service the
system. Be aware of the following differences:
v The model type that is shown on the labels.
v The model description that is shown on the left end cap.
Drives
An enclosure can hold up to 12 3.5-inch drives or up to 24 2.5-inch drives.
The drives are located in the front of the enclosure. The 12 drives are mounted in
four columns with three rows.
Note: The drive slots cannot be empty. A drive assembly or blank carrier must be
in each slot.
svc00609
Figure 1. 12 drives on either 2076-112 or 2076-312
svc00608
Drive indicators
The drives have two LED indicators each. They have no controls or connectors.
The LED color is the same for both drives. The LEDs for the 3.5-inch drives are
placed vertically above and below each other. The LEDs for the 2.5-inch drives are
placed next to each other at the bottom.
svc00607
1
svc00606
1 2
Figure 5 shows where the end caps are located on the front of an enclosure with 12
drives. The end caps are located in the same position for an enclosure with 24
drives.
v 1 Left end cap
v 2 Drives
v 3 Right end cap
Figure 6 shows the indicators on the front of the enclosure end cap.
The left enclosure end caps for both enclosures are identical and contain only
indicators. The left enclosure end cap contains no controls or connectors. The right
enclosure end cap for both enclosures has no controls, indicators, or connectors.
1 2 3
svc00694
1
2
3
4
svc00605
Two canisters are located in the middle of each enclosure. The power supply units
are located on the left and right of the canisters. The left slot is power supply 1
(1), and the right slot is power supply 2 (2). Power supply 1 is top side up,
and power supply 2 is inverted. The upper slot is canister 1 (3), and the lower
slot is canister 2 (4). Canister 1 is top side up, and canister 2 is inverted.
Figure 7 shows the rear view of a model 2076-112 or a model 2076-124 control
enclosure. Figure 8 on page 6 shows the rear view of a model 2076-312 or a model
2076-324 control enclosure with the 10 Gbps Ethernet port (5). Figure 9 on page 6
shows the rear of an expansion enclosure.
1 3 2
svc00662
svc00726
4
1 3 2
svc00610
4
The two power supply units in the enclosure are installed with one unit top side
up and the other inverted. The power supply unit for the control enclosure has six
LEDs.
There is a power switch on each of the power supply units. The switch must be on
for the power supply unit to be operational. If the power switches are turned off,
or the main power is removed, the integrated batteries temporarily continue to
supply power to the node canisters. As a result, the canisters can store
configuration data and cached data to their internal drives. Battery power is
required only if both power supply units stop operating.
Figure 10 on page 7 shows the location of the LEDs 1 in the rear of the power
supply unit.
svc00670
Figure 10. LEDs on the power supply units of the control enclosure
See “Procedure: Understanding the system status using the LEDs” on page 49 for
help in diagnosing a particular failure.
The two power supply units in the enclosure are installed with one unit top side
up and the other inverted. The power supply unit for the expansion enclosure has
four LEDs, two less than the power supply for the control enclosure.
Figure 11 shows the locations of the LEDs 1 in the rear of the power supply unit.
svc00671
Figure 11. LEDs on the power supply units of the expansion enclosure
See “Procedure: Understanding the system status using the LEDs” on page 49 for
help in diagnosing a particular failure.
Note: The reference to the left and right locations applies to canister 1, which is
the upper canister. The port locations are inverted for canister 2, which is the lower
canister.
svc00689
Figure 12. Fibre Channel ports on the node canisters
There are two green LEDs associated with each port: the speed LED and the link
activity LED. These LEDs are in the shape of a triangle. The LEDs are located in
between the two rows of the ports as shown in Figure 13. Figure 13 shows the
LEDs for the Fibre Channel ports on canister 1. Each LED points to the associated
port. The first and second LEDs in each set show the speed state, and the third
and fourth LEDs show the link state.
1 2 3 4 5 6 7 8
1 2
3 4
svc00685
Table 10 provides the status descriptions for the LEDs on the Fibre Channel ports.
Table 10. Fibre Channel port LED status descriptions
Speed state LED Link state LED Link state
Off Off Inactive
Off On or flashing Active low speed (2 Gbps)
Flashing On or flashing Active medium speed (4
Gbps)
On On or flashing Active high speed (8 Gbps)
Fibre Channel ports are identified by their physical port number and by a
worldwide port name (WWPN).
The physical port numbers identify Fibre Channel cards and cable connections
when you perform service tasks. The WWPNs are used for tasks such as Fibre
Channel switch configuration and to uniquely identify the devices on the SAN.
The WWPNs are derived from the worldwide node name (WWNN) that is
allocated to the Storwize V7000 node in which the ports are installed. The WWNN
for each node is stored within the enclosure. When you replace a node canister, the
WWPNs of the ports do not change.
USB ports
Two USB ports are located side by side on each node canister.
Note: The reference to the left and right locations applies to canister 1, which is
the upper canister. The port locations are inverted for canister 2, which is the lower
canister.
svc00690
For the 1 Gbps support, the Ethernet ports are numbered 1 on the left and 2 on the
right as shown in Figure 15 on page 12. Port 1 must be connected; the use of port 2
is optional. Two LEDs are associated with each port.
Note: The reference to the left and right locations applies to canister 1, which is
the upper canister. The port locations are inverted for canister 2, which is the lower
canister.
Figure 15. Ethernet ports on the 2076-112 and 2076-124 node canisters
Figure 16. 10 Gbps Ethernet ports on the 2076-312 and 2076-324 node canisters
The SAS ports are numbered 1 on the left and 2 on the right as shown in Figure 17.
Port 1 is used if you add one expansion enclosure. Port 2 is used if you add a
second expansion enclosure. Each port provides four data channels.
Note: The reference to the left and right locations applies to canister 1, which is
the upper canister. The port locations are inverted for canister 2, which is the lower
canister.
svc00692
SAS ports must be connected to Storwize V7000 enclosures only. See “Problem:
SAS cabling not valid” on page 44 for help in attaching the SAS cables.
Four LEDs are located with each port. Each LED describes the status of one data
channel within the port. The data channel number is shown with the LED.
Table 13. SAS port LEDs on the node canister
LED state Description
Off No link is connected.
Flashing The link is connected and has activity.
On The link is connected.
The three LEDs are located in a horizontal row near the upper right of the canister
1. Figure 18 shows the rear view of the node canister LEDs.
svc00672
Figure 18. LEDs on the node canisters
Note: The reference to the left and right locations applies to canister 1, which is
the upper canister. The port locations are inverted for canister 2, which is the lower
canister.
Table 14. Node canister LEDs
Name Description Color Symbol
System Indicates the status of the node. Green
status v The on status indicates that the node is active, that
is, it is an active member of a clustered system.
When the node is active, do not remove it.
v The off state indicates there is no power to the
canister or the canister is in standby mode. These
conditions can cause the off state:
– The main processor is off and only the service
processor is active.
– A power-on self-test (POST) is running on the
canister.
– The operating system is loading.
v The flashing status indicates that the node is in
candidate state or service state. It is not able to
perform I/O in a system. When the node is in
either of these states, it can be removed. Do not
remove the canister unless directed by a service
procedure.
There is a diagnostic port on the left of the canister. There are no indicators that
are associated with the port. There are no defined procedures that use the port.
Note: The reference to the left and right locations applies to canister 1, which is
the upper canister. The port locations are inverted for canister 2, which is the lower
canister.
Note: The reference to the left and right locations applies to canister 1, which is
the upper canister. The port locations are inverted for canister 2, which is the lower
canister.
1 2
2 1
svc00668
Figure 19. SAS ports and LEDs in rear of expansion enclosure
Four LEDs are located with each port. Each LED describes the status of one data
channel within the port. The data channel is shown with the LED.
Table 15. SAS port LEDs on the expansion canister
LED state Description
Off No link is connected.
Flashing The link is connected and has activity.
On The link is connected.
The two LEDs are located in a vertical row on the left side of the canister.
Figure 20 on page 17 shows the LEDs (1) in the rear of the expansion canister.
svc00673
Figure 20. LEDs on the expansion canisters
Record the following information and ensure that authorized people know how to
access the information:
v The management IP addresses. This address connects to the system using the
management GUI or starts a session that runs the command-line interface (CLI)
commands. The system has two Ethernet ports. Each port can have either an
IPv4 address or an IPv6 address or both. Record this address and any limitations
regarding where it can be accessed from within your Ethernet network.
v The service IP address of the node canisters on the control enclosures is used
only in certain circumstances. The service IP address connects to a node canister
in the control enclosure. Access to the address is sometimes required if the
canister has a fault that stops it from becoming an active member of the system.
Each of the two node canisters can have a service IP address that is specified for
Ethernet port 1. Each address can have either an IPv4 address or an IPv6
address or both. Ensure that the address specified for each node canister is
different.
v The system password for user superuser. The password is required to access the
system through the service IP address. The authentication of superuser is always
local; therefore, the user ID can be used when a remote authentication server
that is used for other users is not available.
Table 17. Access information for your system
Item Value Notes
The management IP address
for the GUI and CLI
The management user ID
(the default is admin)
The management user ID
password (the default is
admin)
The control enclosure
management IP address
Control enclosure service IP
address: node canister 1
Control enclosure service IP
address: node canister 2
The control enclosure
superuser password (the
default is passw0rd)
Use the management GUI or the CLI commands to power off a system. Using
either of these methods ensures that the data that is cached in the node canister
memory is correctly flushed to the RAID arrays.
Do not power off an enclosure unless instructed to do so. If you power off an
expansion enclosure, you cannot read or write to the drives in that enclosure or to
any other expansion enclosure that is attached to it from the SAS ports. Powering
off an expansion enclosure can prevent the control enclosure from flushing all the
data that it has cached to the RAID arrays.
If you set up Call Home to IBM, ensure that the contact details that you configure
are correct and kept up to date as personnel change.
The storage system backs up your control enclosure configuration data to a file
every day. This data is replicated on each control node canister in the system.
Download this file regularly to your management workstation to protect the data.
This file must be used if there is a serious failure that requires you to restore your
system configuration. It is important to back up this file after modifying your
system configuration.
The spare drives are used automatically if a member drive fails. If you have
sufficient spare drives, you do not have to replace them immediately when they
fail. However, monitoring the number, size, and technology of your spare drives,
ensures that you have sufficient drives for your requirements. Ensure that there are
sufficient spare drives available so that your RAID arrays are always online.
If there are a number of unfixed alerts, fixing any one alert might become more
difficult because of the effects of the other alerts.
Check the IBM support website to see if new code releases are available:
If you have only one system, it is relatively easy to identify the enclosures that
make up the system. Identification becomes more difficult when you have multiple
systems in your data center and multiple systems in the same rack.
The enclosure identifier that is displayed on the front of the display is unique
within a system. However, the identifiers can be repeated between different
systems. Do not rely solely on this identifier.
For each system, record the location of the control enclosure and the location of
any expansion enclosures. It is useful to label the enclosures themselves with the
system name and the management IP addresses.
Subscribe to support notifications by visiting the IBM support page on the IBM
website:
By subscribing, you are informed of new and updated support site information,
such as publications, hints and tips, technical notes, product flashes (alerts), and
downloads.
Have the phone number of the support center available. When you call support,
provide the machine type (always 2076) and the serial number of the enclosure
that has the problem. If the problem does not relate to a specific enclosure, provide
the control enclosure serial number. The serial numbers are on the labels on the
enclosures.
Support personnel also ask for your customer number, machine location, contact
details, and the details of the problem.
If the power fails, the cache and state data is written to a local solid-state drive
(SSD) that is held within the canister. The batteries within the control enclosure
provide the power to write the cache and state data to a local drive.
Note: Storwize V7000 expansion canisters do not cache volume data or store state
information in volatile memory. They, therefore, do not require battery power. If ac
power to both power supplies in an expansion enclosure fails, the enclosure
powers off. When ac power is restored to at least one of the power supplies, the
controller restarts without operator intervention.
There are two power supply units in the control enclosure. Each one contains an
integrated battery. Both power supply units and batteries provide power to both
control canisters. Each battery has a sufficient charge to power both node canisters
for the duration of saving critical data to the local drive. In a fully redundant
system with two batteries and two canisters, there is enough charge in the batteries
to support saving critical data from both canisters to a local drive twice. In a
system with a failed battery, there is enough charge in the remaining battery to
support saving critical data from both canisters to a local drive once.
If the ac power to a control enclosure is lost, the canisters do not start saving
critical data to a local drive until approximately 10 seconds after the loss of ac
power is first detected. If the power is restored within this period, the system
continues to operate. This loss in power is called a brown out. As soon as the
saving of the critical data starts, the system stops handling I/O requests from the
host applications, and Metro Mirror and Global Mirror relationships go offline. The
system powers off when the saving of the critical data completes.
If both node canisters shut down without writing the cache and state data to the
local drive, the system is unable to restart without an extended service action. The
system configuration must be restored. If any cache write data is lost, volumes
must be restored from a backup. It is, therefore, important not to remove the
canisters or the power supply units from the control enclosures unless directed to
do so by the service procedures. Removing either of these components might
prevent the node canister from writing its cache and state data to the local drive.
When the ac power is restored to the control enclosure, the system restarts without
operator intervention. How quickly it restarts depends on whether there is a
history of previous power failures.
When the ac power is restored after a power outage that causes both canisters to
save their critical data, the system restarts only when the batteries have sufficient
charge to power both canisters for the duration of saving the critical data again. In
a fully redundant system with two batteries, this condition means that after one ac
power outage and a saving of critical data, the system can restart as soon as the
power is restored. If a second ac power outage occurs before the batteries have
In a system with a failed battery, an ac power failure causes both canisters to save
critical data and completely discharges the remaining battery. When the ac power
is restored, the system starts in service state and does not permit I/O operations to
be restarted until the remaining battery is fully charged. The recharging takes
approximately 1 hour.
There are conditions other than loss of ac power that can cause critical data to be
saved and the nodes to go into service state and not permit I/O operations. The
node canisters save critical data if they detect there is no longer sufficient battery
charge to support a saving of critical data. This situation happens when, for
example, both batteries have two-thirds of a charge. The total charge that is
available in the enclosure is sufficient to support a saving of critical data once;
therefore, both canisters are in active state and I/O operations are permitted. If one
battery fails though, the remaining battery has only two-thirds of a charge, and the
total charge that is available in the enclosure is now insufficient to perform a
saving of the critical data if the ac power fails. Data protection cannot be
guaranteed in this case. The nodes save the critical data by using the ac power and
enter service state. The nodes do not handle I/O operations until the remaining
battery has sufficient charge to support the saving of the critical data. When the
battery has sufficient charge, the system automatically restarts.
If one of the batteries in a system fails and is not replaced, it prevents the other
battery from performing a maintenance discharge. Not only does this condition
reduce the lifetime of the remaining battery, but it also prevents a maintenance
discharge cycle from occurring after the battery has provided protection for at least
2 critical saves or 10 brown outs. Preventing this maintenance cycle from occurring
increases the risk that the system accumulates a sufficient number of power
outages to cause the remaining battery to be discounted when calculating whether
there is sufficient charge to protect the system. This condition results in the system
entering service state while the one remaining battery performs a maintenance
discharge. I/O operations are not permitted during this process. This activity takes
approximately 10 hours.
Chapter 3. Understanding the Storwize V7000 battery operation for the control enclosure 25
26 Storwize V7000: Troubleshooting, Recovery, and Maintenance Guide
|
| The volume virtualization that is provided extends the time when a medium error
| is returned to a host. Because of this difference to non-virtualized systems, the
| Storwize V7000 uses the term bad blocks rather than medium errors.
| The Storwize V7000 allocates volumes from the extents that are on the managed
| disks (MDisks). The MDisk can be a volume on an external storage controller or a
| RAID array that is created from internal drives. In either case, depending on the
| RAID level used, there is normally protection against a read error on a single
| drive. However, it is still possible to ge a medium error on a read request if
| multiple drives have errors or if the drives are rebuilding or are offline due to
| other issues.
| The Storwize V7000 provides migration facilities to move a volume from one
| underlying set of physical storage to another or to replicate a volume that uses
| FlashCopy or Metro Mirror or Global Mirror. In all these cases, the migrated
| volume or the replicated volume returns a medium error to the host when the
| logical block address on the original volume is read. The system maintains tables
| of bad blocks to record where the logical block addresses that cannot be read are.
| These tables are associated with the MDisks that are providing storage for the
| volumes.
| It is possible that the tables that are used to record bad block locations can fill up.
| The table can fill either on an MDisk or on the system as a whole. If a table does
| fill up, the migration or replication that was creating the bad block fails because it
| was not possible to create an exact image of the source volume.
| The system creates alerts in the event log for the following situations:
| v When it detects medium errors and creates a bad block
| v When the bad block tables fill up
| Bad blocks are cleared by deallocating the volume disk extent by deleting the
| volume or by issuing write I/O to the block. It is good practice to correct bad
| blocks as soon as they are detected. This action prevents the bad block from being
| propagated when the volume is replicated or migrated. It is possible, however, for
| the bad block to be on part of the volume that is not used by the application. For
| example, it can be in part of a database that has not been initialized. These bad
| blocks are corrected when the application writes data to these areas. Before the
| correction happens, the bad block records continue to use up the available bad
| block space.
Use the initialization tool to do the initial setup of your system. Use the
management GUI to monitor and maintain the configuration of storage that is
associated with your clustered systems. Perform service procedures from the
service assistant. Use the command-line interface (CLI) to manage your system.
| You use the management GUI to manage and service your system. The Monitoring
| > Events panel provides access to problems that must be fixed and maintenance
| procedures that step you through the process of correcting the problem.
| You can also sort events by time or error code. When you sort by error code, the
| most serious events, those with the lowest numbers, are displayed first. You can
| select any event that is listed and select Actions > Properties to view details about
| the event.
Regularly monitor the status of the system using the management GUI. If you
suspect a problem, use the management GUI first to diagnose and resolve the
problem.
| Use the views that are available in the management GUI to verify the status of the
| system, the hardware devices, the physical storage, and the available volumes. The
| Monitoring > Events panel provides access to all problems that exist on the
| system. Use the Recommended Actions filter to display the most important events
| that need to be resolved.
If there is a service error code for the alert, you can run a fix procedure that assists
you in resolving the problem. These fix procedures analyze the system and provide
more information about the problem. They suggest actions to take and step you
through the actions that automatically manage the system where necessary. Finally,
they check that the problem is resolved.
If there is an error that is reported, always use the fix procedures within the
management GUI to resolve the problem. Always use the fix procedures for both
software configuration problems and hardware failures. The fix procedures analyze
the system to ensure that the required changes do not cause volumes to be
inaccessible to the hosts. The fix procedures automatically perform configuration
changes that are required to return the system to its optimum state.
You must use a supported web browser. Verify that you are using a supported web
browser from the following website:
You can use the management GUI to manage your system as soon as you have
created a clustered system.
1. Start a supported web browser and point the browser to the management IP
address of your system.
The management IP address is set when the clustered system is created. Up to
four addresses can be configured for your use. There are two addresses for
IPv4 access and two addresses for IPv6 access.
2. When the connection is successful, you see a login panel.
3. Log on by using your user name and password.
After all the alerts are fixed, check the status of your system to ensure that it is
operating as intended.
You connect to the service assistant on one node canister through the service IP
address. If there is a working communications path between the node canisters,
you can view status information and perform service tasks on the other node
canister by making the other node canister the current node. You do not have to
reconnect to the other node.
The node canister might be in service state because it has a hardware issue, has
corrupted data, or has lost its configuration data.
The storage system management GUI operates only when there is an online
system. Use the service assistant if you are unable to create a system or if both
node canisters in a control enclosure are in service state.
The service assistant does not provide any facilities to help you service expansion
enclosures. Always service the expansion enclosures by using the management
GUI.
The service assistant provides detailed status and error summaries. You can also
perform the following service-related actions:
v Collect logs to create and download a package of files to send to support
personnel.
v Remove the data for the system from a node.
A number of tasks that are performed by the service assistant cause the node
canister to restart. It is not possible to maintain the service assistant connection to
the node canister when it restarts. If the current node canister on which the tasks
are performed is also the node canister that the browser is connected to and you
lose your connection, reconnect and log on to the service assistant again after
running the tasks.
You must use a supported web browser. Verify that you are using a supported and
an appropriately configured web browser from the following website:
Perform the service assistant actions on the correct node canister. If you did not
connect to the node canister that you wanted to work on, access the Change Node
panel from the home page to select a different current node.
Commands are run on the current node. The current node might not be the node
canister that you connected to. The current node identification is shown on the left
at the top of the service assistant screen. The identification includes the enclosure
serial number, the slot location, and if it has one, the node name of the current
node.
For a full description of the commands and how to start an SSH command-line
session, see the “Command-line interface” topic in the “Reference” section of the
Storwize V7000 Information Center.
Nearly all of the flexibility that is offered by the CLI is available through the
management GUI. However, the CLI does not provide the fix procedures that are
available in the management GUI. Therefore, use the fix procedures in the
management GUI to resolve the problems. Use the CLI when you require a
configuration setting that is unavailable in the management GUI.
You might also find it useful to create command scripts using the CLI commands
to monitor for certain conditions or to automate configuration changes that you
make on a regular basis.
For a full description of the commands and how to start an SSH command-line
session, see the “Command-line interface” topic in the “Reference” section of the
Storwize V7000 Information Center.
To access a node canister directly, it is normally easier to use the service assistant
with its graphical interface and extensive help facilities.
When a USB key is inserted into one of the USB ports on a node canister in a
control enclosure, the node canister searches for a control file on the USB key and
runs the command that is specified in the file. When the command completes, the
command results and node status information are written to the USB key.
When a USB key is plugged into a node canister, the node canister software
searches for a text file named satask.txt in the root directory. If the software finds
the file, it attempts to run a command that is specified in the file. When the
command completes, a file called satask_result.html is written to the root
directory of the USB key. If this file does not exist, it is created. If it exists, the data
is inserted at the start of the file. The file contains the details and results of the
command that was run and the status and the configuration information from the
node canister. The status and configuration information matches the detail that is
shown on the service assistant home page panels.
The satask.txt file can be created on any workstation by using a text editor. If a
Microsoft Windows workstation is being used, the initialization tool can be used to
create the commands that are most often used.
The fault LED on the node canister flashes when the USB service action is being
performed. When the fault LED stops flashing, it is safe to remove the USB key.
The USB key can then be plugged into a workstation and the satask_result.html
file viewed in a web browser.
To protect from accidentally running the same command again, the satask.txt file
is deleted after it has been read.
If no satask.txt file is found on the USB key, the result file is still created, if
necessary, and the status and configuration data is written to it.
If you download the initialization tool, you must copy the file onto the USB key
that you are going to use.
To start the initialization tool, insert the USB key that contains the program into a
USB slot on a suitable personal computer. Run the InitTool.exe program from the
USB drive.
The initialization tool is used to create the satask.txt file on a USB key. After the
satask.txt file is created, follow the instructions in “Using a USB key” on page 34
to run the commands on the node.
The initialization tool prompts you for the task that you want to perform and for
the parameters that are relevant to that task. It prompts you when to put it in the
node canister on the control enclosure. When the commands have run, return the
USB key to your personal computer and start the tool again to see the results.
By using the initialization tool, you can set the USB key to run one of the
following tasks:
v Create a new system.
v Reset the superuser password.
| v Set or reset the service IP address on the node canister on the control enclosure.
| For any other tasks that you want to perform on a node canister on the control
| enclosure, you must create the satask.txt file using a text editor.
satask.txt commands
This topic identifies the commands that can be run from a USB key.
If you are creating the satask.txt command file by using a text editor, the file
must contain a single command on a single line in the file. The commands that
you use are the same as the service CLI commands except where noted. Not all
service CLI commands can be run from the USB key. The satask.txt commands
always run on the node that the USB key is plugged into.
Syntax
Parameters
-serviceip
(Required) The IPv4 address for the service assistant.
-gw
(Optional) The IPv4 gateway for the service assistant.
-mask
(Optional) The IPv4 subnet for the service assistant.
-serviceip_6
(Required) The IPv6 address for the service assistant.
-gw_6
(Optional) The IPv6 gateway for the service assistant.
-prefix_6
(Optional) Resets to the default IPv4 address.
-default
(Required) The IPv6 prefix for the service assistant.
-resetpassword
(Optional) Sets the service assistant password to the default value.
Description
| This command resets the service assistant IP address to the default value. If the
| command is run on the upper canister, the default value is 192.168.70.121 subnet
| mask: 255.255.255.0. If the command is run on the lower canister, the default value
| is 192.168.70.122 subnet mask: 255.255.255.0. If the node canister is active in a
system, the superuser password for the system is reset; otherwise, the superuser
password is reset on the node canister.
If the node canister becomes active in a system, the superuser password is reset to
that of the system. You can configure the system to disable resetting the superuser
password. If you disable that function, this action fails.
This action calls the satask chserviceip command and the satask resetpassword
command.
Syntax
satask resetpassword
Parameters
None.
This command resets the service assistant password to the default value passw0rd.
If the node canister is active in a system, the superuser password for the system is
reset; otherwise, the superuser password is reset on the node canister.
If the node canister becomes active in a system, the superuser password is reset to
that of the system. You can configure the system to disable resetting the superuser
password. If you disable that function, this action fails.
Snap command
Use this command to collect diagnostic information from the node canister and to
write the output to a USB key.
Syntax
Parameters
-options
(Optional) Specifies which diagnostic information to collect.
Description
Syntax
Parameters
-file
(Required) The file name of software installation package.
-ignore
(Optional) Overrides prerequisite checking and forces installation of the
software.
Description
This command copies the file from the USB key to the upgrade directory on the
node canister.
Syntax
Parameters
-clusterip
(Required) The IPv4 address for Ethernet port 1 on the system.
-gw
(Required) The IPv4 gateway for Ethernet port 1 on the system.
-mask
(Required) The IPv4 subnet for Ethernet port 1 on the system.
-clusterip_6
(Required) The IPv6 address for Ethernet port 1 on the system.
gw_6
(Required) The IPv6 gateway for Ethernet port 1 on the system.
prefix_6
(Required) The IPv6 prefix for Ethernet port 1 on the system.
name
(Optional) The name of the new system.
Description
Syntax
sainfo getstatus
Parameters
None.
Description
This command writes the output from each node canister to the USB key.
The following procedures are often used to find and resolve problems:
v Procedures that involve data collection and system configuration
v Procedures that are used for hardware replacement.
Always use the recommended actions of the management GUI as the starting point
to diagnose and resolve a problem. The topics that follow describe the type of
problem that you might experience that are not resolved by using the management
GUI. In those situations, review the symptoms and follow the actions that are
provided here.
Note: After you have created your clustered system, remove hardware components
only when directed to do so by the fix procedures. Failure to follow the procedures
can result in loss of access to data or loss of data. Follow the fix procedures when
servicing a control enclosure.
You can connect to and manage a Storwize V7000 system as soon as you have
created a clustered system. If you cannot create a clustered system, see the problem
that contains information about what to do if you cannot create one. Go to
“Problem: Cannot create a clustered storage system” on page 41.
To run the management GUI, start a supported web browser and point it to the
management IP address of your system. Up to four addresses can be configured
for your use. There are two addresses for IPv4 access, and two addresses for IPv6
access. If you do not know the system management IP address, go to “Problem:
Storage system management IP address unknown” on page 40. After the
connection is successful, you see a login panel. If you are unable to access the login
panel, go to “Problem: Unable to connect to the management GUI” on page 40.
Log on using your user name and password. If you are unable to log on, go to
“Problem: Unable to log on to the storage system management GUI” on page 41.
Select the recommended alert, or any other alert, and run the fix procedure. The fix
procedure steps you through the process of troubleshooting and correcting the
problem. The fix procedure displays information that is relevant to the problem
and provides various options to correct the problem. Where it is possible, the fix
procedure runs the commands that are required to reconfigure the system.
Always use the recommended action for an alert because these actions ensure that
all required steps are taken. Use the recommended actions even in cases where the
service action seems obvious, such as a drive showing a fault. In this case, the
drive must be replaced and reconfiguration must be performed. The fix procedure
performs the reconfiguration for you.
The fix procedure also checks that another existing problem does not result in a fix
procedure that causes volume data to be lost. For example, if a power supply unit
in a node enclosure must be replaced, the fix procedure checks and warns you if
the integrated battery in the other power supply unit is not sufficiently charged to
protect the system.
If possible, fix the alerts in the order shown to resolve the most serious issues first.
Often, other alerts are fixed automatically because they were the result of a more
serious issue.
After all the alerts are fixed, go to “Procedure: Checking the status of your system”
on page 48.
If you do not know the storage system management IP address, it is part of the
data that is shown in the service assistant home panel or the data that is returned
by the USB key. If you know the service address of a node canister, go to
“Procedure: Getting node canister and system information using the service
assistant” on page 48; otherwise, go to “Procedure: Getting node canister and
system information using a USB key” on page 49.
Log on using your user name and password. Follow the suggested actions when
you encounter a specific situation:
v If you are not logging on as superuser, contact your system administrator who
can verify your user name and reset your account password.
v If the user name that you are using is authenticated through a remote
authentication server, verify that the server is available. If the authentication
server is unavailable, you can log on as user name superuser. This user is
always authenticated locally.
v If you do not know the password for superuser, go to “Procedure: Resetting
superuser password” on page 47.
| The create clustered-system function protects the system from loss of volume data.
| If you create a clustered system on a control enclosure that was previously used,
| you lose all of the volumes that you previously had. To determine if there is an
| existing system, use data that is returned by “Procedure: Getting node canister and
| system information using the service assistant” on page 48 or “Procedure: Getting
| node canister and system information using a USB key” on page 49.
v The node canister that you are attempting to create a clustered system on is in
candidate state. The node canister is in candidate state if it is a new canister.
v The partner node canister in the control enclosure is not in active state.
v The latest system ID of the control enclosure is 0.
| If the create function failed because there is an existing system, fix the existing
| clustered system; do not re-create a new clustered system.If you want to create a
| clustered system and do not want to use any data from the volumes used in the
| previous clustered system, go to “Procedure: Deleting a system completely” on
| page 56, and then run the create function again.
| You might not be able to create a cluster if the node canister (the one on which
| you are attempting to create the clustered system) is in service state. Check
| whether the node canister is in service state by using the data returned by
| “Procedure: Getting node canister and system information using the service
| assistant” on page 48 or “Procedure: Getting node canister and system information
| using a USB key” on page 49. If the node is in service state, fix the reported node
| errors. For more information, go to “Procedure: Fixing node errors” on page 56.
| After the node error is corrected, attempt to create a clustered storage system
| again.
You might encounter a number of situations when you cannot connect to the
service assistant.
| v Check that you have entered the “/service” path after the service IP address.
| Point your web browser to <control enclosure management IP address>/service for
| the node that you want to work on. For example, if you set a service address of
| 11.22.33.44 for a node canister, point your browser to 11.22.33.44/service.
v Check that you are using the correct service address for the node canister. To
find the IPv4 and IPv6 addresses that are configured on the node, go to
“Problem: Unknown service address of a node canister” on page 42. Try
accessing the service assistant through these addresses. Verify that the IP
address, subnet, and gateway are specified correctly for IPv4 addresses. Verify
that the IP address, prefix, and gateway are specified for the IPv6 addresses. If
any of the values are incorrect, see “Procedure: Changing the service IP address
of a node canister” on page 57.
v You cannot connect to the service assistant if the node canister is not able to
start the Storwize V7000 code. To verify that the LEDs indicate that the code is
active, see “Procedure: Understanding the system status using the LEDs” on
page 49.
v The service assistant is configured on Ethernet port 1 of a node canister. Verify
that an Ethernet cable is connected to this port and to an active port on your
Ethernet network. See “Procedure: Finding the status of the Ethernet
connections” on page 55 for details.
v Ping the management address to see if the Ethernet network permits the
connection. If the ping fails, check the Ethernet network configuration to see if
there is a routing or a firewall issue. Check that the Ethernet network
configuration is compatible with the gateway and subnet or prefix settings.
Check that you have not used an address that is used by another device on your
Ethernet network. If necessary, change the network configuration or see
“Procedure: Changing the service IP address of a node canister” on page 57 to
change the service IP address of a node.
| v A default service address is initially assigned to each node canister. The service
| IP address 192.168.70.121 subnet mask 255.255.255.0 is preconfigured on Ethernet
| port 1 of the upper canister, canister 1. The service IP address 192.168.70.122
| subnet mask 255.255.255.0 is preconfigured on Ethernet port 2 of the lower
| canister, canister 2.
| You might not be able to access these addresses because of the following
| conditions:
| – These addresses are the same as the addresses that are used by other devices
| on the network.
| – These addresses cannot be accessed on your network.
| – There are other reasons why they are not suitable for use on your network.
| If the previous conditions apply, see “Procedure: Changing the service IP
| address of a node canister” on page 57 to change the service IP address to one
| that works in your environment.
You must use a supported web browser. Verify that you are using a supported web
browser from the following website:
Switch to using a supported web browser. If the problem continues, contact IBM
Support.
To find out how to resolve the node error, go to “Procedure: Fixing node errors”
on page 56.
When installing a new expansion enclosure, follow the management GUI Add
Enclosure wizard, which is available from the Manage Devices Actions menu.
| When installing a new control enclosure, follow the management GUI Add Control
| and Expansion Enclosures wizard, which is available from the Monitoring >
| System Details menu. After selecting the control enclosure from the navigation
| tree, click the Actions menu, and then select Add Enclosures > Control and
| Expansions.
To confirm that the two copies of a mirrored volume are still identical, choose the
volume view that works best for you. Select one of the volume copies in the
volume that you want to check. From the Actions menu, select the Validate
Volume Copies option.
If you want to resolve any differences, you have the following options:
v Consider that one volume is correct and make the other volume copy match the
other copy if any differences are found. The primary volume copy is the copy
that is considered correct.
v Do not assume that either volume copy is correct. If a difference is found, the
sector is marked. A media error is returned if the volume is read by a host
application.
You might encounter this problem during initial setup or when running commands
if you are using your own USB key rather than the USB key that was packaged
with your order.
You can use this procedure to reset the superuser password if you have forgotten
the password. This command runs differently depending on whether you run it on
a node canister that is active in a clustered system.
Note: If a node canister is not in active state, the superuser password is still
required to log on to the service assistant.
To use a USB key to reset the superuser password, see “USB key and Initialization
tool interface” on page 33.
If the node canister is active in a clustered system, the password for superuser is
changed on the clustered system. If the node canister is not in active state, the
superuser password for the node canister is changed. If the node canister joins a
clustered system later, the superuser password is reset to that of the clustered
system.
Because of the differences between the enclosures, you must be able to distinguish
between the control enclosures and the expansion enclosures when you service the
system. Be aware of the following differences:
v The model type that is shown on the labels. Model types 2076-112, 2076-124,
2076-312, and 2076-324 are control enclosures. Model types 2076-212 and
2076-224 are expansion enclosures.
v The model description that is shown on the left end cap. The description shows
either Control or Expansion.
v The number of ports at the rear of the enclosure. Control enclosures have
Ethernet ports, Fibre Channel ports, and USB ports on the canisters. Expansion
enclosures do not have any of these ports.
v The number of LEDs on the power supply units. The power supply units in the
control enclosure have six LEDs. The power supply units in the expansion
enclosure have four LEDs.
Note: Use the Manage Device options from the management GUI to change the
ID of an enclosure. Use this option to set a unique ID on all your enclosures.
Use the following management GUI functions to find a more detailed status:
| v Monitoring > System Details
| v Home > Manage Device
| v Pools > MDisks by Pools
| v Volumes > Volumes
| v Monitoring > Events and then use the filtering options to display alerts,
| messages, or event types.
To obtain the information, connect to and log on to the service assistant using the
starting service assistant procedure. For more information, go to “Accessing the
service assistant” on page 32.
1. Log on to the service assistant.
Note: If the node that you want to see information about is not the current
node, change it to the current node from the home page.
3. Examine the data shown for the current node.
The home page shows a table of node errors that exist on the node canister and a
table of node details for the current node. The node errors are shown in priority
order.
The node details are divided into several sections. Each section has a tab. Examine
the data that is reported in each tab for the information that you want. The Node
tab shows general information about the node canister that includes the node state
and whether it is a configuration node. The Hardware tab shows information
about the hardware. The Access tab shows the management IP address and the
service address for this node. The Location tab identifies the enclosure in which
the node canister is located.
Use any USB key with a FAT32 file system, a EXT2 file system, or a EXT3 file
system on its first partition.
1. Ensure that the USB key does not contain a file named satask.txt in the root
directory.
If satask.txt does exist in the directory, the node attempts to run the
command that is specified in the file. The information that is returned is
appended to the satask_result.html file. Delete this file if you no longer want
the previous output.
2. Insert the USB key in one of the USB ports of the node canister from which you
want to collect data.
3. The node canister fault LED flashes. It continues to flash while the information
is collected and written to the USB key.
4. Wait until the LED stops flashing before removing the USB key.
Because the LED is a fault indicator, it might remain permanently on or off.
5. View the results in a web browser.
The file contains the details and results of the command that was run and the
status and the configuration information from the node canister.
The LEDs provide a general idea of the system status. You can obtain more detail
from the management GUI and the service assistant. Examine the LEDs when you
are not able to access the management GUI or the service assistant, or when the
The procedure shows the status for the enclosure chassis, power supply units and
batteries, and canisters. It does not show the status for the drives.
The first step is to determine the state of the control enclosure, which includes its
power supply units, batteries, and node canisters. Your control enclosure is
operational if you can manage the system using the management GUI. You might
also want to view the status of the individual power supply units, batteries, or
node canisters.
Find the control enclosure for the system that you are troubleshooting. There is one
control enclosure in a system. If you are unsure which one is the control enclosure,
go to “Procedure: Identifying which enclosure or canister to service” on page 47.
1. Use the state of the ac power failure, power supply OK, fan failure, and dc
power failure LEDs on each power supply unit in the enclosure to determine if
there is power to the system, or if there are power problems. Figure 21 on page
51 shows the LEDs on the power supply unit for the 2076-112 or 2076-124. The
LEDs on the power supply units for the 2076-312 and 2076-324 are similar, but
they are not shown here.
svc00670
Figure 21. LEDs on the power supply units of the control enclosure
2. At least one power supply in the enclosure must indicate Power supply OK or
Power supply firmware downloading for the node canisters to operate. For this
situation, review the three canister status LEDs on each of the node canisters.
Start with the power LED.
Table 20. Power LEDs
Power LED
status Description
Off There is no power to the canister. Try reseating the canister. Go to “Procedure:
Reseating a node canister” on page 60. If the state persists, follow the
hardware replacement procedures for the parts in the following order: node
canister, enclosure chassis.
status Description
Slow Power is available, but the canister is in standby mode. Try to start the node
flashing (1 canister by reseating it. Go to “Procedure: Reseating a node canister” on page
Hz) 60.
Fast The canister is running its power-on self-test (POST). Wait for the test to
flashing (2 complete. If the canister remains in this state for more than 10 minutes, try
Hz) reseating the canister. Go to “Procedure: Reseating a node canister” on page
60. If the state persists, follow the hardware replacement procedure for the
node canister.
svc00672
Figure 22. LEDs on the node canisters
3. If the power LED is on, consider the states of the clustered-system status and
fault LEDs.
Table 21. System status and fault LEDs
System status
To review the status of the control enclosure batteries, see Table 22.
Table 22. Control enclosure battery LEDs
Battery Good Battery Fault
+ - + -
Description Action
On Off Battery is good and fully None
charged.
Flashing off Battery is good but not fully None
charged. The battery is either
charging or a maintenance
discharge is being performed.
Off On Nonrecoverable battery fault. Replace the battery. If
replacing the battery
does not fix the issue,
replace the power supply
unit.
Ensure that the software is active on the node before you begin this procedure. Go
to “Procedure: Understanding the system status using the LEDs” on page 49.
Ethernet port 1 must be connected to an active port on your Ethernet network.
Determine the state of the Ethernet LEDs by using one of the following methods:
v Use the USB key to obtain the most comprehensive information for the node
status. Go to “Procedure: Getting node canister and system information using a
USB key” on page 49.
The status, speed, and MAC address are returned for each port. Information is
returned that identifies whether the node is the configuration node and if any
node errors were reported.
v Examine the LEDs of the Ethernet ports. For the status of the LEDs, go to
“Ethernet ports and indicators” on page 11.
The activity LED flashes when there is activity on the connection. The link state
LED must be permanently on. If it is off, the link is not connected.
If your link is not connected, perform the following actions to check the port status
each time until it is corrected or connected:
1. Verify that each end of the cable is securely connected.
2. Verify that the port on the Ethernet switch or hub is configured correctly.
3. Connect the cable to a different port on your Ethernet network.
4. If the status is obtained using the USB key, review all the node errors that are
reported.
5. Replace the Ethernet cable.
6. For the 10 Gbps Ethernet port, replace the small form-factor pluggable (SFP)
transceiver. See “Replacing an SFP transceiver” on page 83.
7. Follow the hardware replacement procedure for a node canister. See “Replacing
a node canister” on page 79.
Attention: If the enclosure reaches a point where the system data is not available
on any node canister in the system, you have to perform a system recovery. This
recovery is an extended service action and might not be able to recover all of your
volumes. Do not perform this action to remove the system data from a node unless
there is a node canister with saved system information in the enclosure. Do not
remove the system data from a node unless instructed to do so by a service
procedure.
1. Start the service assistant on the node canister.
2. Use the service assistant node action to hold the node in service state.
3. Use the Manage System option to remove the system data from the node.
The node restarts in service state. When you want the node canister to be active
again, use the service assistant home page action to leave service state.
Attention: This procedure makes all the volume data that you have on your
system inaccessible. You cannot recover the data. This procedure affects all
volumes that are managed by your system, which includes the drives in the
control enclosure, in the expansion enclosures, and on the managed disks on
external storage systems. The only volumes not affected are image mode volumes
on external storage systems.
Do not continue unless you are certain that you want to remove all the volume
data and configuration data from your system. This procedure is not used as part
of any recovery action.
There are two stages to this procedure. First, the node canisters are reset. Second,
the enclosure data is reset.
1. Start the service assistant on one of the node canisters.
2. Use the service assistant node action to hold the node in service state.
3. Use the Manage System option to remove the system data from the node.
4. Perform the previous steps on the second node canister in the enclosure and
then on every node in every other enclosure in the system.
5. On one node in every enclosure, open the service assistant Configure
Enclosure and select the Reset System ID option.
This action causes the system to reset.
6. Power each enclosure off and on before creating a system.
When you change an IPv4 address, you change the IP address, the subnet, mask,
and gateway. When you change an IPv6 address, you change the IP address,
prefix, and gateway.
Which method to use depends on the status of the system and the other node
canisters in the system. Follow the methods in the order shown until you are
successful in setting the IP address to the required value.
You can set an IPv4 address, an IPv6 address, or both, as the service address of a
node. Enter the required address correctly. If you set the address to 0.0.0.0 or
0000:0000:0000:0000:0000:0000:0000, you disable the access to the port on that
protocol.
In these situations, you must manually create an satask.txt file on a USB key to
initialize your clustered system. Use the USB key that was supplied with your
system or any USB key that is formatted with a FAT32, an EXT2, or an EXT3 file
system on its first partition.
1. Open a file editor that can create ASCII text files.
2. Create a file called satask.txt.
3. Add a single line of command text to the file.
If you are creating a clustered system with an IPv4 address, the command line
is like the following string:
satask mkcluster -clusterip aaa.aaa.aaa.aaa
-gw ggg.ggg.ggg.ggg -mask mmm.mmm.mmm.mmm
For more information about using the USB key, see “USB key and Initialization
tool interface” on page 33.
Note: The service assistant gives you the option to create a clustered system only
if the node state is candidate.
To initialize a clustered system using the service assistant, perform the following
steps:
| 1. Point your web browser to the service address of the upper node canister in
| your control enclosure: 192.168.70.121 subnet mask: 255.255.255.0.
2. Log on with the superuser password. The default password is passw0rd.
If you are unable to connect, see “Problem: Cannot connect to the service
assistant” on page 43.
If the default service assistant address cannot be used in your network
environment, connect using a direct Ethernet connection. To make this
connection, see “Procedure: Accessing a canister using a directly attached
Ethernet cable.”
3. Select Manage System.
4. Enter the system name and the management IP address.
5. Click Create System.
Attention: Without a USB key to service the system, it is not possible to reset the
superuser password or to change the service IP addresses in the event of a fault
that prevents access to the management interface. It is essential that you take steps
to record this information for use in the event of a failure.
| Perform this procedure if you are not authorized to use a USB key in your data
| center and when the service address of your nodes cannot be accessed over your
| Ethernet network. This situation might occur for a new installation where the
Note: Do not attempt to use a directly attached Ethernet cable to a canister that is
active in a clustered system. You might disrupt access from host applications or the
management GUI. If the node is active, use the management GUI network
configuration options to set the service IP to an address that is accessible on the
network.
Perform the following steps to access a canister using a directly attached Ethernet
cable:
1. Connect one end of an Ethernet cable to Ethernet port 1 of the upper node
canister.
Verify that you are reseating the correct node canister and that you use the correct
canister handle for the node that you are reseating. Handles for the node canisters
are located next to each other. The handle on the right operates the upper canister.
The handle on the left operates the lower canister.
1. Verify the clustered-system status LED on the node canister. If it is
permanently on, the node is active. If the node is active, no reseating is
required.
2. Verify that you have selected the correct node canister and verify why you are
reseating it. Go to “Procedure: Identifying which enclosure or canister to
service” on page 47.
If you reseat a node that is active, it cannot store its state data and cannot
restart without other service actions.
If the other node canister in the enclosure is not active, reseating the node
canister while it is active results in loss of the data on your volumes and the
system is unavailable to hosts.
3. Grasp the handle between the thumb and forefinger.
4. Squeeze them together to release the handle.
To power off your Storwize V7000 system, use the following steps:
1. Stop hosts.
| 2. Shut down the system by using the management GUI. Click Monitoring >
| System Details. From the Actions menu, select Shut Down System.
3. Wait for the power LED on both node canisters in all control enclosures to start
flashing, which indicates that the shutdown operation has completed.
The following figure shows the LEDs on the node canisters. The power LED is
the LED on the left when the canister is top-side up.
1
svc00672
The management GUI and the service assistant have features to assist you in
collecting the required information. The management GUI collects information
from all the components in the system. The service assistant collects information
from a single node canister. When the information that is collected is packaged
together in a single file, the file is called a snap.
Always follow the instructions that are given by the support team to determine
whether to collect the package by using the management GUI or the service
assistant. Instruction is also given for which package content option is required.
| v If you are collecting the package by using the management GUI, select Settings
| > Support. Click Download Support Package. Follow the instructions.
v If you are collecting the package by using the service assistant, ensure that the
node that you want to collect logs from is the current node. Select the Collect
Logs option from the navigation. You can collect a support package or copy an
individual file from the node canister. Follow the instructions to collect the
information.
A failure has indicated that the node software is damaged and must be reinstalled.
Use the service assistant to reinstall the software.
1. Ensure that the node you want to reinstall the software on is the current node.
Go to “Accessing the service assistant” on page 32.
2. Select Reinstall Software from the navigation.
3. Select Rescue from another node.
SAN failures might cause Storwize V7000 drives to be inaccessible to host systems.
Failures can be caused by SAN configuration changes or by hardware failures in
SAN components.
The following list identifies some of the hardware that might cause failures:
v Power, fan, or cooling switch
v Application-specific integrated circuits
v Installed small form-factor pluggable (SFP) transceiver
v Fiber-optic cables
Perform the following steps if you were sent here from the error codes:
1. Verify that the power is turned on to all switches and storage controllers that
the Storwize V7000 system uses, and that they are not reporting any hardware
failures. If problems are found, resolve those problems before proceeding
further.
2. Verify that the Fibre Channel cables that connect the systems to the switches
are securely connected.
3. If you have a SAN management tool, use that tool to view the SAN topology
and isolate the failing component.
The following items can indicate that a single Fibre Channel link has failed:
v The customer's SAN monitoring tools
v The Fibre Channel status LEDs at the rear of the node canister
v An error that indicates that a single port has failed
Attempt each of the following actions, in the following order, until the failure is
fixed:
1. Ensure that the Fibre Channel cable is securely connected at each end.
2. Replace the Fibre Channel cable.
3. Replace the SFP transceiver for the failing port on the Storwize V7000 node.
Note: Storwize V7000 nodes are supported with both longwave SFP
transceivers and shortwave SFP transceivers. You must replace an SFP
transceiver with the same type of SFP transceiver. If the SFP transceiver to
replace is a longwave SFP transceiver, for example, you must provide a suitable
replacement. Removing the wrong SFP transceiver could result in loss of data
access.
4. Perform the Fibre Channel switch service procedures for a failing Fibre Channel
link. This might involve replacing the SFP transceiver at the switch.
5. Contact IBM Support for assistance in replacing the node canister.
The following guidelines apply to all storage systems that are attached to the
Storwize V7000 system:
v Always follow the service instructions that are provided in the documentation
for your storage system.
v Ensure that there are no unfixed errors in the event log before you perform any
service procedures.
v After you perform a service procedure, check the event log and fix any errors.
Expect to see the following types of errors:
– MDisk error recovery procedures (ERPs)
– Reduced paths
Attention: Perform service actions only when directed by the fix procedures. If
used inappropriately, service actions can cause loss of access to data or even data
loss. Before attempting to recover a storage system, investigate the cause of the
failure and attempt to resolve those issues by using other fix procedures. Read and
understand all of the instructions before performing any action.
Attention: Do not attempt the recovery procedure unless the following conditions
are met:
v All hardware errors are fixed.
v All node canisters have candidate status.
| v All node canisters must be at the same level of software that the storage system
| had before the system failure. If any node canisters were modified or replaced,
| use the service assistant to verify the levels of software, and where necessary, to
| upgrade or downgrade the level of software.
The system recovery procedure is one of several tasks that must be performed. The
following list is an overview of the tasks and the order in which they must be
performed:
1. Preparing for system recovery
a. Review the information regarding when to run the recover system
procedure
b. Fix your hardware errors
c. Remove the system information for node canisters with error code 550 or
error code 578 by using the service assistant.
2. Performing the system recovery. After you prepared the system for recovery
and met all the pre-conditions, run the system recovery.
Note: Run the procedure on one system in a fabric at a time. Do not perform
the procedure on different node canisters in the same system. This restriction
also applies to remote systems.
3. Performing actions to get your environment operational
v Recovering from offline VDisks (volumes) by using the CLI
v Checking your system, for example, to ensure that all mapped volumes can
access the host.
Attention: If you experience failures at any time while you are running the
recover system procedure, call the IBM Support Center. Do not attempt to do
further recovery actions because these actions might prevent IBM Support from
restoring the system to an operational status.
Certain conditions must be met before you run the recovery procedure. Use the
following items to help you determine when to run the recovery procedure:
Note: It is important that you know the number of control enclosures in the
system, and when the instructions indicate that every node is checked, you must
check the status of both nodes in every control enclosure. For some system
problems or Fibre Channel network problems, you must run the service assistant
directly on the node to get its status.
v Check to see if any node in the system has a node status of active. This status
means that the system is still available. In this case, recovery is not necessary.
v Do not recover the system if the management IP address is available from
another node. Ensure that all service procedures have been run.
v Check the node status of every node canister that is part of this system. Resolve
all hardware errors except node error 578 or node error 550.
– All nodes must be reporting either a node error 578 or a node error 550.
These error codes indicate that the system has lost its configuration data. If
any nodes report anything other than these error codes, do not perform a
recovery. You can encounter situations where non-configuration nodes report
other node errors, such as a 550 node error. The 550 error can also indicate
that a node is not able to join a system.
– If any nodes show a node error 550, record the error data that is associated
with the 550 error from the service assistant.
- In addition to the node error 550, the report can show data that is
separated by spaces in one of the following forms:
v Node identifiers in the format: <enclosure_serial>-<canister slot ID><7
characters, hyphen, 1 number), for example, 01234A6-2
v Quorum drive identifiers in the format: <enclosure_serial>:<drive slot
ID>[<drive 11S serial number>] (7 characters, colon, 1 or 2 numbers,
open square bracket, 22 characters, close square bracket), for example,
01234A9:21[11S1234567890123456789]
v Quorum MDisk identifier in the format: WWPN/LUN (16 hexadecimal
digits followed by a forward slash and a decimal number), for example,
1234567890123456/12
- If the error data contains a node identifier, ensure that the node that is
referred to by the ID is showing node error 578. If the node is showing a
node error 550, ensure that the two nodes can communicate with each
other. Verify the SAN connectivity, and if the 550 error is still present,
restart one of the two nodes by clicking Restart Node from the service
assistant.
- If the error data contains a quorum drive identifier, locate the enclosure
with the reported serial number. Verify that the enclosure is powered on
and that the drive in the reported slot is powered on and functioning. If the
node canister that is reporting the fault is in the I/O group of the listed
Note: If after resolving all these scenarios, half or greater than half of the
nodes are reporting node error 578, it is appropriate to run the recovery
procedure. You can also call IBM Support for further assistance.
– For any nodes that are reporting a node error 550, ensure that all the missing
hardware that is identified by these errors is powered on and connected
without faults. If you cannot contact the service assistant from any node,
isolate the problems by using the LED indicators.
– If you have not been able to restart the system and if any node other than the
current node is reporting node error 550 or 578, you must remove system
data from those nodes. This action acknowledges the data loss and puts the
nodes into the required candidate state.
v Do not attempt to recover the system if you have been able to restart it.
v If back-end MDisks are removed from the configuration, those volumes that
depended on that hardware cannot be recovered. All previously configured
back-end hardware must be present for a successful recovery.
v Any nodes that were replaced must have the same WWNN as the nodes that
they replaced.
| v If any of the node canisters were replaced, they must not have participated in
| any other system. You can resolve this issue by performing a node rescue on the
| affected canister by using the service assistant. Do not perform this action on
| any of the other node canisters.
v The configuration backup file must be up to date. If any configuration changes
had been made since the backup was taken, the data is inconsistent and further
investigation is needed. Manual changes are required after the system is
recovered.
Obtain a basic understanding about the hardware failure. In most situations when
there is no clustered system, a power issue is the cause. For example, both power
supplies might have been removed.
| Before performing this task, ensure that you have read the introductory
| information in the overall recover system procedure.
To remove system information from a node canister with an error 550 or 578,
follow this procedure using the service assistant:
1. Point your browser to the service IP address of one of the nodes, for example,
https://node_service_ip_address/service/.
If you do not know the IP address or if it has not been configured, you must
assign an IP address using the initialization tool.
2. Log on to the service assistant.
3. Select Manage System.
4. Click Remove System Data.
5. Confirm that you want to remove the system data when prompted.
6. Remove the system data for the other nodes that display a 550 or a 578 error.
All nodes previously in this system must have a node status of Candidate and
have no errors listed against them.
7. Resolve any hardware errors until the error condition for all nodes in the
system is None.
8. Ensure that all nodes in the system display a status of candidate.
When all nodes display a status of candidate and all error conditions are None,
you can run the recovery procedure.
| All node canisters must be at the same level of software that the storage system
| had before the system failure. If any node canisters were modified or replaced, use
| the service assistant to verify the levels of software, and where necessary, to
| upgrade or downgrade the level of software.
Note: Your web browser must not block pop-up windows; otherwise, progress
windows cannot open.
The recovery can be run from any node canisters in the system. The node canisters
must not have participated in any other system.
| Note: Each individual stage of the recovery procedure might take significant time
| to complete. The time to complete depends on your specific configuration.
| Before performing this task, ensure that you have read the introductory
| information in the overall recover system procedure.
1. Point your browser to the service IP address of one of the node canisters.
If you do not know the IP address or if it has not been configured, you must
assign an IP address using the initialization tool.
2. Log on to the service assistant.
3. Select Recover System from the navigation.
4. Follow the online instructions to complete the recovery procedure.
Verify the date and time of the last quorum time. The time stamp must be less
than 10 minutes before the failure. The time stamp format is YYYYMMDD
hh:mm, where YYYY is the year, MM is the month, DD is the day, hh is the
hour, and mm is the minute.
Attention: If the time stamp is not less than 10 minutes before the failure, call
IBM Support.
Verify the date and time of the last backup date. The time stamp must be less
than 24 hours before the failure. The time stamp format is YYYYMMDD hh:mm,
where YYYY is the year, MM is the month, DD is the day, hh is the hour, and
mm is the minute.
Attention: If the time stamp is not less than 24 hours before the failure, call
IBM Support.
Changes made after the time of this backup date might not be restored.
After the recovery completes successfully, perform the checks to get your
environment operational.
| If any errors are logged in the error log after the system recovery procedure
| completes, use the fix procedures to resolve these errors, especially the errors
| related to offline arrays.
If you have performed the recovery procedure, and it has completed successfully
but there are offline volumes, you can perform the following steps to bring the
volumes back online. Any volumes that are offline and are not thin-provisioned
volumes are offline because of the loss of write-cache data during the event that
led both nodes to lose their hardened data. These volumes might need additional
recovery steps after the volume is brought back online.
| Note: If you encounter errors in the error log after running the recovery procedure
| that are related to offline arrays, use the fix procedures to resolve the offline array
| errors before fixing the offline volume (VDisk) errors.
Perform the following steps to recover an offline volume after the recovery
procedure has completed:
1. Delete all IBM FlashCopy® function mappings and Metro Mirror or Global
Mirror relationships that use the offline volumes.
| 2. Run the recovervdisk or recovervdiskbysystem command.
| You can recover individual volumes by using the recovervdisk command. You
| can recover all the volumes in a clustered system by using the
| recovervdiskbysystem command.
3. Recreate all FlashCopy mappings and Metro Mirror or Global Mirror
relationships that use the volumes.
Configuration data for the system provides information about your system and the
objects that are defined in it. The backup and restore functions of the svcconfig
command can back up and restore only your configuration data for the Storwize
V7000 system. You must regularly back up your application data by using the
appropriate backup methods.
You can maintain your configuration data for the system by completing the
following tasks:
v Backing up the configuration data
v Restoring the configuration data
v Deleting unwanted backup configuration data files
Before you back up your configuration data, the following prerequisites must be
met:
v No independent operations that change the configuration for the system can be
running while the backup command is running.
v No object name can begin with an underscore character (_).
Note:
v The default object names for controllers, I/O groups, and managed disks
(MDisks) do not restore correctly if the ID of the object is different from what is
recorded in the current configuration data file.
v All other objects with default names are renamed during the restore process. The
new names appear in the format name_r where name is the name of the object in
your system.
Before you restore your configuration data, the following prerequisites must be
met:
v You have the Security Administrator role associated with your user name and
password.
v You have a copy of your backup configuration files on a server that is accessible
to the system.
v You have a backup copy of your application data that is ready to load on your
system after the restore configuration operation is complete.
v You know the current license settings for your system.
v You have not removed any hardware since the last backup of your
configuration.
v No zoning changes have been made on the Fibre Channel fabric which would
prevent communication between the Storwize V7000 and any storage controllers
which are present in the configuration.
v For configurations with more than one I/O group, if a new system is created on
which the configuration data is to be restored, the I/O groups for the other
control enclosures must be added.
The Storwize V7000 analyzes the backup configuration data file and the system to
verify that the required disk controller system nodes are available.
Before you begin, hardware recovery must be complete. The following hardware
must be operational: hosts, Storwize V7000, drives, the Ethernet network, and the
SAN fabric.
Before you back up your configuration data, the following prerequisites must be
met:
v No independent operations that change the configuration can be running while
the backup command is running.
v No object name can begin with an underscore character (_).
v The default object names for controllers, I/O groups, and managed disks
(MDisks) do not restore correctly if the ID of the object is different from what is
recorded in the current configuration data file.
v All other objects with default names are renamed during the restore process. The
new names appear in the format name_r.
where name is the name of the object in your system.
You must regularly back up your configuration data and your application data to
avoid data loss. If a system is lost after a severe failure occurs, both configuration
of the system and application data is lost. You must reinstate the system to the
exact state it was in before the failure, and then recover the application data.
The SSH coding examples that are provided are samples using the PuTTY scp
(pscp) application code. The pscp application is available when you install an SSH
client on your host system. You can access the pscp application through a
Microsoft Windows command prompt.
The following output is an example of the messages that are displayed during
the backup process:
CMMVC6112W io_grp io_grp1 has a default name
CMMVC6112W io_grp io_grp2 has a default name
CMMVC6112W mdisk mdisk14 ...
CMMVC6112W node node1 ...
CMMVC6112W node node2 ...
....................................................
The svcconfig backup CLI command creates three files that provide
information about the backup process and the configuration. These files are
created in the /tmp directory of the configuration node.
The following table describes the three files that are created by the backup
process:
If the process fails, resolve the errors, and run the process again.
7. Issue the following command to exit the system:
exit
8. Issue the following command to copy the backup files to a location that is not
in your system:
pscp -i ssh_private_key_file superuser@cluster_ip:/tmp/svc.config.backup.*
/offclusterstorage/
where cluster_ip is the IP address or DNS name of the system and
offclusterstorage is the location where you want to store the backup files.
Tip: To maintain controlled access to your configuration data, copy the backup
files to a location that is password-protected.
9. Ensure that the copies of the backup files are stored in the location that you
specified in step 8 on page 73.
You can rename the backup files to include the configuration node name either at
the start or end of the file names so that you can easily identify these files when
you are ready to restore your configuration.
Issue the following command to rename the backup files that are stored on a Linux
or IBM AIX® host:
mv /offclusterstorage/svc.config.backup.xml
/offclusterstorage/svc.config.backup.xml_myconfignode
where offclusterstorage is the name of the directory where the backup files are
stored and myconfignode is the name of your configuration node.
To rename the backup files that are stored on a Windows host, right-click the name
of the file and select Rename.
You must regularly back up your configuration data and your application data to
avoid data loss. If a system is lost after a severe failure occurs, both configuration
for the system and application data is lost. You must reinstate the system to the
exact state it was in before the failure, and then recover the application data.
Important: There are two phases during the restore process: prepare and execute.
You must not change the fabric or system between these two phases.
If you do not understand the instructions to run the CLI commands, see the
command-line interface reference information.
Note: A node that is powered off might not show up in this list of nodes
for the system. Diagnose hardware problems directly on the node using
the service assistant IP address and by physically verifying the LEDs for
the hardware components.
2. Verify that all nodes are available as candidate nodes with blank system fields.
Perform the following steps on one node in each control enclosure:
a. Connect to the service assistant on either of the nodes in the control
enclosure.
b. Select Configure Enclosure.
c. Select the Reset the system ID option. Do not make any other changes on
the panel.
d. Click Modify to make the changes.
3. Use the initialization tool that is available on the USB key to initialize the
system with the IP address. Go to “Using the initialization tool” on page 34.
4. In a supported browser, enter the IP address that you used to initialize the
system and the default superuser password (passw0rd).
5. At this point the setup wizard is shown. Be aware of the following items:
a. Accept the license agreements.
b. Set the values for the system name, date and time settings, and the system
licensing. The original settings are restored during the configuration
restore process.
c. Verify the hardware. Only the control enclosure on which the clustered
system was created and directly attached expansion enclosures are
displayed. Any other control enclosures and expansion enclosures in other
I/O groups will be added to the system in step 6.
d. On the Configure Storage panel, deselect Yes, automatically configure
internal storage now. Any internal storage configuration is recovered after
the system is restored.
6. For configurations with more than one I/O group, if a new system is created
on which the configuration data is to be restored, add the rest of the control
enclosures into the clustered system.
| a. From the management GUI, select Monitoring > System Details.
b. Select the system name in the tree.
c. Go to Actions > Add Enclosures > Control and Expansions
Note: Because the RSA host key has changed, a warning message might
display when you connect to the system using SSH.
9. Identify the configuration backup file that you want to restore from.
The file can be either a local copy of the configuration backup XML file that
you saved when backing up the configuration or an up-to-date file on one of
the nodes.
Configuration data is automatically backed up daily at 01:00 system time on
the configuration node.
Attention: You must copy the required backup file to another computer
before you continue. To save a copy of the data, perform the following steps
to check for backup files on both nodes:
| a. From the management GUI, click Settings > Support.
b. Click Show full log listing.
c. Find the file name that begins with svc.config.cron.xml.
d. Double-click the file to download the file to your computer.
10. Issue the following CLI command to remove all of the existing backup and
restore configuration files that are located on your configuration node in the
/tmp directory:
svcconfig clear -all
11. The XML files contain a date and time that can be used to identify the most
recent backup. After you identify the backup XML file that is to be used when
you restore the system, rename the file to svc.config.backup.xml. From your
desktop, issue the following command to copy the file back on to the system.
pscp -i ssh_private_key_file
full_path_to_identified_svc.config.backup.xml
superuser@cluster_ip:/tmp/
12. Issue the following CLI command to compare the current configuration with
the backup configuration data file:
svcconfig restore -prepare
This CLI command creates a log file in the /tmp directory of the configuration
node. The name of the log file is svc.config.restore.prepare.log.
Note: You might receive a warning stating that a licensed feature is not
enabled. This message means that after the recovery process, the current
license settings do not match the previous license settings. The recovery
process continues normally and you can enter the correct license settings in
the management GUI at a later time.
When you log into the CLI again over SSH, you see this output:
IBM_2076:your_cluster_name:superuser>
You can remove any unwanted configuration backup and restore files from the
/tmp directory on your configuration by issuing the following CLI command:
Even though many of these procedures are hot-swappable, these procedures are
intended to be used only when your system is not up and running and performing
I/O operations. Unless your system is offline, go to the management GUI and
follow the fix procedures.
Each replaceable unit has its own removal procedure. Sometimes you can find that
a step within a procedure might refer you to a different remove and replace
procedure. You might want to complete the new procedure before you continue
with the first procedure that you started.
Be careful when you are replacing the hardware components that are located in the
back of the system that you do not inadvertently disturb or remove any cables that
you are not instructed to remove.
First, read the safety precautions in the IBM Storwize V7000 Safety Notices. These
guidelines help you safely work with the Storwize V7000.
Be careful when you are replacing the hardware components that are located in the
back of the system that you do not inadvertently disturb or remove any cables that
you are not instructed to remove.
Attention: Do not replace one type of node canister with another type. For
example, do not replace a model 2076-112 node canister with a model 2076-312
node canister.
Attention: Even if a node canister is powered off, it is still possible to lose data.
Do not remove a node canister unless directed to do so by a service procedure.
Note: Ensure that you are opening the correct handle. The handle locations
for the node canisters and expansion canisters are slightly different.
| Handles for the node canisters are located in close proximity to each other.
| The handle with the finger grip on the right removes the upper canister (1).
| The handle with the finger grip on the left removes the lower canister (2).
svc00744
Even though many of these procedures are hot-swappable, these procedures are
intended to be used only when your system is not up and running and performing
I/O operations. Unless your system is offline, go to the management GUI and
follow the fix procedures.
Be careful when you are replacing the hardware components that are located in the
back of the system that you do not inadvertently disturb or remove any cables that
you are not instructed to remove.
Note: Ensure that you are opening the correct handle. The handle locations
for the node canisters and expansion canisters are slightly different.
Handles for the upper and lower expansion canisters overlap each other. The
handle with the finger grip on the left removes the upper canister (1). The
handle with the finger grip on the right removes the lower canister (2).
2 svc00745
Even though many of these procedures are hot-swappable, these procedures are
intended to be used only when your system is not up and running and performing
I/O operations. Unless your system is offline, go to the management GUI and
follow the fix procedures.
Be careful when you are replacing the hardware components that are located in the
back of the system that you do not inadvertently disturb or remove any cables that
you are not instructed to remove.
CAUTION:
Some laser products contain an embedded Class 3A or Class 3B laser diode.
Note the following information: laser radiation when open. Do not stare into the
beam, do not view directly with optical instruments, and avoid direct exposure
to the beam. (C030)
Perform the following steps to remove and then replace an SFP transceiver:
Important: The Fibre Channel links in the enclosures are supported with both
longwave SFP transceivers and shortwave SFP transceivers. A longwave SFP
transceiver has some blue components that are visible even when the SFP
transceiver is plugged in. You must replace an SFP transceiver with the same
type of SFP transceiver that you are replacing. If the SFP transceiver to replace
is a longwave SFP transceiver, for example, you must replace with another
longwave SFP transceiver. Removing the wrong SFP transceiver might result in
loss of data access.
2. Remove the optical cable by pressing the release tab and pulling the cable out.
Be careful to exert pressure only on the connector and do not pull on the
optical cables.
3. Remove the SFP transceiver. There are a number of different handling or
locking mechanisms that are used on the SFP transceivers. Some SFP
transceivers might have a plastic tag. If so, pull the tag to remove the SFP
transceiver.
Important: Always check that the SFP transceiver that you replace matches the
SFP transceiver that you remove.
4. Push the new SFP transceiver into the aperture and ensure that it is securely
pushed home. The SFP transceiver usually locks into place without having to
swing the release handle until it locks flush with the SFP transceiver. Figure 28
illustrates an SFP transceiver and its release handle.
svc00418
DANGER
When working on or around the system, observe the following precautions:
(D005)
Attention: A powered-on enclosure must not have a power supply removed for
more than five minutes because the cooling does not function correctly with an
empty slot. Ensure that you have read and understood all these instructions and
have the replacement available, and unpacked, before you remove the existing
power supply.
Even though many of these procedures are hot-swappable, these procedures are
intended to be used only when your system is not up and running and performing
I/O operations. Unless your system is offline, go to the management GUI and
follow the fix procedures.
Be careful when you are replacing the hardware components that are located in the
back of the system that you do not inadvertently disturb or remove any cables that
you are not instructed to remove.
Ensure that you are aware of the procedures for handling static-sensitive devices
before you replace the power supply.
| A replacement power supply unit is not shipped with a battery; therefore, transfer
| the battery from the existing power supply unit to the replacement unit. To
| transfer a battery, go to “Replacing a battery in a power supply unit” on page 93.
b. Grip the handle to pull the power supply out of the enclosure as shown in
Figure 30.
svc00632
6. Insert the replacement power supply unit into the enclosure with the handle
pointing towards the center of the enclosure. Insert the unit in the same
orientation as the one that you removed.
If required, return the power supply. Follow all packaging instructions, and use
any packaging materials for shipping that are supplied to you.
DANGER
When working on or around the system, observe the following precautions:
(D005)
Attention: A powered-on enclosure must not have a power supply removed for
more than five minutes because the cooling does not function correctly with an
empty slot. Ensure that you have read and understood all these instructions and
have the replacement available, and unpacked, before you remove the existing
power supply.
Even though many of these procedures are hot-swappable, these procedures are
intended to be used only when your system is not up and running and performing
I/O operations. Unless your system is offline, go to the management GUI and
follow the fix procedures.
Be careful when you are replacing the hardware components that are located in the
back of the system that you do not inadvertently disturb or remove any cables that
you are not instructed to remove.
Ensure that you are aware of the procedures for handling static-sensitive devices
before you replace the power supply.
To replace the power supply unit in an expansion enclosure, perform the following
steps:
1. Read the safety information to which “Preparing to remove and replace parts”
on page 79 refers.
2. Examine the Identify LED that is lit on the front of the enclosure to identify
the correct enclosure.
3. Turn off the power to the power supply units using the switches at the back
of the units.
4. Disconnect the cable retention bracket and the power cord from the power
supply that you are replacing.
5. Remove the power supply unit. Record the orientation of the power supply
unit. Power supply unit 1 is top side up, and power supply unit 2 is inverted.
a. Depress the black locking catch from the side with the colored sticker as
shown in Figure 31 on page 91.
b. Grip the handle to pull the power supply out of the enclosure as shown in
Figure 32.
svc00632
6. Insert the replacement power supply unit into the enclosure with the handle
pointing towards the center of the enclosure. Insert the unit in the same
orientation as the one that you removed.
If required, return the power supply. Follow all packaging instructions, and use
any packaging materials for shipping that are supplied to you.
DANGER
When working on or around the system, observe the following precautions:
(D005)
Even though many of these procedures are hot-swappable, these procedures are
intended to be used only when your system is not up and running and performing
I/O operations. Unless your system is offline, go to the management GUI and
follow the fix procedures.
Be careful when you are replacing the hardware components that are located in the
back of the system that you do not inadvertently disturb or remove any cables that
you are not instructed to remove.
Each power supply unit in a control enclosure contains an integrated battery that
is used during temporary short-term power outages. You must replace the battery
with the exact same model.
To replace the battery in the power supply unit of the control enclosure, perform
the following steps:
1. Read the safety information to which “Preparing to remove and replace parts”
on page 79 refers.
2. Follow the removing steps of the replacing a power-supply unit procedure. Go
to “Replacing a power supply unit for a control enclosure” on page 85.
3. Remove the battery, as shown in Figure 33 on page 95.
svc00611
Figure 33. Removing the battery from the control enclosure power-supply unit
If required, return the battery. Follow all packaging instructions, and use any
packaging materials for shipping that are supplied to you.
Be careful when you are replacing the hardware components that are located in the
back of the system that you do not inadvertently disturb or remove any cables that
you are not instructed to remove.
Each cable retention bracket comes attached to the back of the power supply unit
by the power cord plug-in.
Attention: If your drive is configured for use, go to the management GUI and
follow the fix procedures. Performing the replacement actions without the
assistance of the fix procedures results in loss of data or access to data.
Attention: Do not leave a drive slot empty. Do not remove a drive or drive
assembly before you have a replacement available.
The drives can be distinguished from the blank carriers by the color-coded striping
on the drive. The drives are marked with an orange striping. The blank carriers are
marked with a blue striping.
To replace the drive assembly or blank carrier, perform the following steps:
1. Read the safety information to which “Preparing to remove and replace parts”
on page 79 refers.
2. Unlock the assembly by squeezing together the tabs on the side.
svc00613
Attention: If your drive is configured for use, go to the management GUI and
follow the fix procedures. Performing the replacement actions without the
assistance of the fix procedures results in loss of data or access to data.
Attention: Do not leave a drive slot empty. Do not remove a drive or drive
assembly before you have a replacement available.
To replace the drive assembly or blank carrier, perform the following steps:
1. Read the safety information to which “Preparing to remove and replace parts”
on page 79 refers.
2. Unlock the module by squeezing together the tabs at the top.
svc00614
Be careful when you are replacing the hardware components that are located in the
back of the system that you do not inadvertently disturb or remove any cables that
you are not instructed to remove.
Note: If you are replacing a single cable, this step is not necessary.
svc00651
Figure 38. SAS cable
Note: Ensure that you know the type of enclosure chassis that you are replacing.
The procedures for replacing a control enclosure chassis are different from those
procedures for replacing an expansion enclosure chassis. For information about
replacing an expansion enclosure chassis, see “Replacing an expansion enclosure
chassis” on page 105.
(D005)
Be careful when you are replacing the hardware components that are located in the
back of the system that you do not inadvertently disturb or remove any cables that
you are not instructed to remove.
Ensure that you are aware of the procedures for handling static-sensitive devices
before you remove the enclosure.
Note: If you continue while the system is still active, you risk losing the
clustered system configuration and volume cache data that is stored in the
canister.
6. Turn off the power to the enclosure using the switches on the power supply
units.
7. Record which data cables are plugged into the specific ports. The cables must
be inserted back into the same ports after the replacement is complete;
otherwise, the system cannot function properly.
8. Disconnect the cable retention brackets and the power cords from the power
supply units.
9. Disconnect the data cables for each canister.
10. Remove the power supply units from the enclosure.
11. Remove the canisters from the enclosure. Record the location of each canister.
They must be inserted back into the same location in the new enclosure.
12. Remove all the drives and blank drive assemblies from the enclosure. Record
the location for each drive. They must be inserted back into the same location
in the new enclosure.
13. Remove both enclosure end caps from the enclosure. Keep the left end cap
because it is used again.
14. Remove the clamping screws that attached the enclosure to the rack cabinet.
15. Remove the enclosure chassis from the front of the rack cabinet and take the
chassis to a work area.
16. Install the new enclosure chassis in the rack cabinet.
17. Remove the end caps from the new enclosure and install the clamping screws
that attach the enclosure to the rack cabinet.
18. Replace the end caps. Use the new right end cap and use the left end cap that
you removed in step 13.
Using the left end cap that you removed preserves the model and serial
number identification.
19. Reinstall the drives in the new enclosure. The drives must be inserted back
into the same location from which they were removed on the old enclosure.
20. Reinstall the canisters in the enclosure. The canisters must be inserted back
into the same location from which they were removed on the old enclosure.
21. Install the power supply units.
22. Reattach the data cables to each canister using the information that you
recorded previously.
Note: The cables must be inserted back into the same ports from which they
were removed on the old enclosure; otherwise, the system cannot function
properly.
23. Attach the power cords and the cable retention brackets to the power supply
units.
24. Write the old enclosure machine type and model (MTM) and serial number on
the repair identification (RID) tag that is supplied. Attach the tag to the left
flange at the back of the enclosure.
| Note: There are situations where the canisters restart and report critical
| node error 508. If the node canisters fail to become active after they restart
| when the enclosure is updated, check their status by using the service
| assistant. If both node canisters show critical node error 508, use the
| service assistant to restart the nodes. For any other node error, see
| “Procedure: Fixing node errors” on page 56. To restart a node from the
| service assistant, perform the following steps:
| 1) Log on to the service assistant.
| 2) From the home page, select the node that you want to restart from the
| Changed Node List.
| 3) Select Actions > Restart.
| d. The system starts and can handle I/O requests from the host systems.
| Note: The configuration changes that are described in the following steps
| must be performed to ensure that the system is operating correctly. If you
| do not perform these steps, the system is unable to report certain errors.
| 26. Start the management GUI and select Monitoring > System Details. You see
| an additional enclosure in the system list because the system has detected the
| replacement control enclosure. The original control enclosure is still listed in
| its configuration. The original enclosure is listed with its original enclosure ID.
| It is offline and managed. The new enclosure has a new enclosure ID. It is
| online and unmanaged.
27. Select the original enclosure in the tree view.
Verify that it is offline and managed and that the serial number is correct.
Note: Ensure that you know the type of enclosure chassis that you are replacing.
The procedures for replacing an expansion enclosure chassis are different from
those procedures for replacing a control enclosure chassis. For information about
replacing a control enclosure chassis, see “Replacing a control enclosure chassis”
on page 100.
(D005)
Even though many of these procedures are hot-swappable, these procedures are
intended to be used only when your system is not up and running and performing
I/O operations. Unless your system is offline, go to the management GUI and
follow the fix procedures.
Be careful when you are replacing the hardware components that are located in the
back of the system that you do not inadvertently disturb or remove any cables that
you are not instructed to remove.
Ensure that you are aware of the procedures for handling static-sensitive devices
before you remove the enclosure.
Note: If your system is online, replacing an expansion enclosure can cause one or
more of your volumes to go offline or your quorum disks to be inaccessible. Before
you proceed with these procedures, verify which volumes might go offline. From
the management GUI, go to Home > Manage Devices. Select the enclosure that
you want to replace. Then select Show Dependent Volumes in the Actions menu.
Note: The cables must be inserted back into the same ports from which they
were removed on the old enclosure; otherwise, the system cannot function
properly.
19. Attach the power cords and the cable retention brackets to the power supply
units.
20. Write the old enclosure machine type and model (MTM) and serial number on
the repair identification (RID) tag that is supplied. Attach the tag to the left
flange at the back of the enclosure.
21. Turn on the power to the enclosure by using the switches on the power
supply units.
The system records an error that indicates that an enclosure FRU replacement was
detected. Go to the management GUI to use the fix procedure to change the
machine type and model and serial number in the expansion enclosure.
2
svc00650
4. Working from the front of the rack cabinet, remove the clamping screw from
the rail assembly on both sides of the rack cabinet.
Note: Ensure that the rail is level between the front and the back.
13. Insert the rail location pins through the holes in the rack cabinet.
14. Insert a clamping screw into the upper mounting hole between the rail
location pins.
15. Tighten the screw to secure the rail to the rack from the back side.
16. Repeat the steps to secure the opposite rail to the rack cabinet.
The following methods are used to notify you and the IBM Support Center of a
new event:
v If you enabled Simple Network Management Protocol (SNMP), an SNMP trap is
sent to an SNMP manager that is configured by the customer.
v If enabled, log messages can be forwarded from a sender to a receiver on an IP
network by using the syslog protocol.
v If enabled, event notifications can be forwarded from a sender to a receiver
through Call Home email.
| v If Call Home is enabled, critical faults generate a problem management record
| (PMR) that is sent directly to the appropriate IBM Support Center.
Understanding events
When a significant change in status is detected, an event is logged in the event log.
Error data
| You can view the event log by using the Monitoring > Events options in the
| management GUI. The event log contains many entries. You can, however, select
only the type of information that you need.
You can also view the event log by using the command-line interface (lseventlog).
See the “Command-line interface” topic for the command details.
Table 24 describes some of the fields that are available to assist you in diagnosing
problems.
Table 24. Description of data fields for the event log
Data field Description
Event ID This number precisely identifies why the event was logged.
Error code This number describes the service action that should be followed to
resolve an error condition. Not all events have error codes that are
associated with them. Many event IDs can have the same error code
because the service action is the same for all the events.
Sequence number A number that identifies the event.
Event count The number of events coalesced into this event log record.
Object type The object type to which the event log relates.
Object ID A number that uniquely identifies the instance of the object.
Fixed When an alert is shown for an error condition, it indicates if the
reason for the event was resolved. In many cases, the system
automatically marks the events fixed when appropriate. There are
some events that must be manually marked as fixed. If the event is a
message, this field indicates that you have read and performed the
action. The message must be marked as read.
First time The time when this error event was reported. If events of a similar
type are being coalesced together, so that one event log record
represents more than one event, this field is the time the first error
event was logged.
Last time The time when the last instance of this error event was recorded in the
log.
Root sequence If set, this number is the sequence number of an event that represents
number an error that probably caused this event to be reported. Resolve the
root event first.
Sense data Additional data that gives the details of the condition that caused the
event to be logged.
Event notifications
Storwize V7000 can use Simple Network Management Protocol (SNMP) traps,
syslog messages, and Call Home email to notify you and the IBM Support Center
when significant events are detected. Any combination of these notification
methods can be used simultaneously. Notifications are normally sent immediately
after an event is raised. However, there are some events that might occur because
Each event that Storwize V7000 detects is assigned a notification type of Error,
Warning, or Information. When you configure notifications, you specify where the
notifications should be sent and which notification types are sent to that recipient.
Power-on self-test
When you turn on the system, the node canisters perform self-tests.
If a critical failure is detected during the POST, the software is not loaded and the
fault LED is illuminated. To determine if there is a POST error on a canister, go to
“Procedure: Understanding the system status using the LEDs” on page 49.
When the software is loaded, additional testing takes place, which ensures that all
of the required hardware and software components are installed and functioning
correctly.
Error codes help you to identify the cause of a problem, the failing
field-replaceable units (FRUs), and the service actions that might be needed to
solve the problem.
Event IDs
The Storwize V7000 software generates events, such as informational events and
error events. An event ID or number is associated with the event and indicates the
reason for the event.
Error events are generated when a service action is required. An error event maps
to an alert with an associated error code. Depending on the configuration, error
events can be notified through email, SNMP, or syslog.
Informational events
The informational events provide information about the status of an operation.
Informational events are recorded in the event log and, depending on the
configuration, can be notified through email, SNMP, or syslog.
Because node errors are specific to a node, for example, memory has failed, the
errors are only reported on that node. However, some of the conditions that the
node detects relate to the shared components of the enclosure. In these cases both
node canisters in the enclosure report the error.
There are two types of node errors: critical node errors and noncritical node errors.
Critical errors
A critical error means that the node is not able to participate in a clustered system
until the issue that is preventing it from joining a clustered system is resolved. This
error occurs because part of the hardware has failed or the system detects that the
software is corrupt. If it is possible to communicate with the canister with a node
error, an alert that describes the error is logged in the event log. If the system
cannot communicate with the node canister, a Node missing alert is reported. If a
node has a critical node error, it is in service state, and the fault LED on the node
is on. The exception is when the node cannot connect to enough resources to form
a clustered system. It shows a critical node error but is in the starting state. The
range of errors that are reserved for critical errors are 500 - 699.
Noncritical errors
A noncritical error code is logged when there is a hardware or software failure that
is related to just one specific node. These errors do not stop the node from entering
active state and joining a clustered system. If the node is part of a clustered
system, there is also an alert that describes the error condition. The node error is
shown to make it clear which of the node canisters the alert refers to. The range of
errors that are reserved for noncritical errors are 800 - 899.
You must perform software problem analysis before you can perform further
operations to avoid the possibility of corrupting your configuration.
Node errors
1. Review the saved location information of the node
500 Incorrect enclosure
canister and the status of the other node canister in
Explanation: The node canister has saved cluster the enclosure (the partner canister). If the node
information, which indicates that the canister is now canister has been inadvertently swapped, the other
located in a different enclosure chassis from where it node canister will have the same error.
was previously used. Using the node canister in this 2. If the canisters have been swapped, use the
state might corrupt the data held on the enclosure hardware remove and replace canister procedure to
drives. swap the canisters. The system should start.
User response: Follow troubleshooting procedures to 3. If the partner canister is in candidate state, use the
relocate the nodes to the correct location. hardware remove and replace canister procedure to
1. Review the saved location information of the node swap the canisters. The system should start.
canister and the status of the other node canister in 4. If the partner canister is in active state, it is running
the enclosure (the partner canister). Determine the cluster on this enclosure and has replaced the
whether the enclosure is part of an active system original use of this canister. You must follow the
with volumes that contain required data. procedure to remove cluster data from this node
2. If you have unintentionally moved the canister into canister. The node canister will then become active
this enclosure, move the canister back to its original in the cluster in its current slot.
location, and put the original canister back in this 5. If the partner canister is in service state, review its
enclosure. Follow the hardware remove and replace node error to determine the correct action.
canister procedures. Generally, you will fix the errors reported on the
3. If you have intentionally moved the node canister partner node in priority order, and review the
into this enclosure, check whether it is safe to situation again after each change. If you have to
continue or whether a loss of data can result. Do replace the partner canister with a new one you
not continue if any of the following conditions should move this canister back to the correct
apply, contact IBM technical support instead: location at the same time.
a. You require the volume data on the system from
Possible Cause-FRUs or other:
which the node canister was removed, and that
system is not running with two online nodes. v None
b. You require the volume data on this system, and
the partner node is not online. 502 No enclosure identity exists and a status
4. If you have determined that the node canister can from the partner node could not be
be used in this location, follow procedures to obtained.
remove cluster data from the node canister. Explanation: The enclosure has been replaced and
communication with the other node canister (partner
Possible Cause-FRUs or other: node) in the enclosure is not possible. The partner node
v None could be missing, powered off or unable to boot, or an
internode communication failure may exist.
501 Incorrect slot User response: Follow troubleshooting procedures to
fix the hardware:
Explanation: The node canister has saved cluster
information, which indicates that the canister is now 1. Follow the service procedures to get the partner
located in the expected enclosure chassis, but in a node started. An error will still exist because the
different slot from where it was previously used. Using enclosure has no identity. If the error has changed,
the node canister in this state might mean that hosts follow the service procedure for that error.
are not able to connect correctly. 2. If the partner has started and is showing a location
error (probably this one), then the PCI link is
User response: Follow troubleshooting procedures to
probably broken. Since the enclosure chassis was
relocate the nodes to the correct location.
recently replaced, this is likely the problem. Obtain
a replacement enclosure chassis, and restart the relocate nodes to the correct location.
hardware remove and replace control enclosure 1. Review the saved location information of the node
chassis procedure. canister and the saved location information of the
3. If this action does not resolve the issue, contact IBM other node canister in the enclosure (the partner
technical support. They will work with you to canister). The correct node canister is the one that
ensure that the cluster state data is not lost while comes from the enclosure chassis that is being
resolving the problem. replaced. The drives now in this enclosure should
also come from that enclosure.
Possible Cause-FRUs or other: 2. Decide what to do with the node canister that did
v Enclosure chassis (100%) not come from the enclosure that is being replaced.
a. If the other node canister from the enclosure
503 Incorrect enclosure type being replaced is available, use the hardware
remove and replace canister procedures to
Explanation: The node canister has been moved to an remove the incorrect canister and replace it with
expansion enclosure. A node canister will not operate the second node canister from the enclosure
in this environment. being replaced. The two node canister should
show node error 504 and the actions for that
User response: Follow troubleshooting procedures to
error should be followed.
relocate the nodes to the correct location.
b. If the other node canister from the enclosure
1. Review the saved location information of the node
being replaced is not available, check the
canister to determine which control enclosure the
enclosure of the node canister that did not come
node canister should be in.
from the replaced enclosure. Do not use this
2. Use the hardware remove and replace canister canister in this enclosure if you require the
procedures to move the node canister to the correct volume data on the system from which the node
location and move the expansion canister, that is canister was removed, and that system is not
probably in that location, to here. If there is a node running with two online nodes. You should
canister that is in active state where this node return the canister to its original enclosure and
canister should be, do not replace that node canister use a different canister in this enclosure.
with this one.
c. When you have checked it is not required
elsewhere, follow the procedure to remove
504 No enclosure identity and partner node cluster data from the node canister that did not
matches. come from the enclosure that is being replaced.
Restart both nodes. Expect node error 506 to
Explanation: The enclosure vital product data
now be reported and follow the service
indicates that the enclosure chassis has been replaced.
procedures for that error.
This node canister and the other node canister in the
enclosure were previously operating in the same
Possible Cause-FRUs or other:
enclosure chassis.
v None
User response: Follow troubleshooting procedures to
configure the enclosure.
506 No enclosure identity and no node state
1. This is an expected situation during the hardware
on partner
remove and replace procedure for a control
enclosure chassis. Continue following the remove Explanation: The enclosure vital product data
and replace procedure and configure the new indicates that the enclosure chassis has been replaced.
enclosure. There is no cluster state information on the other node
canister in the enclosure (the partner canister), so both
Possible Cause-FRUs or other: node canisters from the original enclosure have not
v None been moved to this one.
User response: Follow troubleshooting procedures to
505 No enclosure identity and partner has relocate nodes to the correct location:
cluster data that does not match. 1. Review the saved location information of the node
canister and why the second node canister from the
Explanation: The enclosure vital product data
original enclosure was not moved into this
indicates that the enclosure chassis has been replaced.
enclosure.
This node canister and the other node canister in the
enclosure do not come from the same original 2. If you are sure that this node canister came from
enclosure. the enclosure that is being replaced, and the original
partner canister is available, use the hardware
User response: Follow troubleshooting procedures to remove and replace node canister procedure to
install the second node canister in this enclosure. Note: If both nodes are reporting this node error, the
Restart the node canister. The two node canister cause may be a damaged enclosure.
should show node error 504 and the actions for that Possible Cause-FRUs or other:
error should be followed. v Service procedure error (90%)
3. If you are sure this node canister came from the v Enclosure chassis (10%)
enclosure that is being replaced, and that the
original partner canister has failed, continue
following the remove and replace procedure for an 509 The enclosure identity cannot be read.
enclosure chassis and configure the new enclosure. Explanation: The canister was unable to read vital
product data (VPD) from the enclosure. The canister
Possible Cause-FRUs or other: requires this data to be able to initialize correctly.
v None
User response: Follow troubleshooting procedures to
fix the error:
507 No enclosure identity and no node state 1. Follow the node canister reseat procedure to reseat
Explanation: The node canister has been placed in a the canister.
replacement enclosure chassis. The node canister is also 2. If the error is still present after the canister has been
a replacement, or has had all cluster state removed reseated, proceed to the hardware troubleshooting
from it. procedures below.
User response: Follow troubleshooting procedures to a. Check errors reported on the other node canister
relocate the nodes to the correct location. in this enclosure (the partner canister).
1. Check the status of the other node in the enclosure. b. If it is reporting the same error, follow the
It should show node error 506. Unless it also shows hardware remove and replace procedure to
error 507, use these procedures to resolve the errors replace the enclosure chassis.
on the other node. c. If the partner canister is not reporting this error,
2. If the other node in the enclosure is also reporting follow the hardware remove and replace
507, the enclosure and both node canisters have no procedure to replace this canister.
state information. You should contact IBM technical
support. They will assist you in setting the Note: If a newly installed system has this error on both
enclosure vital product data and running cluster node canister, the data that needs to be written to the
recovery. enclosure will not be available on the canisters, you
should contact IBM technical support for the WWNNs
Possible Cause-FRUs or other: to use.
Possible Cause-FRUs or other:
v None
v Node canister (50%)
v Enclosure chassis (50%)
508 Cluster identifier is different between
enclosure and node
510 The detected memory size does not
Explanation: The node canister location information
match the expected memory size.
shows it is in the correct enclosure, however the
enclosure has had a new cluster created on it since the Explanation: The amount of memory detected in the
node was last shut down. Therefore, the cluster state node canister is less than the expected memory. The
data stored on the node is not valid. error code date shows the detected memory, in MB,
followed by the expected memory, in MB.
User response: Follow troubleshooting procedures to
relocate the nodes to the correct location. User response: Follow troubleshooting procedures to
1. Check whether a new cluster has been created on fix the hardware:
this enclosure while this canister was not operating 1. Use the hardware remove and replace node canister
or whether the node canister was recently installed procedure to install a new node canister.
in the enclosure.
Possible Cause-FRUs or other:
2. If this node canister is the one to be used in this
v Node canister (100%)
enclosure, you should follow the procedure to
remove cluster data from the node canister. It will
then join the cluster. 523 The internal disk file system is
3. If this is not the node canister that you intended to damaged.
use, follow the hardware remove and replace Explanation: The node startup procedures have found
canister procedure to replace the node canister with problems with the file system on the internal disk of
the intended one. the node.
Possible Cause-FRUs or other: 2. Check all enclosures in the system are powered on
and that the SAS cabling between the enclosures
v Environment issue (100%) has not been disturbed. If any wiring changes have
been made check all cables are securely connected
535 Canister internal PCIe switch failed and that the cabling rules have been followed.
3. If a quorum drive in a system enclosure is shown as
Explanation: The PCI Express switch has failed or
missing, find the drive and check that it is working.
cannot be detected. In this situation, the only
The drive may have been moved from the location
connectivity to the node canister is through the
shown, in that case find the drive and ensure it is
Ethernet ports.
installed and working. If the drive is not located in
User response: Follow troubleshooting procedures to the control enclosure, try moving it to the control
fix the hardware: enclosure, a problem in SAS connectivity may be
1. Follow the procedures to reseat the node canister. the issue.
2. If reseating the canister does not resolve the
Note: If you are able to reestablish the systems
situation, follow the hardware remove and replace
operation you will be able to use the extra
node canister procedures to replace the canister.
diagnostics the system provides to diagnose the node canister should have been written down. If
problems on SAS cables and expansion enclosures. the correct WWNN cannot be determined contact
4. If a quorum disk on an external storage system is your support center for assistance.
shown as missing, find the storage control and 4. Use the service assistant to modify the incorrect
confirm that the LUN is available, check the Fibre WWNN. If it is the node showing the error that
Channel connections between the storage controller should be modified, this can safely be done
and the 2076 are working and that any changes immediately. If it is an active node that should be
made to the SAN configuration and zoning have modified, use caution because the node will restart
not effected the connectivity. Check the status of the when the WWNN is changed. If this node is the
Fibre Channel ports on the node and resolve any only operational node in an enclosure, access to the
issues. volumes that it is managing will be lost. You should
5. If all nodes have either node error 578 or 550, ensure that the host systems are in the correct state
attempt to reestablish a cluster by following the before you change the WWNN.
service procedures for the nodes showing node 5. If the node showing the error had the correct
error 578. If this is not successful, follow the cluster WWNN, it can be restarted, using the service
recovery procedures. assistant, after the node with the duplicate WWNN
is updated.
556 A duplicate WWNN has been detected. 6. If you are unable to find a Storwize V7000 node
canister with the same WWNN as the node canister
Explanation: The node canister has detected another showing the error, use the SAN monitoring tools to
device that has the same World Wide Node Name determine whether there is another device on the
(WWNN) on the Fibre Channel network. A WWNN is SAN with the same WWNN. This device should not
16 hexadecimal digits long. For a Storwize V7000, the be using a WWNN assigned to a Storwize V7000, so
first 11 digits are always 50050768020. The last 5 digits you should follow the service procedures for the
of the WWNN are given in the additional data of the device to change its WWNN. Once the duplicate
error. The Fibre Channel ports of the node canister are has been removed, restart the node canister.
disabled to prevent disruption of the Fibre Channel
network. One or both node canisters with the same Possible Cause-FRUs or other:
WWNN can show the error. Because of the way
v None
WWNNs are allocated, a device with a duplicate
WWNN is normally another Storwize V7000 node
canister. 562 The nodes hardware configuration does
not meet the minimum requirements.
User response: Follow troubleshooting procedures to
configure the WWNN of the node: Explanation: The node hardware is not at the
1. Find the Storwize V7000 node canister with the minimum specification for the node to become active in
same WWNN as the node canister reporting the a cluster. This may be because of hardware failure, but
error. The WWNN for a Storwize V7000 node is also possible after a service action has used an
canister can be found from the node Vital Product incorrect replacement part.
Data (VPD) or from the node canister details shown User response: Follow troubleshooting procedures to
by the service assistant. The node with the duplicate fix the hardware:
WWNN need not be part of the same cluster as the
node reporting the error; it could be remote from 1. It is not possible to service parts within the node
the node reporting the error on a part of the fabric canister. Reseat the existing node canister to see
connected through an inter-switch link. The two whether the problem fixes. If it does not, use the
node canisters within a control enclosure must have hardware node canister remove and replace
different WWNNs. The WWNN of the node canister procedures to change the node canister.
is stored within the enclosure chassis, so the
duplication is most likely caused by the Possible Cause-FRUs or other:
replacement of a control enclosure chassis. v Node canister (100%)
2. If a Storwize V7000 node canister with a duplicate
WWNN is found, determine whether it, or the node 564 Too many software crashes have
reporting the error, has the incorrect WWNN. occurred.
Generally, it is the node canister that has had its
enclosure chassis that was recently replaced or had Explanation: The node has been determined to be
its WWNN changed incorrectly. Also consider how unstable because of multiple resets. The cause of the
the SAN is zoned when making your decision. resets can be that the system encountered an
unexpected state or has executed instructions that were
3. Determine the correct WWNN for the node with the
not valid. The node has entered the service state so that
incorrect WWNN. If the enclosure chassis has been
diagnostic data can be recovered.
replaced as part of a service action, the WWNN for
Explanation: Parts of the node software package are User response: Follow troubleshooting procedures to
receiving unexpected results; there may be an correct connectivity issues between the cluster nodes
inconsistent set of subpackages installed, or one and the quorum devices.
subpackage may be damaged. 1. Manual intervention is required once the node
User response: Follow troubleshooting procedures to reports this error.
reload the software. 2. Attempt to reestablish the cluster using other nodes.
1. Follow the procedure to run a node rescue. This may involve fixing hardware issues on other
nodes or fixing connectivity issues between nodes.
2. If the error occurs again, contact IBM technical
support. 3. If you are able to reestablish the cluster, remove the
cluster data from the node showing 578 so it goes to
Possible Cause-FRUs or other: candidate state, it will then be automatically added
v None back to the cluster. If the node does not
automatically add back to the cluster, note the name
and I/O group of the node, then delete the node
from the cluster configuration (if this has not
already happened) and then add the node back to
the cluster using the same name and I/O group.
4. If all nodes have either node error 578 or 550, Possible Cause-FRUs or other:
follow the cluster recovery procedures. v None
5. Attempt to determine what caused the nodes to
User response:
shut down.
1. Wait for the node to automatically fix the error
Possible Cause-FRUs or other: when sufficient charge becomes available.
v None 2. If possible, determine why one battery is not
charging. Use the battery status shown in the node
canister hardware details and the indicator LEDs on
671 The available battery charge is not the PSUs in the enclosure to diagnose the problem.
enough to allow the node canister to If the issue cannot be resolved, wait until the cluster
start. Two batteries are charging. is operational and use the troubleshooting options
Explanation: The battery charge within the enclosure in the management GUI to assist in resolving the
is not sufficient for the node to safely become active in issue.
a cluster. The node will not start until sufficient charge
exists to store the state and configuration data held in Possible Cause-FRUs or other:
the node canister memory if power were to fail. Two v Battery (33%)
batteries are within the enclosure, one in each of the v Control power supply (33%)
power supplies. Neither of the batteries indicate an
v Power cord (33%)
error–both are charging.
The node will start automatically when sufficient
673 The available battery charge is not
charge is available. The batteries do not have to be
enough to allow the node canister to
fully charged before the nodes can become active.
start. No batteries are charging.
Both nodes within the enclosure share the battery
Explanation: A node cannot be in active state if it
charge, so both node canisters report this error.
does not have sufficient battery power to store
The service assistant shows the estimated start time in configuration and cache data from memory to internal
the node canister hardware details. disk after a power failure. The system has determined
that both batteries have failed or are missing. The
Possible Cause-FRUs or other:
problem with the batteries must be resolved to allow
v None the system to start.
User response: Wait for the node to automatically fix User response: Follow troubleshooting procedures to
the error when sufficient charge becomes available. fix hardware:
Possible Cause-FRUs or other: 1. Resolve problems in both batteries by following the
v None procedure to determine status using the LEDs.
2. If the LEDs do not show a fault on the power
supplies or batteries, power off both power supplies
672 The available battery charge is not in the enclosure and remove the power cords. Wait
enough to allow the node canister to 20 seconds, then replace the power cords and
start. One battery is charging. restore power to both power supplies. If both node
Explanation: The battery charge within the enclosure canisters continue to report this error replace the
is not sufficient for the node to safely become active in enclosure chassis.
a cluster. The node will not start until sufficient charge
exists to store the state and configuration data held in Possible Cause-FRUs or other:
the node canister memory if power were to fail. Two v Battery (33%)
batteries are within the enclosure, one in each of the
v Power supply (33%)
power supplies. Only one of the batteries is charging,
so the time to reach sufficient charge will be extended. v Power cord (33%)
v Enclosure chassis (1%)
The node will start automatically when sufficient
charge is available. The batteries do not have to be
fully charged before the nodes can become active. 690 The node is held in the service state.
Both nodes within the enclosure share the battery Explanation: The node is in service state and has been
charge, so both node canisters report this error. instructed to remain in service state. While in service
state, the node will not run as part of a cluster. A node
The service assistant shows the estimated start time,
must not be in service state for longer than necessary
and the battery status, in the node canister hardware
while the cluster is online because a loss of redundancy
details.
will result. A node can be set to remain in service state
either because of a service assistant user action or other node canister in the enclosure and then use
because the node was deleted from the cluster. the management GUI to resolve the problem.
User response: When it is no longer necessary to hold 3. Use a USB key to get the status of the node and
the node in the service state, exit the service state to check whether it is the configuration node. If the
allow the node to run: node showing this error is the configuration node,
use a USB key or the service assistant to get the
1. Use the service assistant action to release the service
status of the other node in the enclosure. If the
state.
other node is active, run the service assistant on the
other node, but make this node the active node.
Possible Cause-FRUs or other: Check whether the active node is reporting node
v None error 805. Use the service assistant to hold the node
in service state, which will cause the other node in
the enclosure to become the configuration node.
801 Memory reduced.
Once you have set the node canister into service
Explanation: Memory is reduced but sufficient state, you should immediately release the service
memory exists to run I/O operations. state so that the node becomes active in the cluster.
Now use the troubleshooting fix procedures in the
User response: Follow troubleshooting procedures to
management GUI to correct the associated cluster
fix the hardware.
error.
Features
This list includes the major accessibility features in the management GUI:
v You can use screen-reader software and a digital speech synthesizer to hear what
is displayed on the screen. The following screen reader has been tested: JAWS
11.
v Most of the GUI features are accessible by using the keyboard. For those features
that are not accessible, equivalent function is available by using the
command-line interface (CLI).
Navigating by keyboard
You can use keys or key combinations to perform operations and initiate many
menu actions that can also be done through mouse actions. You can navigate the
management GUI and help system from the keyboard by using the following key
combinations:
v To navigate between different GUI panels, select the Low-graphics mode option
on the GUI login panel. You can use this option to navigate to all the panels
without manually typing the web addresses.
v To go to the next frame, press Ctrl+Tab.
v To move to the previous frame, press Shift+Ctrl+Tab.
v To navigate to the next link, button, or topic within a panel, press Tab inside a
frame (page).
v To move to the previous link, button, or topic within a panel, press Shift+Tab.
v To select GUI objects, press Enter.
v To print the current page or active frame, press Ctrl+P.
v To expand a tree node, press the Right Arrow key. To collapse a tree node, press
the Left Arrow key.
v To scroll all the way up, press Home; to scroll all the way down, press End.
v To go back, press Alt+Left Arrow key.
v To go forward, press Alt+Right Arrow key.
v For actions menus:
– Press Tab to navigate to the grid header.
– Press the Left or Right Arrow keys to reach the drop-down field.
– Press Enter to open the drop-down menu.
– Press the Up or Down Arrow keys to select the menu items.
– Press Enter to launch the action.
v For filter panes:
– Press Tab to navigate to the filter panes.
– Press the Up or Down Arrow keys to change the filter or navigation for
nonselection.
You can find the HTML version of the IBM Storwize V7000 information at the
following website:
publib.boulder.ibm.com/infocenter/storwize/ic/index.jsp
You can access this information using screen-reader software and a digital speech
synthesizer to hear what is displayed on the screen. The information was tested
using the following screen reader: JAWS Version 10 or later.
IBM may not offer the products, services, or features discussed in this document in
other countries. Consult your local IBM representative for information on the
products and services currently available in your area. Any reference to an IBM
product, program, or service is not intended to state or imply that only that IBM
product, program, or service may be used. Any functionally equivalent product,
program, or service that does not infringe any IBM intellectual property right may
be used instead. However, it is the user's responsibility to evaluate and verify the
operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter
described in this document. The furnishing of this document does not give you
any license to these patents. You can send license inquiries, in writing, to:
For license inquiries regarding double-byte (DBCS) information, contact the IBM
Intellectual Property Department in your country or send inquiries, in writing, to:
The following paragraph does not apply to the United Kingdom or any other
country where such provisions are inconsistent with local law:
INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS
PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER
EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS
FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of express or
implied warranties in certain transactions, therefore, this statement may not apply
to you.
Licensees of this program who wish to have information about it for the purpose
of enabling: (i) the exchange of information between independently created
programs and other programs (including this one) and (ii) the mutual use of the
information which has been exchanged, should contact:
IBM Corporation
Almaden Research
650 Harry Road
Bldg 80, D3-304, Department 277
San Jose, CA 95120-6099
U.S.A.
The licensed program described in this document and all licensed material
available for it are provided by IBM under terms of the IBM Customer Agreement,
IBM International Program License Agreement or any equivalent agreement
between us.
All statements regarding IBM's future direction or intent are subject to change or
withdrawal without notice, and represent goals and objectives only.
This information is for planning purposes only. The information herein is subject to
change before the products described become available.
This information contains examples of data and reports used in daily business
operations. To illustrate them as completely as possible, the examples include the
names of individuals, companies, brands, and products. All of these names are
fictitious and any similarity to the names and addresses used by an actual business
enterprise is entirely coincidental.
COPYRIGHT LICENSE:
If you are viewing this information softcopy, the photographs and color
illustrations may not appear.
Trademarks
IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of
International Business Machines Corp., registered in many jurisdictions worldwide.
Other product and service names might be trademarks of IBM or other companies.
A current list of IBM trademarks is available on the web at Copyright and
trademark information at www.ibm.com/legal/copytrade.shtml.
Adobe and the Adobe logo are either registered trademarks or trademarks of
Adobe Systems Incorporated in the United States, and/or other countries.
Intel, Intel logo, Intel Xeon, and Pentium are trademarks or registered trademarks
of Intel Corporation or its subsidiaries in the United States and other countries.
Microsoft, Windows, Windows NT, and the Windows logo are trademarks of
Microsoft Corporation in the United States, other countries, or both.
UNIX is a registered trademark of The Open Group in the United States and other
countries.
Java and all Java-based trademarks and logos are trademarks or registered
trademarks of Oracle and/or its affiliates.
Other product and service names might be trademarks of IBM or other companies.
This equipment has been tested and found to comply with the limits for a Class A
digital device, pursuant to Part 15 of the FCC Rules. These limits are designed to
provide reasonable protection against harmful interference when the equipment is
operated in a commercial environment. This equipment generates, uses, and can
radiate radio frequency energy and, if not installed and used in accordance with
the instruction manual, might cause harmful interference to radio communications.
Notices 145
Operation of this equipment in a residential area is likely to cause harmful
interference, in which case the user will be required to correct the interference at
his own expense.
Properly shielded and grounded cables and connectors must be used in order to
meet FCC emission limits. IBM is not responsible for any radio or television
interference caused by using other than recommended cables and connectors, or by
unauthorized changes or modifications to this equipment. Unauthorized changes
or modifications could void the user's authority to operate the equipment.
This device complies with Part 15 of the FCC Rules. Operation is subject to the
following two conditions: (1) this device might not cause harmful interference, and
(2) this device must accept any interference received, including interference that
might cause undesired operation.
Responsible Manufacturer:
"Warnung: Dieses ist eine Einrichtung der Klasse A. Diese Einrichtung kann im
Wohnbereich Funk-Störungen verursachen; in diesem Fall kann vom Betreiber
verlangt werden, angemessene Mabnahmen zu ergreifen und dafür
aufzukommen."
Dieses Gerät ist berechtigt, in übereinstimmung mit dem Deutschen EMVG das
EG-Konformitätszeichen - CE - zu führen.
Verantwortlich für die Einhaltung der EMV Vorschriften ist der Hersteller:
International Business Machines Corp.
New Orchard Road
Armonk,New York 10504
Tel: 914-499-1900
IBM Deutschland
Technical Regulations, Department M456
IBM-Allee 1, 71137 Ehningen, Germany
Tel: +49 7032 15-2937
e-mail: mailto: tjahn @ de.ibm.com
Notices 147
Generelle Informationen: Das Gerät erfüllt die
Schutzanforderungen nach EN 55024 und EN 55022 Klasse A.
rusemi
Taiwan Class A compliance statement
Notices 149
150 Storwize V7000: Troubleshooting, Recovery, and Maintenance Guide
Index
Numerics blank carrier
replacing 96, 98
control enclosure
detection error 45
2.5" drive assembly browsers power supply unit 6
replacing 98 supported 44 control enclosure chassis
3.5" drive assembly replacing 100
replacing 96 create cluster command 38
C critical
node errors 130
A cable retention bracket
releasing 96
Customer replaced 109
about this document Canadian electronic emission notice 146
sending comments xvi canister
accessibility expanison 81 D
keyboard 141 identification 47 deleting
repeat rate node 79 backup configuration files
up and down buttons 141 replacing 79, 81 using the CLI 77
shortcut keys 141 changing system 56
accessing service IP address 57 detection error
canisters CLI control location 45
Ethernet cable 59 cluster (system) commands 33 expansion location 45
cluster (system) CLI 33 service commands 33 determining
management GUI 30 cluster (system) CLI SAN problem 62
publications 141 accessing 33 Deutschsprachiger EU Hinweis 147
service assistant 32, 61 when to use 33 documentation
service CLI 33 cluster (system) commands improvement xvi
actions CLI 33 drive characteristics
reset service IP address 35 clustered storage system best practices 21
reset superuser password 35 failure to create 42 drives
alerts clustered system 2.5-inch drives 2
best practices 21 initializing 3.5-inch drives 2
apply software command 37 with service assistant 59 LED indicator 2
with USB key 58
restore 70
B clustered systems E
backing up error codes 130
electronic emission notices
best practices 21 recovery codes 130
Avis de conformité à la
system configuration files 72 restore 66
réglementation d'Industrie
backup configuration files T3 recovery 66
Canada 146
deleting codes
Deutschsprachiger EU Hinweis 147
using the CLI 77 node error
European Union (EU) 146
restoring 74 critical 130
Federal Communications Commission
bad blocks 27 noncritical 130
(FCC) 145
battery commands
French Canadian 146
maintenance 23, 24 apply software 37
Germany 147
removing 93 create cluster 38
Industry Canada 146
best practices query status 38
International Electrotechnical
alerts 21 reset service assistant password 36
Commission (IEC) 148
backing up data 21 satask.txt 35
Japanese Voluntary Control Council
drive characteristics 21 snap 37
for Interference (VCCI) 148
inventory reporting 21 svcconfig backup 72
Korean 148
IP address 19 svcconfig restore 74
New Zealand 146
notifications 20 comments, sending xvi
People's Republic of China 148
passwords 19 components
Taiwan 149
power management 20 enclosure 2
United Kingdom 148
RAID 21 illustration 5
EMC statement, People's Republic of
record end cap
China 148
location information 22 indicators 4
enclosure
subscribe hardware 1
components 5
notifications 22 contact information
identification 47
troubleshooting 19 European 149
enclosure end cap
warranty agreement Taiwan 149
replacing 99
maintenance agreement 22
Index 153
U
understanding
clustered-system recovery codes 130
error codes 116
event log 113
United Kingdom electronic emission
notice 148
USB key
detection error 46
using 34
when to use 34
USB ports 11
using
GUI interfaces 29
initialization tool 35
initialization tool interface 34
management GUI 29
service assistant 31
USB key 34
V
VDisks (volumes)
recovering from offline
using CLI 70
viewing
event log 113
node canister
status 48, 49
system
status 48, 49
volumes (VDisks)
recovering from offline
using CLI 70
W
warranty agreement
best practices 22
when to use
cluster (system) CLI 33
management GUI interface 30
service assistant 31
service CLI 33
USB key 34
worldwide port names (WWPNs)
description 10
Printed in USA
GC27-2291-02