Version 3.2
Installation and User's Guide
Notes, Cautions, and Warnings
NOTE: A NOTE indicates important information that helps you make better use of your computer.
CAUTION: A CAUTION indicates either potential damage to hardware or loss of data and tells you
how to avoid the problem.
WARNING: A WARNING indicates a potential for property damage, personal injury, or death.
Copyright 2016 Dell Inc. All rights reserved. This product is protected by U.S. and international copyright and
intellectual property laws. Dell and the Dell logo are trademarks of Dell Inc. in the United States and/or other
jurisdictions. All other marks and names mentioned herein may be trademarks of their respective companies.
2016 - 04
Rev. 110-6235-EN-R1
Contents
Preface....................................................................................................................... 9
Audience................................................................................................................................................ 9
Related Documentation........................................................................................................................9
Dell Online Services.............................................................................................................................. 9
Dell EqualLogic Storage Solutions........................................................................................................9
Technical Support and Customer Service .........................................................................................10
Contacting Dell....................................................................................................................................10
3
3 Get Started with SAN HQ.................................................................................. 30
SAN HQ Client GUI Navigation...........................................................................................................30
All Groups Summary Information................................................................................................. 33
Group Information Categories..................................................................................................... 38
Context Link Bar Options..............................................................................................................49
95th Percentile Reporting............................................................................................................. 51
Display Data from Different Times............................................................................................... 53
Missing Data Points in Graphs...................................................................................................... 56
Identify Volumes............................................................................................................................56
Add a Group for Monitoring................................................................................................................57
SAN Headquarters Prerequisites for Adding a Group.................................................................. 57
Add a Group from the SAN Headquarters GUI.............................................................................57
Add a Group to the Monitoring List from the Command Line....................................................58
Manage Server Monitoring................................................................................................................. 60
Add a Server................................................................................................................................... 61
Change the Default Server............................................................................................................ 61
Remove a Server............................................................................................................................ 61
Upgrade the Client Monitoring Multiple Servers..........................................................................62
Verify and Modify Access to Log Files.......................................................................................... 63
4
Launch Group Manager with Single Sign-On.................................................................................... 73
Single Sign-On Requirements.......................................................................................................74
Enable Single Sign-On and Modify Login Credentials.................................................................74
Disable Single Sign-On for a Group............................................................................................. 75
Delete Login Credentials for a Group...........................................................................................75
Launch Dell Storage Update Manager from SAN HQ........................................................................75
Dell Storage Update Manager Overview and Prerequisites.........................................................76
Options for Launching Dell Storage Update Manager.................................................................76
Launch Dell Storage Update Manager for Standalone Firmware Updates................................. 77
Add and Manage Favorite Views......................................................................................................... 77
Add a View to the Favorites List.................................................................................................... 78
Set a View As the Home Page.......................................................................................................78
Monitor Synchronous Replication Volumes...................................................................................... 78
About Synchronous Replication................................................................................................... 79
Display SyncRep Volumes in SAN Headquarters..........................................................................79
Live View Data..................................................................................................................................... 81
Prerequisites for Establishing a Live View Session....................................................................... 82
Run a Live View Session................................................................................................................ 82
RAID Policy Evaluation and Reliability................................................................................................84
Use the RAID Evaluator................................................................................................................. 84
RAID Policy Reliability Reporting.................................................................................................. 85
SSD Usage Statistics............................................................................................................................ 92
View SSD Usage Statistics............................................................................................................. 92
View Recoverable Volumes ............................................................................................................... 98
About the PS Series Firmware Volume Undelete Feature........................................................... 98
Display Recoverable Volumes in SAN Headquarters................................................................... 99
Space Borrowing...............................................................................................................................100
Snapshot Space Borrowing Overview........................................................................................ 101
Space Borrowing Statistics.......................................................................................................... 101
Snapshot and Replica Compression................................................................................................ 103
Compression Overview...............................................................................................................103
Prerequisites and Considerations .............................................................................................. 103
SAN HQ Compression Reporting............................................................................................... 104
Syslog Event Logging........................................................................................................................ 107
Configure Syslog Event Logging for a Group............................................................................ 107
Disable Syslog Event Logging for a Group.................................................................................108
Change the Syslog Configuration to Use Specific Interfaces................................................... 108
Disable the SAN Headquarters Syslog Server.............................................................................109
SNMP Notifications........................................................................................................................... 109
About SNMP Notifications...........................................................................................................110
SAN HQ SNMP Notification Configuration.................................................................................110
SNMP Notifications Configuration File....................................................................................... 110
5
Requirements for SNMP Notifications.........................................................................................111
Configure SNMP Notifications.....................................................................................................111
Test and Troubleshoot SNMP Notifications............................................................................... 114
SAN HQ Support for VMware Virtual Volumes.................................................................................114
Storage Container Information................................................................................................... 114
Virtual Volume Information......................................................................................................... 116
Virtual Machine Information........................................................................................................ 117
SAN HQ Support for Delegated Space in Multiple Pools.................................................................118
Delegated Space in Capacity of Group View............................................................................. 119
Manage Group Network Address Changes......................................................................................120
Diagnose and Solve Monitoring Problems....................................................................................... 121
Additional Group Monitoring Concepts........................................................................................... 121
How Data Is Compressed in Log Files........................................................................................ 121
How Group Performance Affects SNMP Polling........................................................................122
How Your Group Monitoring Environment Affects TCP Retransmission Reporting................ 123
Dependency on Software and Firmware Versions.....................................................................125
6
Copy Audit Logs to the Clipboard.............................................................................................. 149
7
Data Collected By SupportAssist................................................................................................ 192
SupportAssist Requirements....................................................................................................... 192
Configure SupportAssist................................................................................................................... 194
Prerequisites for Configuring SupportAssist.............................................................................. 194
Configuring SupportAssist When Adding a New Group............................................................ 197
Configuring SupportAssist for an Existing Group...................................................................... 203
Configure SupportAssist From the CLI.......................................................................................207
Modify the SupportAssist Configuration..........................................................................................208
SupportAssist Configuration Settings........................................................................................ 208
SupportAssist Activity Log........................................................................................................... 212
Modify SupportAssist Data Collection Settings for a Group......................................................213
Modifying SupportAssist Syslog Settings for a Group................................................................214
Update or Configure Contact Data............................................................................................ 214
Enable and Disable Weekly Data Collection.............................................................................. 215
Enable and Disable Automatic Upload for a Group................................................................... 215
Delete the SupportAssist Configuration for a Group.................................................................216
Use SupportAssist..............................................................................................................................216
On-Demand Data Collection..................................................................................................... 216
Decrypt Local Data Packages......................................................................................................217
Procedure for Decrypting One or More Local Data Packages..................................................218
Display Support Case Information..............................................................................................218
Monitor and Troubleshoot SupportAssist........................................................................................ 219
General Troubleshooting........................................................................................................... 220
Considerations When Enabling SupportAssist...........................................................................220
Performance Impact When Using SupportAssist...................................................................... 220
SupportAssist Alerts.................................................................................................................... 220
Display Groups Configured for SupportAssist............................................................................221
Offline Data Uploads Using SupportAssist................................................................................. 221
Use a Different Server to Upload Diagnostic Information........................................................ 224
SupportAssist Reference Information.............................................................................................. 225
Data Collection and Upload Settings......................................................................................... 225
Data Sorting Options...................................................................................................................225
SupportAssist Command Reference.......................................................................................... 226
8
Preface
Preface
Dell EqualLogic SAN Headquarters (SAN HQ) enables you to monitor multiple PS Series groups from a
single graphical interface. It gathers and formats performance data and other vital group information.
Analyzing the data might help you improve performance and more effectively allocate group resources.
Audience
This manual is designed for administrators responsible for installing SAN Headquarters and using it to
monitor PS Series groups.
Administrators are not required to have extensive network or storage system experience. However, it is
useful to understand:
Related Documentation
For detailed information about PS Series arrays, groups, volumes, array software, and host software, log in
to the support site at eqlsupport.dell.com.
9
Technical Support and Customer Service
Dell support service is available to answer your questions about PS Series arrays and FS Series appliances.
Contacting Dell
If you are a customer in the United States or Canada in need of technical support, call
1-800-945-3355.
If you are outside of the United States or Canada, visit eqlsupport.dell.com/.
If you have an Express Service Code, have it ready. The code helps the Dell automated support
telephone system direct your call more efficiently.
10
1
Overview of SAN Headquarters
SAN Headquarters (SAN HQ) enables you to monitor multiple PS Series groups from a single graphical
user interface (GUI). It gathers and formats performance data and other vital group information. Analyzing
the data can help you improve performance and more effectively allocate group resources.
SAN HQ does not disrupt access to group storage, or degrade performance on the hosts or groups.
SAN HQ ServerIssues SNMP requests to collect configuration, status, and performance data from
one or more PS Series groups and stores the information in log files. To keep you informed of
potential problems, the SAN HQ Server can send email notification of group alerts. The SAN HQ
Server also includes a syslog server. You can configure groups to log events to the syslog server.
NOTE: For the best performance, Dell recommends that you have only one SAN HQ Server
installation monitor a group. Do not have multiple servers monitoring the same group.
Log filesContain the data that the SAN HQ Server collects from a group. The SAN HQ Server
maintains one set of log files for each monitored group. Each set of log files can contain up to 1-
years worth of data. After a year, the SAN HQ Server overwrites the oldest data. You can put the log
files on a network-accessible resource to share the data with computers running the SAN HQ Server.
SAN HQ ClientProvides a graphical user interface (GUI) for managing the SAN HQ environment and
viewing data collected by one or several SAN HQ servers. A SAN HQ Client accesses the log files
maintained by the SAN HQ Server and formats the group data into tables and graphs. You can run the
SAN HQ Client on multiple computers.
SupportAssistProvides a diagnostics gathering facility configured and run from the SAN HQ Client to
collect diagnostics and configuration information from the PS Series arrays and upload it to Dell
Support for analysis. For more information, see How SupportAssist Works.
After you add a group, the SAN HQ Server issues regular SNMP requests to the group and collects
configuration, status, and performance data. By default, the polling period the time between
consecutive polling operations is two minutes.
NOTE: SNMP polling has no impact on group performance, because serving SNMP requests is not a
high priority for a group.
The SAN HQ Server stores the data in the group log files. Computers running the SAN HQ Client access
the log files and display the data in the SAN HQ GUI.
Figure 1. Single-Server Environment shows the general layout of a SAN HQ single-server environment.
In Figure 1. Single-Server Environment, a remote SAN HQ Client accesses the local SAN HQ client/server
via the network. The SAN HQ Server issues a series of SNMP requests (polls) to each group for
configuration, status, and performance information. When the first set of polls returns from a group, the
server stores this baseline information in the group data and log files for future reference. It issues
The SAN HQ Server also includes a Syslog Server to which a PS Series group can log events (see Syslog
Event Logging).
In Figure 2. Multi-Server Environment, a remote SAN HQ Client accesses two SAN HQ servers at different
sites. Both SAN HQ servers are monitoring groups on separate networks. The SAN HQ Client can access
the data and log files monitored by those servers, displaying the information in the client GUI.
The SAN HQ Server can monitor multiple groups. The server issues a series of polls to each group it
monitors for configuration, status, and performance information.
Either double-click the SAN HQ shortcut on the desktop, or click Start Programs EqualLogic
SAN Headquarters. You can also double-click SANHQClient.exe in the directory where you installed
SAN HQ. By default, the installation location is: C:\Program Files\EqualLogic\SAN HQ.
You can also start SAN HQ from a command line. See Start SAN Headquarters GUI Using a Command
Line.
When you start the GUI, the Servers and Groups window appears, showing all groups monitored by the
default server. The Servers and Groups window displays the status of each connected server and
monitored group and any active alerts. You can also modify GUI settings. When you configure additional
servers for monitoring, a similar window appears for each server.
Figure 3. Servers and Groups Window shows a typical Servers and Groups window.
Expandable list of servers and groups. Select a server to show all groups monitored by that server. Select
a group name to display group data in graphs and tables.
The icon next to each item in the tree hierarchy indicates the health condition of a group member:
A green arrow indicates a server with its associated groups is available. A yellow diamond in the icon
indicates that the default server, which cannot be removed. For multiple servers, you can change the
default server by right-clicking an available server and selecting Make This Default Server from the
drop-down menu. (The client always has one default server that is originally configured during
installation.)
A check mark in a green circle indicates that no health conditions exist
A question mark (?) in a blue circle (and the group name in yellow) indicates that a Caution level
condition exists
An exclamation mark (!) in a yellow triangle (and the group name in yellow) indicates that a Warning
level condition exists
An "x" in a red circle (and the group name or IP address in red) indicates that an event requiring
immediate attention exists or that SAN HQ cannot reach the group
Lists the groups monitored by the server and shows the following information for each group:
Group name
Monitoring status (Connecting, Connected, Disconnected, or Failed to Connect)
Amount of time elapsed since the group was last polled (Last Poll Update)
PS Series firmware on the group members (displays mixed if members are running different versions)
Group network address (group IP address, DNS name, or management address)
Location and description (based on the groups Group Manager settings)
Lists any active alerts, events, or audit logs for the selected group. Alerts indicate when an alarm (typically
a hardware problem), or a performance condition exists in a group. Events display syslog tracking of
system operations. Audits are syslog events about administrator actions. For information, see Reported
Alerts, Syslog Events, and Audit Messages.
Click the Servers and Groups button to display the All Groups window, or to exit a Settings window.
General settings SAN HQ GUI appearance, installation settings, client startup settings, tooltip
behavior, chart display settings, temperature display settings, alert display settings
Contact groups settings
E-Mail settings
Displays the configuration settings for SupportAssist, customer support cases reported back to SAN HQ,
and the Activity Log panel showing all recent SupportAssist activity. For more information, see
SupportAssist Configuration Settings.
You can also display tool tip help in the GUI windows by placing the pointer on a graph legend or on the
question mark next to a table title.
Table 1. SAN HQ Command-Line Options describes the options that can be specified with this command
to customize the view.
NOTE: The default behavior is to start with the normal view, which is the standard latest 8-hour
format.
Table 1. SAN HQ Command-Line Options
-View:" value" Specifies the view type. Options for value include:
"Summary"
"Capacity"
"Inbound Replica" or "Inbound"
"Outbound Replica" or "Outbound"
"Combined"
"IO"
"Events"
"Hardware "or "Firmware"
-TimeRange:"value" Specifies the time to be applied to the display, in the format: -TimeRange:"
start- end" or "start to end" where options for start and end include:
dd/mm/yyyy hh:MM:ss AM|PM
dd/mm/yyyy 24 hour clock
If not specified, the page loads in the standard latest 8-hour format.
-TimeLatest:"value" Loads data from the specified time. Options for value include:
"X" or "X hours" for X hours
"Y days" for Y days
"Z months" for Z months
If not specified, the page loads in the standard latest 8-hour format.
For example:
The following table describes the actions that occur if the command line contains any errors when you
enter it.
1. Identify the groups that you want to monitor. Make sure the groups meet the requirements described
in PS Series Group Requirements.
2. Identify the computer that will run SAN HQ Server and monitor the previously identified groups. SAN
HQ Server stores group data in log files that must be accessible to the computers running SAN HQ
Client.
See Requirements for Running SAN Headquarters Server and Log File Directory Requirements.
NOTE: Do not have multiple computers running SAN HQ Server monitor the same group.
3. Identify the computers that will run SAN HQ Client. SAN HQ Client enables you to run SAN HQ GUI,
which obtains group data from the log files maintained by SAN HQ Server and formats the data in
graphs and tables.
NOTE: If you install SAN HQ Server, SAN HQ Client is also installed on the computer.
No dedicated management configuration SAN HQ must have access to at least one Ethernet
port on each member
Dedicated management configuration SAN HQ must have access to the dedicated
management port configured on each member
Microsoft .NET 4.5 or later
NOTE: SAN Headquarters installation application installs Microsoft .NET 4.5 Client Profile if it is
not detected on the host system.
Host O/S recommended memory requirement, plus 512 MB base memory, plus 20 MB RAM per
member for each monitored group
512 MB of free local disk space to save temporary files
Directory for log files. SAN Headquarters Server stores group data in the log files. The directory must
meet the requirements described in Log File Directory Requirements.
Table 3. Protocol and Port Information for Standard SAN HQ Functionality lists the protocols and
ports for standard SAN HQ functionality.
HTTP (80) SAN HQ Client communication to the Web for retrieving recent updates
information (http://psonlinehelp.dell.com/V8.1/. For more information, see
Obtain SAN HQ Update Notifications.
UDP (514) (Optional) SAN HQ Server communication to the PS Series array, to display
syslog entries on the Events and Audits panel.
SNMP/UDP (161) Communication between SAN HQ Server and PS Series array for standard data
collection.
SSH (22) Secure Socket Shell communication between SAN HQ Server and PS Series
array, and for the initial SupportAssist configuration.
TCP/IP (2606) SupportAssist data collection port for communicating with the PS Series array.
For each monitored group, the directory must have at least the following amount of free space:
log_file_size x 13 + 100MB
You specify the log file size during the installation. The default data log file size is 5MB, the minimum
size is 2MB, and the maximum size is 10MB.
To use the computer running the SAN HQ Server as a syslog server to store event messages and
audits, the default size of the event and audit log file is 5MB, the minimum size is 2MB, and the
maximum size is 20MB. You can later modify the syslog size. When messages have consumed all the
free space in the event and audit file, new messages overwrite the oldest messages. For more
information about event messages, see Syslog Event Logging. For information about audit logs, see
Audit Messages
Local device on the server. If you want, you can set up this directory as a network share.
Requirement: You must specify a directory (for example, C:\SANHQ\Log). The directory cannot be
a root drive, such as C:\. By default, Windows hides the Program Data folder.
Network device. For example, you can use a directory on a PS Series group volume or a network
file share.
Dell recommends that you specify the Universal Naming Convention (UNC) name for a network
file share (for example, \\monservice\log). If you specify a mapped network drive, SAN HQ
converts it to a UNC name. Because of the potential for network disruption, Dell recommends that
you not locate the log file directory on the same group that you are monitoring.
By default, the SAN HQ Server (EQLXPerf) runs as a local user with the name LocalSystem. If you are
using a log file directory that is on a network file share, you must configure EQLXPerf to run as a
domain user that has full access to the network file share. During the SAN HQ installation, you will be
prompted for the domain user name and password.
For example, to assign a domain user name to EQLXPerf:
a. In the Windows Control Panel, click Administrative Tools and then Services.
b. Right-click EqlxPerf in the list of services and select Properties.
c. Click the Log On tab.
d. Select This account, enter the domain user name and password for EQLXPerf, and then click
Apply.
Optionally, you can assign a domain user name to EQLXPerf from the SAN HQ menu bar by selecting
Settings, then Server Settings. After you modify the EQLXPerf login credentials, you must stop and
start the SAN HQ Server, as described in Restart the SAN Headquarters Server.
Each computer running the SAN HQ Client must have read access to the log file directory and any
network resources being used.
In addition, if you want to allow a SAN HQ Client computer to add groups to the monitoring list,
configure email notification, or change the SNMP community name for a group, the computer must have
read/write access to the directory and any network resources. If you choose to enable TCP/IP
communication during a full client/server installation (see Installation Procedure), SAN HQ uses the
folder's permission to determine user access to the log file data.
Use the standard Microsoft NTFS security mechanisms for the log file directory. Right-click the directory
and select Properties. Then, click the Security tab and specify the access information.
a. Go to https://eqlsupport.dell.com.
b. Log in to your support account.
c. Go to the Downloads area. You can also contact your PS Series support provider to obtain the
SAN HQ software. Copy the software to a location accessible to the computer on which you
want to install SAN HQ.
Archives minimally require approximately 1.5 to 2 times the log size of the amount of data selected
Reports minimally require approximately 5MB per group per report, depending on the selected
time range
Exports minimally require approximately 50MB per group, depending on the selected time range,
data object, and size of the group's log files
Pre-installation Review
For each set of groups that you want to monitor, you must install SAN HQ Server on one computer. You
can install SAN HQ Client on additional computers. If you are already running SAN HQ, you can upgrade
to a new version and retain the data already collected.
Before starting SAN HQ software installation, review the prerequisites for installing SAN HQ described in
Prerequisites for Installing SAN Headquarters and perform the necessary pre-installation tasks.
If you are upgrading from a previous version of SAN HQ, see Upgrade SAN Headquarters before installing
the software.
Installation Procedure
To begin installing SAN Headquarters:
1. Double-click the SAN HQ executable file, SANHQSetup32And64.exe. This file can be found in the
Downloads area of the Dell support site at eqlsupport.dell.com.
The welcome screen appears. Click Next.
NOTE: The SAN Headquarters installation application installs the Microsoft .NET 4.5 Client
Profile if it is not detected on the host system.
2. Accept the end-user license agreement (EULA).
3. Accept the default directory for the application files (C:\Program files\EqualLogic\SAN HQ), then
click Next.
4. Choose to install the SAN HQ Server and Client. See Full Installation Server and Client.
Dell recommends that you set an individual rule for all SAN HQ executables for IPv4 and, if
configured, IPv6.
For detailed information about adding and configuring firewall rules, refer to the TechNet article,
Configuring Firewall Rules, at the Microsoft Windows Server TechCenter: http://
technet.microsoft.com/en-us/library/dd448559(WS.10).aspx
4. An installation dialog box displays as the application is installed. When complete, click Finish.
By default, Launch SAN Headquarters is selected. You will be asked to start the SAN HQ service
(EQLXPerf), which will poll and record requested group monitor data. If you do not start this service,
you can continue by using existing data logs, but not gather new data. For more information, see Log
File Directory Requirements.
5. (Optional) Enter the login credentials for remote SAN HQ server login. If you do not provide the user
name and password to authenticate the remote server, the connection to the server will be
degraded, and advanced SAN HQ features will be disabled.
Client-Only Installation
1. Specify the local cache directory where group performance data will be stored. Each monitored
group requires 30MB. Dell recommends using a directory on drive C: if you have sufficient space.
2. Specify the network file share where the log files are located, in the form:
IP-address-of-server\Monitor
If the TCP/IP settings were enabled during the SAN HQ Server installation, you will be asked to provide
user credentials of the SAN HQ Server before continuing. The client-only installation asks for the
credentials on first use. If the system is not on a domain, supply machine login information using
either of these forms:.\user-name, or system-name\user-name.
If you are installing the SAN HQ Server for the first time, the Add Group wizard starts, enabling you to
add groups to the monitoring list.
If you were previously running SAN HQ, and you are using the same log files, the SAN HQ Server
automatically locates the log files and resumes monitoring the groups. Some data points might be
missing for the time period that SAN HQ was not installed.
See Get Started with SAN Headquarters for information about adding groups to the monitoring list and
performing other post-installation tasks.
Preserve Data
SAN HQ provides several methods that you can use to preserve group performance data. For example,
you can:
You can preserve data at the current time or use a command line to perform the task regularly.
You must run SAN HQ from the same domain user account and from the same computer to start the
Group Manager GUI as a standalone application without entering login credentials. If single sign-on is not
enabled, Group Manager is launched in the default web browser, and you must enter the group
administration credentials.
NOTE: If you already configured single sign-on from Group Manager, do not configure single-sign
on in SAN HQ. Configuring single sign-on does not share credential information when the group is
already configured to use the firmware version of single sign-on.
For more information about launching the Group Manager GUI with single sign-on, see Launch Group
Manager with Single Sign-On.
Alerts table in the All Groups window (see Start SAN Headquarters GUI)
Alerts panel at the bottom of each window (see Reported Alerts)
Optionally, you can set up email event notification and designate email addresses to receive messages
when an alert occurs in the group. See Configure Email Notification for Group Alerts.
Optionally, you can configure a group to log events to the Syslog server that is part of the SAN HQ
Server. Events appear in the Events panel at the bottom of each window and also in the Events
window. See Syslog Event Logging.
Optionally, you can set up audit logs to track administrator actions. See Display Audit Logs.
Back Up Log Files and User Configuration File Before Uninstalling SAN HQ
NOTE: You need Windows administrator privileges to change the log file directory.
3. Record the location of the log files. You will need to specify the location of the log files when
reinstalling SAN HQ.
Back Up Log Files and User Configuration File Before Uninstalling SAN HQ
You might want to back up the log files and the user.config user configuration file before uninstalling
the software. Dell recommends that you back up your log file data, particularly before performing a full
client/server reinstallation.
All log files are located in the log file folder you specified when installing the software, usually in
\documents and settings\all users\Application Data\Equallogic\San HQ\Logs. To
locate the log files, see Record Location of Log File Data.
You can restore these files when the reinstallation is complete and the settings will be restored.
Uninstallation Procedure
To uninstall SAN HQ:
When you upgrade SAN HQ, the SAN HQ Server attempts to reconnect to the groups via SNMP. You
should expect minimal to no performance impact to your groups.
When you install a new version of SAN HQ without first removing the current software, SAN HQ
automatically uses the log files from the previous installation
When you upgrade an existing full client and server installation, the installation wizard asks if you want
to first back up the existing SAN HQ log file data. Accept the Backup SAN HQ data to the specified
location option (selected by default) when running the installation wizard to upgrade a full
installation.
CAUTION: Do not back up your data to the EqualLogic folder on the same drive under any
circumstances. Dell strongly recommends that you back up your data to a different physical drive
before upgrading SAN HQ. If you do not back up the log file data, you might not be able to
restore your existing environment.
If you remove SAN HQ before reinstalling, you can also choose to keep the existing log files. Make
sure you specify the current log file directory when prompted by the installation procedure.
If you upgrade to a new version of SAN HQ, any new type of data that the version collects will not
appear for dates before the time of the upgrade
If you upgrade to version 3.0 and later of SAN HQ, the installation automatically archives all data from
previous versions.
NOTE: After upgrading, the oldest supported compressed archive file (.grpx) from the previous
version of SAN HQ will not be viewable.
Automated news updatesConfigure a set amount of time before SAN HQ checks for more recent
information when launched. By default, SAN HQ checks for updates every 7 days. If new updates are
available, an icon is displayed in the Updates Notification button in the lower-right corner of the SAN
HQ Client.
ManuallyAt any time, select Check for Updates from the Help menu to gather the most recent
updates available. The Recent Updates dialog box displays, displaying recent updates. Click OK to
The Summary of Group window displays an overview of group status, capacity, I/O performance, and
network performance. You can navigate through the SAN HQ GUI windows in several ways and display
increasingly detailed group data in different contexts.
Table 4. SAN HQ Menu Bar shows the menu bar items and where to find additional information.
Group Launch Group Manager See Launch Group Manager with Single Sign-
On.
Launch Dell Storage Update Manager See Launch Dell Storage Update
Manager from SAN HQ
Export Group Data See Group Data Exports
Stop Monitoring See Stop Group Monitoring
Start Monitoring See Add a Group for Monitoring
Increase Log File Size See Increase the Log File Size
Remove from Monitoring List See Remove a Group from Monitor List
Hide Group Information See Hide Groups in the SAN HQ GUI
Help About, Navigation Help, User Guide, Release Notes, and Check for Updates
See Start SAN Headquarters GUI
Under the menu bar, the tool bar buttons perform the following actions:
Back or Forward Returns to a previously visited window or moves forward to a previously visited
window. You can also click the down arrow next to the Forward button to display the navigation
history.
Add New Server Adds an extra server to monitor groups on that server. Similar to selecting SAN HQ
from the menu bar, and then Add New Server.
Add Group Adds a new group to monitor. Similar to selecting SAN HQ from the menu bar, and then
Add New Group.
New Window Produces a new window displaying the current view. When you exit from the new
window, the original window remains open.
Launch Group Manager When you select a group from the Servers and Groups tree, the Launch
Group Manager button appears, letting you start the Group Manager GUI for the group. Similar to
selecting Group from the menu bar, and then Launch Group Manager.
Launch Dell Storage Update Manager Launches the Dell Storage Update Manager tool that analyzes
your current Dell EqualLogic storage configuration and guides you through the PS and FS Series
firmware update process.
Create Archive Saves group data to a compressed archive (.grpx) file. Similar to selecting SAN HQ
from the menu bar, and then Create Archive.
Print Prints the current window.
A single SAN HQ Client can monitor multiple servers. The Servers and Groups tree shows all servers that
have been configured and their associated groups. For multiple servers, one server is designated as the
default server and cannot be removed. Additional servers are secondary servers that can be added and
removed.
Summary information for all groups is available at the top of the Servers and Groups tree. This view
presents comprehensive information, across all groups, allowing you to make quick comparisons and
analyses of your infrastructure. Links are provided on the summary pages to take you to specific
information for a group. For more information, see All Groups Summary Information.
You can obtain more detailed information by expanding the group in the Servers and Groups tree and
selecting a subcategory: capacity, combined graphs, event/audit logs, hardware/firmware, I/O, iSCSI
connections, and network. You can further refine the information by expanding the tree. For example, to
see data for all disks for a member, select Hardware/Firmware Disks. For more information about all
subcategories, see Group Information Categories.
Callout 5 Favorites
Lists views saved as favorites that you added from Favorites on the menu bar. Right-click a saved favorite
to make it your home page, delete it, or rename it. See Add and Manage Favorite Views.
You can identify the category and object for which the GUI is displaying data by examining the window
title. In most cases, the title includes the group name, category, and object. For example: Capacity of All
Pools on Group - group06
Within each information category, you can view groupwide data (Group), or data that is restricted to a
specific group object: Pools, Members, Volumes, Volume Collections, and optionally, Storage Containers,
VVols, and VMs. For more information, see Context Link Bar Options.
The All Groups Summaries page provides links to summary views across all groups. Hover over the link
for a brief description of the summary view. Click a link (for example, Capacity Summary) to see the All
Groups summary for capacity. Data on the summary pages is refreshed every 2 minutes, or manually
when you click the Refresh button.
Alerts Summary Provides a quick look at the total number of alerts by severity level: Critical,
Warning, Caution, and Information. The color of this panel indicates alert severity (for example, green
when no alerts exist, red for critical alerts). Click the Alerts Summary link to navigate to the Alerts
panel for the default server. For more information about SAN HQ alerts, see Reported Alerts.
Capacity Summary Displays the capacity information gathered from the most recent poll of each
group. For more information, see Capacity Information.
Hardware/Firmware Summary Displays the Hardware/Firmware information gathered from the
most recent poll of each group. For more information, see Hardware_Firmware Summary.
Replication Summary Displays outbound replication information gathered from the most recent
poll of each group. For more information, see Replication Summary.
SyncRep Summary Displays synchronous replication information per group or individual volume,
gathered from the most recent poll of each group. For more information, see SyncRep Summary.
Volume Capacity Summary Displays volume capacity information per group or individual volume,
gathered from the most recent poll of each group. For more information, see Volume Capacity
Summary.
Volume I/O Summary Displays the individual volume I/O information gathered from the most
recent poll of each group. For more information, see Volume I/O Summary.
Capacity Summary
The All Groups Capacity Summary page displays capacity information gathered from the most recent poll
of each group. For each group, SAN HQ provides:
Unreserved Space Amount of unallocated space for thin-provisioned volumes in the group. The
reported size of a thin-provisioned volume can be larger than the volume reserve.
Free Space Percentage of free group space required to fulfill the maximum in-use space for all
thin-provisioned volumes.
Thin Provisioned Volumes Number of thin-provisioned volumes in the group. The reported size
of a thin-provisioned volume can be larger than the volume reserve.
Template Volumes Number of thin-provisioned volumes in the group that are template
volumes.
Thin Clone Volumes Number of thin-provisioned volumes in the group that are thin clone
volumes.
SyncRep Utiliaztion Total, in use, and free space allocated for synchronous replication volumes, and
the number of active synchronous replication (SyncActive) volumes in the group.
For all information categories, you can click the icon in the column heading to sort by size criteria. For
additional definitions of capacity and replication terms, see Capacity and Replication Terms.
Hardware/Firmware Summary
The All Groups Hardware/Firmware Summary page displays the hardware and firmware configuration for
all groups monitored by the server. The information is gathered from the most recent poll of each group.
Group and Member Name of the group where the Hardware/Firmware information is displayed.
Expand the group name to display individual members of the group
Firmware Full firmware revision number on the group or member
Model of the PS Series array
Controller List of controller models for the group or member
Member Status Summary of the current member status for the group or the individual status of the
member: online, offline, unconfigured, initializing, vacating-in-progress, or vacated
Service Tag Service tags of the groups and members
RAID status and RAID policy The current RAID status and RAID policy type for the group or
member. A degraded, reconstructing, or verifying RAID status might adversely affect performance.
RAID reliability score and RAID reliability rating A SAN HQ feature that helps you make decisions
about the resilience of the RAID policy on your group, reported both as a numerical score, and starred
rating. For information, see RAID Policy Reliability Reporting.
For related information about potential hardware problems, see Hardware Problems.
Replication Summary
The All Groups Replication Summary page displays outbound replication gathered from the most recent
poll of each group. SAN HQ displays the following information:
Group/Volume Group or volume name. Expand a group to display individual volumes configured
for replication.
Replication Partner List of all groups currently configured as an outbound replication partner
Status Current status of outbound replication for a group or volume. For a group, the value
indicates the group's configured replication status. For a volume, the value indicates a volume's
individual replication status.
Active Replicas Active replicas count; that is, the number of volumes actively replicating data to a
partner site. An individual volume indicates if it is actively replicating to a partner site.
Paused Replicas Outbound replication paused count; that is, the number of volumes currently
paused from replicating data to a partner site. An individual volume indicates if it is currently paused,
and the Status column provides additional details.
Waiting Replicas Number of volumes waiting for outbound replication to begin. An individual
volume indicates if it is currently waiting for replication to begin.
For more information about replication terms, see Capacity and Replication Terms.
SyncRep Summary
The All Groups SyncRep Summary page displays information about volumes configured for synchronous
replication. The information shows the total number of SyncRep volumes, in-sync volumes, out-of-sync
volumes, paused volumes, amount of changes for the group since the last SyncRep operation, and the
remaining amount of changes for the group to be synced.
For conceptual information about SyncRep, see About Synchronous Replication. To display SyncRep
volumes from the Servers and Groups tree, see Display SyncRep Volumes in SAN Headquarters.
Group/Volume Name of the group displaying the volume capacity information. Expand the group/
volume name to display the individual volumes for the group.
For more information about these items and additional capacity and replication terms, see Capacity and
Replication Terms.
Group/Volume Name of the group displaying the volume I/O information. Expand the group/
volume name to display the individual volumes for the group.
Volume type Template for a template volume; Thin Clone for a thin clone of a template volume;
Thin for a thin-provisioned volume; Standard for a standard, fully provisioned volume, or VVol for a
virtual volume
Pool Name of the pool containing the volume
Member Names of the members contained in the volume
Storage Container Name of the storage container containing the virtual volume.
Average I/O rate Average rate of iSCSI read and write data transfers to and from the volume
Volume average latency Average time it takes to complete the iSCSI read and write operation for the
volume
Volume average I/O size Average size of the I/O read and write operations for the volume
Volume average queue depth Average number of outstanding I/Os at the time of each incoming
I/O. A queue depth of zero indicates no outstanding I/Os.
Volume read and write percentages Read and write operations for the volume, expressed as
percentages of the total I/O for the volume
Volume member binding Member to which a volume is bound (by default, none). Load balancing
across members does not occur if a volume is bound to a member.
For more information about these items and additional I/O terms, see I/O Terms.
The following sections describe the information that the categories and related subcategories provide.
Capacity Information
For a selected group in the Servers and Groups tree, click Capacity to show the Capacity of Group
window. This view shows similar group capacity information as on the All Groups Capacity Summary
page (see Capacity Summary), and shows total unused space in the group that is reserved, but currently
not storing data.
Table 5. Inbound Replicas describes the information below the inbound replicas chart.
Volume size Volume reserve; that is, the actual amount of pool space allocated to the volume
Reported size Reported size of the replica volume, as seen by the iSCSI initiators. This number can be
more or less than the volume reserve size.
Reserve size Current amount of replica reserve for the volume, shown in GBs. The replica reserve is
the portion of delegated space reserved for storing replicas of the volume.
Reserved Current amount of replica reserve for the volume, shown as a percentage.
percent
Reserve free Amount of replica reserve that is not in use (not storing replicas of the volume), shown
in GB.
Reserve free Amount of replica reserve that is not in use, shown as a percentage.
percent
Reserve in use Amount of replica reserve in use (storing replicas of the volume), shown in GB.
Expanded size Amount of space that would be taken if compression was not enabled.
Outbound Replicas
The Outbound Replicas view displays all outbound replicas for the group. (Replication of the volume is
considered outbound from the view of the primary group.) You can also set the context link bar option
Table 6. Outbound Replicas describes the information below the outbound replicas chart.
Column Description
Replica Set Replica set name, based on the volume name and dot-number extension for
uniqueness. The number in the extension reflects the order in which each
partner was configured as a replication partner to the group.
Partner Replication partner to the primary group, to which the volume data is
replicated.
Volume Size Volume reserve; that is, the actual amount of pool space allocated to the
volume.
Reported Size Reported size of the replica volume, as seen by the iSCSI initiators. For
volumes that are not thin provisioned, the reported size is the same as the
volume reserve. For thin-provisioned volumes, the reported size can be
larger than the volume reserve.
Reserve Size Current amount of replica reserve allocated from delegated space for storing
replicas of the volume, shown in GB and as a percentage (Reserved Percent).
Delegated Space on Amount of space on the replication partner delegated to the group.
Partner
Status Replication status for a volume. See the Group Manager online help for a
complete list of replication statuses.
Most Recent Duration The amount of time to complete the most recent volume replication.
Most Recent Transfer Amount of data transferred during the most recent volume replication.
Size
Space Borrowed on Amount of space borrowed from remote delegated space for replicas of the
Partner selected volume.
For complete information about replication, see the Dell EqualLogic Group Manager Administrators
Guide.
Recoverable Volumes
The Recoverable Volumes view displays all pending deleted volumes on the group that can be undeleted
and recovered. SAN HQ displays information in two categories: General and Original. For information
about the fields for this view, see Display Recoverable Volumes in SAN Headquarters. For conceptual
information about recoverable volumes, see About the PS Series Firmware Volume Undelete Feature.
You can also set the context link bar option (see Context Link Bar Options) to show recoverable volumes
on all pools in the group.
Space borrowing allows you to temporarily increase the available snapshot space for a volume by
borrowing from the snapshot reserve of other volumes and from pool free space. For more information
about this view, see Snapshot Space Borrowing.
Combined Graphs
For a selected group in the Servers and Groups tree, click the Combined Graphs link to show the
Combined Graphs of Group window. Except for capacity information (Capacity pane), all information
presented is at the group level and will be the same when you select a context link bar option (see
Context Link Bar Options) .
Network and I/O Presents several graphs showing network and I/O data: average I/O size (KB),
average number of IOPS operations, average latency (ms), average I/O rate (KB/sec), TCP retransmit
percentage, network rate (KB/sec), and average queue depth. This information is similar to
information displayed on the I/O and Network information windows (see I/O Information and
Network Information). For information about I/O terms, see I/O Terms. For information about network
terms, see Network Terms.
Capacity For the selected time period, shows total capacity; in-use, free, and total snapshot reserve
capacity; and snapshot reserve in-use and free space. This information will change when you select a
different context (for example, VVols), from the context link bar.
Total Counts For the selected time period, shows total iSCSI connections and snapshots, I/O
operations per second against the weighted I/O latency, and I/O throughput rate against the average
I/O latency.
This view shows similar information to the Events and Audit Logs tabs in the bottom of any view. You can
search, export, or copy events or audit logs. For complete information, see Displaying Events and Display
Audit Logs.
Hover over an increment in the graph to display this information for a given point in time.
Table 7. Disk Status describes the information below the Disk Status chart.
Column Description
Member Lists all members in the group. Select a member name to navigate to the
hardware information for that particular member, including health conditions
for that member (for example, more spare disks were expected for the
member's RAID policy).
Model PS Series array model number, set if the supported firmware is running on that
member.
Service Tag PS Series array service tag, set if the supported firmware is running on that
member.
RAID Policy RAID level (RAID 10, RAID 5, RAID 50, RAID 6, or RAID 6 Accelerated)
configured on the members disk drives and number of spare drives.
Enclosure Serial PS Series array serial number, set if the supported firmware is running on that
Number member.
Disk Model Model for the disks present on the group or member.
Controller Type Type of control module (must be the same type on an array).
Estimated Last Boot Date and time when the secondary control module was turned on and started.
Expand the tree for Hardware/Firmware information to show the Disks subcategory (see Disk
Information).
Disk Information
The Disks on Group view shows read and write IOPS in graphical form and as a total and percentage for
disks on the member.
Table 8. Disk Information provides information about all disks for the member.
Column Description
Member Lists all members in the group. Select a member name to navigate to the
hardware information for that particular member, including health conditions
for that member (for example, more spare disks were expected for the
member's RAID policy).
Average Disk Queue The queue depth for the disk measured as the average number of outstanding
Depth I/O operations at the start of each incoming I/O operation. A queue depth of
zero indicates no outstanding I/Os are present.
Model PS Series array model number, set if the supported firmware is running on that
member.
Disk Type Specifies the disk type as either a hard disk drive (HDD) or a solid state drive
(SSD).
LifeTime Remaining (SSDs only) The estimated lifetime of the drive based on manufacturers
specifications and the number of write operations performed on the drive.
I/O Information
For a selected group in the Servers and Groups tree, click I/O to show the I/O for Group window.
Table 9. I/O for Group Information describes the I/O information in the panels on the left side.
General Information Percentage of read and write IOPs for the group. Shows the group
estimated I/O load, relative to the maximum capability of the group.
Average I/O size Average I/O size of the iSCSI read and right operations in the group.
Average IOPs Average number of iSCSI read and write processed each second in the
group.
Average I/O rate Average rate of iSCSI read and write data transfers in the group.
Average queue depth Average number of outstanding I/Os at the time of each incoming I/O.
Table 10. Group I/O describes the group I/O rate information displayed in the Group I/O chart.
Column Description
Group average I/O size Average size of the iSCSI read and write operations for the group.
Group inbound replication Average size of the iSCSI operations associated with inbound replication
I/O size to the group (for PS Series groups running firmware version 5.0 and
later).
Group average IOPS Average number of iSCSI read and write operations processed each
second in the group.
Group inbound replication Number of write operations processed each second by the group for
IOPS inbound replication (for PS Series groups running firmware version 7.0
and later).
Group average latency Average time it takes to complete iSCSI read and write operations in the
group.
Group inbound replication Average time to complete an inbound replication iSCSI operation to the
latency group (for PS Series groups running firmware version 7.0 and later).
Group average I/O Average rate of iSCSI read and write data transfers in the group.
Group inbound replication Average rate of iSCSI data transfers for inbound replication to the group
I/O (for PS Series groups running firmware version 7.0 and later).
Table 11. Group I/O Load Space Distribution describes the group I/O rate information displayed in the
Group I/O Load Space Distribution chart.
Column Description
Total Load Total amount of I/O load space.
High load1 Amount of I/O load space with a large number of I/Os running
at a given time.
Medium load1 Amount of I/O load space with an average number of I/Os
running at a given time.
Low load1 Amount of I/O load space with a low number of I/Os running at
a given time.
SSD space If the array includes SSD drives, amount of space available on all
SSD drives (displayed only on groups with at least one member
using tiered storage).
For more information about I/O terms used on the I/O for Group window, see I/O Terms.
From the Live View Sessions for Group window, you can:
For more information about the Live View feature, see Live View Data.
For prerequisites for running a Live View session, see Prerequisites for Establishing a Live View Session.
For step-by-step instructions for running a Live View session, see Run a Live View Session.
Experimental Analysis
The Experimental Analysis view provides estimated information about PS Series group performance
relative to a specific workload, to help you understand your group performance and plan storage
expansion. Because this information is only a theoretical potential, you should not use this data as the
sole measure of your group's performance potential.
This view provides the following estimated data, graphically and in tables:
Estimated IOPS workload For the group, provides the percentage of average IOPS relative to the
average estimated maximum IOPS, the percentage of read and write operations in the group, and the
RAID policies configured on a member's disk drives and spare drives. The workload percentage is an
indicator of how much work the group is performing relative to how much it can do, based on
maximum IOPS. "Work" is defined as the number of small (8KB), random I/O operations the group
performs each second.
Estimated Max IOPS Estimated maximum IOPS the group can process, based on the current
configuration and I/O load. This estimate also shows the estimated IOPS that can be processed when
a RAID set is degraded.
For information about accessing the Experimental Analysis view, see Access the Experimental Analysis
Window. For examples of interpreting Experimental Analysis estimated performance data, see Examples
of Interpreting Performance Data.
RAID Evaluator
The RAID Evaluation for Group view displays the current RAID policy information and lets you apply a
different RAID policy to compare and analyze performance benefits. Information is presented in tables
and graphs, and you can immediately see the results of applying a new RAID policy in the RAID policy
reliability meter. This view also provides a link to best practices for choosing a RAID member policy.
Refer to the following sections for complete information about RAID evaluation for a group:
For conceptual information about RAID evaluation, see RAID Policy Evaluation and Reliability.
For information about using the RAID evaluator, see Use the RAID Evaluator.
For information about RAID policy reliability reporting, which includes understanding the RAID
reliability score shown in the table data and meters, see RAID Policy Reliability Reporting. This section
also provides examples of evaluating the reliability of your current RAID policy and possible
improvements, describes how to display RAID policy reliability alerts, and provides additional
information in the form of frequently asked questions.
SyncRep
For PS Series groups running firmware version 6.0 and later, the SyncRep for Group view lets you monitor
information about synchronous replication (SyncRep) volumes that have been configured for the group.
By default, this view shows a summary of all SyncRep volumes, the number of in-sync and out-of-sync
volumes, paused volumes, the amount of changes for this group since the last SyncRep operation, and
the remaining amount of changes for the group to be synced. The timeline graph shows the remaining
changes to be completed.
For conceptual information about SyncRep volumes, see About Synchronous Replication.
For more information about the SyncRep for Group view, see Display SyncRep Volumes in SAN
Headquarters.
The All iSCSI Connections chart shows the total iSCSI connections for the group, the maximum
connections allowed, and the sample period for the data. (The Network for Group window shows the
number of the iSCSI connections to the group, which includes connections to volumes, snapshots, active
inbound replication, replication management, and host control management. It also shows the total data
read from and written to the group targets.)
Table 12. All iSCSI Connections describes the group iSCSI connection information provided for each iSCSI
target.
Column Description
Target Name of the connected volume or snapshot.
Host Initiator Name iSCSI initiator name of the host currently using the connection.
Host IP Address IP address of the host currently using the iSCSI connection.
Total Read Amount of data that has been read over the iSCSI connection.
Total Write Amount of data that has been written over the iSCSI connection.
Volume IQN iSCSI qualified name of the volume or snapshot, using the format: IQN, year-
month date, reversed domain name, and optional prefix.
Network Information
For a selected group in the Servers and Groups tree, click Network to show the Network for Group
window. Click an option on the context link bar to view network information for all pools on the group or
network information for all members on the group (see Context Link Bar Options).
For the sample period for the data, the following charts show the total number of iSCSI connections,
network load percentage, percentage of all TCP segment packets sent to the group, and sent I/O and
received I/O (MB/sec).
Table 13. General Network Configuration Information describes the general configuration information
provided for the group.
Column Description
Number of iSCSI Number of iSCSI connections to the group, which includes snapshots,
connections volumes, active inbound replication, replication management, and host
control management. For a detailed look at all iSCSI connections to the
group, see iSCSI Connection Information.
Management network Yes or no. If yes, the group has configured a dedicated management
enabled network, which separates iSCSI traffic from group network management
traffic. If no, a management network is not configured.
Ports with DCB support Total number of Ethernet ports with data center bridging enabled and
active active.
Ports with DCB support Total number of Ethernet ports with data center bridging enabled but
inactive inactive.
Ports incompatible with Total number of Ethernet ports with data center bridging disabled.
DCB
Column Description
Total link speed Sum of the negotiated link speeds for all active network interfaces in the
group, reported at half-duplex data transmission rate.
Active ports Number of active network interfaces in the group. Ports connected to
redundant control modules are counted as one network interface.
TCP retransmit Percentage of all retransmitted TCP segment packets sent by the group.
percentage This TCP retransmit percentage is tracked on each member, not on each
network interface. A TCP retransmit percentage greater than 1% might
indicate insufficient bandwidth. For more information, see How Your
Group Monitoring Environment Affects TCP Retransmission Reporting.
TCP retransmits Number of TCP segment packets retransmitted during the sample data
period.
TCP out packets Number of TCP segment packets successfully sent during the sample data
period.
Table 15. Network Rate Information describes the network rate information provided for the group.
Column Description
Received Group-received network trafficAverage per-second rate of network traffic
received by the group. Most received network traffic is for iSCSI write
operations, although group operations between members and replication
activity can increase this rate.
The Network Link meter indicates the estimated network load on the group as a percentage of total
bandwidth. The percentage displayed is either for sent data or received data, whichever is higher.
Expand the tree for Network information to show the Ports subcategory (see Network Port Information).
Column Description
Member Member in the group, repeated for each network interface. Select a member
name to view network information for that member (for example, glengrp -
eth1).
Interface Network interface name (for example, eth1). Selecting a member in the table
shows information for that particular network interface.
Link Speed Negotiated link speed for all active network interfaces, reported at half-duplex
transmission rate.
Operational Status Up or down. Indicates whether the network interface is operational. This status
can appear as down even though the administrator status is up.
Last Modified Date and time the network interface was last modified or when the array
restarted.
Port Type Indicates whether the port can be used for iSCSI and management traffic or
only for management traffic.
Data Center Bridging Indicates whether the Ethernet port is compatible and enabled for DCB, or
Status disabled.
Group (default)Displays groupwide data for the group information category you are viewing. For
example, for I/O information, SAN HQ displays the average I/O size, average IOPS, average latency,
average I/O rate, and average queue depth.
PoolsDisplays data for all pools on the group (for example, the capacity of all pools on a group) or
for an individual pool. When viewing all pools on the group, you can access individual pools from the
list in the lower panel.
MembersDisplays data for all members or for individual members. When viewing all members in the
group, you can access individual members from the list in the lower panel.
VolumesDisplays data for all online volumes. The drop-down menu lists the individual volumes you
can select. The volume icons indicate the type of volume: standard (fully provisioned), thin-
provisioned, template, thin clone, and NAS (see Identify Volumes).
Volume CollectionsDisplays data for all volume collections or for individual volume collections.
In some cases, the Volume Collections windows display data for the group. Click the GUI tooltip for
details.
If you have groups with VMware virtual volumes, the context link bar options shown in the following
figure are displayed. These options show storage containers, virtual volumes, and virtual machines.
Figure 8. Context Link Bar Options for Groups With Virtual Volumes provides the following information:
Storage containers Displays data for all storage containers. Lists all storage containers in the drop-
down menu; select among the list
VVols Displays all virtual volumes in the selected storage container or in all storage containers.
Select a VVol from the drop down menu
VMs Displays all VMs on the group
An option on the context link bar that does not apply to an information category appears dimmed, with
the exception of the Group object, which always appears dimmed when selected. If your group does not
contain virtual volumes, these context link bar options do not display.
You apply 95th percentile reporting on both groupwide (summary) data or data restricted to a specific
object (groups, pools, members, volumes, and volume collections). Table 17. 95th Percentile Reporting
shows where you can set the display to the 95th percentile for summary or detailed data in the SAN HQ
GUI.
By default, SAN HQ sets the view to Standard. For example, the standard view for group I/O might
resemble the example in Figure 9. Standard View for Group I/O, where write I/O data spikes reach 30KB
and write IOPS spikes are 1100 seconds.
To set 95th percentile reporting, click the 95th% button. . When you set the view to 95th% in the example
in Figure 9. Standard View for Group I/O, SAN HQ filters out the data spikes so the view looks like Figure
10. 95th Percentile View for Group I/O. In this case, the maximum write I/O is under 12KB and write IOPS
under 900 seconds). For more information, see Control Chart Display.
NOTE: When switching from Standard view to 95th percentile reporting, both summary and specific
data can show missing data if either the beginning or ending data points appearing in Standard view
are the high point of the graph. Because 95th percentile reporting factors out the top 5% spikes, the
polling period appears shorter at either the beginning or end of the graph.
For example, by default, graphs display data from the most recent 8-hour time period, and tables and
circle graphs display data from the most recent poll.
You can display data from a different time period by using the GUI navigation features shown in Figure 11.
GUI Timeline, Graphs, and Tables.
Shown as a gray rectangle on the timeline, controls the time range for the information that appears in the
graphs. You can select the time range selector and stretch it along the timeline to set the time range, or
you can use the zoom links. You can also select and move the selector along the timeline to change the
dates of the time range.
By default, when you first view the Summary of Group window, the timeline shows the most recent 10
days, and the time range selector is set to display data from the most recent 8-hour time period.
Enable you to quickly set the value of the time range selector and also control the range of dates in the
timeline:
Click Show Latest to show data up to the most recent poll. If you move the selector to the far right of
the timeline, Show Latest is automatically selected. If you move the selector to the left, Show Latest is
automatically deselected.
Click 1hr, 8hr, 1d, 7d, or 30d to set the time range selector to 1 hour, 8 hours, 1 day, 7 days, or 30
days, respectively. The link for a specific time range appears only after SAN HQ has gathered data for
this amount of time.
Click All to show all the dates for which data is available in the timeline
Click Custom to specify a specific range of dates to show in the timeline
Callout 4 Available data
Shows the current setting for the time range. Graphs show data obtained during this time period.
Shows the date and time for the data displayed in the tables and circle graphs. By default, the tables and
circle graphs show data from the most recent time in the selected time range.
Show data from the selected time range. Place the cursor over a graph legend to obtain help information
about the data.
To obtain a data point for a graph, the SAN HQ Server calculates the average of the data collected from
two consecutive polling operations.
Place the cursor over a point in time on a graph to display a tooltip with information about the data point,
including the polling (or sampling) period used to obtain the data.
Click a point in time to display data in tables and circle graphs from that time.
Use the pointer to select a time range in a graph. Tables and circle graphs will show data that is
averaged over this time range.
Use the left and right arrow keys to move back and forth in a graph.
If your pointer has a wheel, use it to zoom into and zoom out of a graph.
If SAN HQ is unable to communicate with a group because of a significant event in the group, such as
control module restart, failover, a firmware upgrade operation, or if you have an unsupported
hardware configuration, some data points might be missing from the graphs.
By default, shows data from the latest time in the selected time range. However:
If you click a point in time in a graph, shows data from that specific time
If you select a time range in the graphs, shows data averaged over that time range
Place the cursor over the question mark next to a table, or over a circle graph legend to obtain help
information about the data.
Toggles between a display of the currently selected data in either 95th percentile or Standard mode. The
95th percentile mode factors out the top 5 percent of data spikes for a more accurate representation of
your activity. Standard mode displays all data points including spikes.
An increasing polling period might indicate that the group has prioritized other high use activities on the
group (for example, replication in progress, rebuilding RAID, and so on.). As such, an increasing polling
period does not always indicate something is wrong with the group. Based on the responsiveness of the
array, it reviews the results of each poll to determine if the polling interval needs to expand, contract, or
remain the same. This process uses historical evidence of a group's poll responsiveness to determine the
optimal frequency for the group and then makes adjustments accordingly.
If SAN HQ is unable to communicate with a group because of a significant event in the group, such as
control module restart, failover, or a firmware upgrade operation
If you have an unsupported hardware configuration that prevents SAN HQ from obtaining certain
information
You reinstalled SAN HQ (using the previous installations log file directory) after some period of time.
No data was collected for the time during which SAN HQ was not installed.
Identify Volumes
SAN HQ provides several ways to quickly identify online volumes by type. When you select a group from
the Servers and Groups tree to display groupwide data, click the down arrow for the Volumes context link
option. A drop-down list of all volumes in the group appears, with icons indicating the volume type:
standard (fully provisioned), thin-provisioned, template, thin clone,and NAS (see Table 18. Volume Type
Icons). For information about these volume types, see the Dell EqualLogic Group Manager Administrators
Guide.
NAS Purple cylinder (NAS volumes use the NAS cluster name as a prefix)
Groups must meet the requirements described in PS Series Group Requirements. In addition, to
successfully monitor a PS Series group, you must meet the SAN HQ prerequisites described in SAN
Headquarters Prerequisites for Adding a Group.
For valid information, SAN HQ tries to connect to the group. This attempt might take several minutes. If
SAN HQ cannot connect to the group, it displays a dialog box allowing you to reenter the group IP
address or DNS name, and the SNMP community name. Click Retry to attempt the connection again.
After successfully adding a group, the SAN HQ Server starts to gather group data. Data appears in the SAN
HQ GUI graphs and tables as successive polling operations occur. Polling can take up to 10 minutes,
depending on the group workload. You can also add a group using a script/CLI-based method, which
provides a subset of the options in the GUIs Add Group wizard. For information, see Add a Group to the
Monitoring List from the Command Line.
Table 19. AddGroup CLI Command Options defines the command-line options and associated
parameters.
Table 20. Optional and Required Parameters for AddGroup defines the required and optional
parameters for the -AddGroup parameter.
Add Group Email SMTP Port= SMTP-servers- Specifies the port number to be used when
port-number sending alert email notifications to the specified
SMTP server. This command cannot be used with
UseGroupEmail or Use-
GroupEmailSettings.
If one or more servers are running a newer version of the client, a dialog box appears for you to upgrade
the client (see Upgrade the Client Monitoring Multiple Servers).
Data monitored from the servers is stored in separate log file locations. If the log files are incompatible,
you can remove the server from monitoring, or upgrade to the newer log file format.
The Servers and Groups tree shows a summary of all servers available. Selecting a server shows all
groups monitored by that server.
To add a server, see Add a Server. To remove a server, see Remove a Server. Note that you cannot
remove the default server, nor the last server monitored. To upgrade the client monitoring the server, see
Upgrade the Client Monitoring Multiple Servers.
1. Select the SAN HQ menu item, and then select Add New Server, or click Add New Server from the
toolbar. Alternatively, right-click SAN HQ Servers from the top of the Servers and Groups tree, and
then select Add New Server. The Add New SAN Headquarters Server wizard appears, which guides
you to provide all required information. Click Next.
NOTE: The servers you add can be in different physical locations, provided that you have
network connectivity between the client and server.
2. Enter server connection information required to initialize the connection to the server:
a. Enter the remote log directory path of the server to be monitored. You can specify either the
server DNS name or IP address. Specify the share name when supplying an IP address.
b. Enter a server display name to reference the SAN HQ server you are adding. You can choose any
name, or use the server IP address. (This field is automatically populated using the DNS name or
IP address from step a.)
c. Click Next.
3. When you add a server with TCP/IP enabled, you must enter your authentication credentials for
access to the remote server. The supplied credentials generate a validation ID for subsequent login
without needing to store the password. If the credentials you enter cannot be validated, then the
connection to the server will be degraded. This degradation might result in slower performance and
disable advanced features such as Live View. After you enter valid credentials, the server's
communication interface is established. The default TCP/IP port is 8000 during the installation,
although you can configure the server to use a different port. Click Next.
NOTE: If the server port has been updated, but the server has not been restarted to reflect the
changes, you will not be able to add the server until the server restarts to reflect the changed
port number.
4. The Add Server Connection Summary dialog box shows a summary of the information you entered
for the new connection. If the information is correct, click Add Server. Otherwise, use the Back
button to reenter information. When you click Add Server, you are connected to the server you
added. It might take a few minutes for the groups on the server to poll. If the connection cannot be
made, a message appears indicating the problem (for example, invalid credentials).
Remove a Server
To remove a new server from the SAN HQ GUI:
1. Right-click the server you want to remove from the Servers and Groups tree, and then select
Remove Server. You are asked to confirm removing the server.
2. Click Yes to remove the server from monitoring.
NOTE: You cannot remove the default server to SAN HQ. The Remove Server options appear
dimmed.
Scenario 1: A newer version of the client is running on the default server (vm2). In this case, you must
upgrade to the newer version. SAN HQ selected the newest server by default. Click the Upgrade button
to copy log files and install the newer version of the client (Figure 12. Scenario 1: SAN HQ Client Requires
Upgrade).
Scenario 2: A newer version of the client exists on two servers. In this scenario, the selected server
(fservice) is at a lower client version than the default server (vm2). Continuing to select that server will
require a second upgrade, as indicated in the warning message at the bottom of the dialog box (Figure
13. Scenario 2SAN HQ Client Requires Upgrade).
Scenario 3: In this scenario, SAN HQ detected a newer version of two servers running the client. The
non-default server (fservice) is running the newest version and therefore is selected by default. However,
When you click Remove, the dialog box reappears with the default server (vm1) selected. As before, the
top message indicates that the current SAN HQ Client version (2.5) should be upgraded to the selected
version (in this case, server vm1 at version 3.0, Figure 14. Scenario 3Incompatible Server Removed).
Figure 15. Log Directory Path Verification Dialog Box shows a dialog box that indicates the server name
and associated log file directory, and whether it is reachable. This example uses five servers, for one of
which the log directory is unreachable. When you click the Retry link, a dialog box presents information
about all the connected servers. You can review this information and, if you want, remove the non-
default server. Otherwise, you can ignore the condition, or exit the client.
General Settings Set the GUI theme and graph colors, installation settings (and log file directory),
tooltips settings, display format of charts, temperature settings, and alert display settings. For more
information, see:
The SupportAssist link below the Servers and Groups tree on the main page lets you navigate to the
SupportAssist Configuration Settings page, from which you can view and enable SupportAssist for your
groups, and view an activity log for groups monitored by SupportAssist (see SupportAssist Configuration
Settings). You can also view Dell Support customer cases created for your groups and reported back to
SupportAssist (see Display Support Case Information). Use the navigation links in the upper-left pane to
access these views. For more information about setting up and using SupportAssist to collect and upload
PS Series group diagnostics, see Configure and Use SupportAssist.
The Help link shows the current versions of the SAN HQ Client and Server, links to the Dell EqualLogic
support site, and provides the SAN HQ user guide and release notes PDF documentation from the
software kit.
1. Click Settings in the lower-left GUI panel. The General Settings page appears by default.
2. In the GUI Appearance panel:
Set the theme for the GUI windowsYou can use the theme that matches your operating system,
or you can select another theme available on the system
Change the colors used in the GUI graphsYou can use a predefined color scheme, or select
your own colors
3. Click Apply to implement and save the changes.
1. Click Settings in the lower-left panel. The General Settings page appears by default.
2. In the Installation Settings panel, the following information appears:
Installation TypeClient Only, when connected remotely to a SAN HQ Server; or Full, when
running on the system with the SAN HQ Server
Installation DirectoryWhere you installed the SAN HQ software
Log File DirectoryLocation of the log files maintained by the SAN HQ Server. If you have the
correct permission, you can change the location of the log files. See Change the Log File
Directory. This information does not appear if you are running a multiserver remote client.
Log File Directory In-Use SpaceServer only. The current amount of space in use by the log files.
Server InformationFor a client-only installation, provides the installation directory location,
servers monitored, and client cache directory location
Client Cache DirectoryLocal directory for caching data related to SAN HQ. The directory does
not appear if you are running the SAN HQ Server.
3. Click any directory link to open the directory.
To modify the log file directory, you must be running the SAN HQ Server
Before changing the log file directory, close any SAN HQ GUI sessions that are accessing the log files,
except for the GUI session that will perform the change
Make sure the new log file directory meets the requirements described in Log File Directory
Requirements.
NOTE: SAN HQ automatically restarts the SAN HQ Server after the log file location is changed.
1. Click Settings in the lower-left GUI panel. The General Settings page displays.
2. In the Installation Settings panel, click Change. The Change Log Files Directory Welcome dialog
box appears. Click Next to start the wizard.
3. In the Change Log Files DirectoryNew Location dialog box, specify:
New log file directory location
Optionally, whether to delete the original log files and directory after the files are copied to the
new location
4. Click Copy to start the copy operation.
5. When the Change Log Files Directory Restart GUI dialog box appears, click Restart GUI.
If you chose to keep the original log file directory, you can manually delete the log files and directory.
1. Click Settings in the lower-left GUI panel. The General Settings page appears.
2. In the Client Startup Settings panel, select Check for Windows Firewall rules at Client startup.
3. Click Apply to implement and save the changes.
Control Tooltips
From the General Settings page, you can determine how tooltips appear in the GUI. By default, when you
move the pointer over a point in time in a graph, tooltips appear that display data from the selected time.
Similarly, when you move the pointer over a graph legend, tooltips appear with definitions of the data.
To control tooltip behavior:
1. Click Settings in the lower-left GUI panel. The General Settings page appears.
2. In the Tooltip Settings panel, turn tooltips on or off by selecting or deselecting the choices listed.
3. Click Apply to implement and save the changes.
1. Click Settings in the lower-left GUI panel. The General Settings page appears.
2. In the Chart Display Settings panel, select either Standard or 95th percentile from the drop-down
list. Optionally, you can check the box to restore all displays to the default.
3. Click Apply to implement and save the changes.
1. Click Settings in the lower-left GUI panel. The General Settings page appears.
2. In the Temperature Display Settings panel, select either Fahrenheit or Celsius.
3. Click Apply to implement and save the changes.
1. Click Settings in the lower-left GUI panel. The General Settings page appears.
2. In the Alert Display Settings panel, select Hide all disabled email alerts from the User Interface.
3. Click Apply to implement and save the changes.
Startup Settings Start the Windows service used by SAN HQ using either the Local Service account
(default) or with specific user account credentials. For the SAN HQ service to use a network log
directory, the service must be running as a domain user. If you have administrative privileges, you can
change the startup settings for the service in the Startup Settings panel. Dell recommends that you
configure to run as a specific user and supply a user name and password.
Connection Settings for Client Change the communication settings for advanced features.
The SAN HQ Server is capable of using an advanced communication infrastructure. You can specify a
host name (or IP address), and a port number that will be used by all clients for advanced
communication. SAN HQ clients that do not have the correct connection information will use a
degraded connection format that retrieves only a subset of data from the server. (For more
information, see Degraded Connection Status.)
Click Browse to locate and select the standalone Dell Storage Update Manager kit on a local drive.
If you previously configured a local kit, click Delete to remove the older version.
When you launch the Dell Storage Update Manager, you will be directed to the Dell Support to obtain
the latest kit (see Options for Launching Dell Storage Update Manager).
1. Right-click the server name in the Servers and Groups tree and select Change Login Credential.
2. In the dialog box, resupply the credentials of the SAN HQ Server monitoring this group. Note that
these credentials must be resolved by the server to validate successfully.
To correct this problem, add a rule to allow the SAN HQ Server process (SanHQServer.exe)
communication to the port specified during installation.
To correct this problem, verify that the SAN HQ Server process, eqlxperf, is restarted and listed in
Windows services.
The SupportAssist Settings Configuration Settings page appears. Use the upper-left panel to navigate to
the activity log for recent SupportAssist activity, to a list of customer support cases reported back to
SupportAssist, or back to the SupportAssist configuration page.
For information about the SupportAssist configuration settings, see SupportAssist Configuration Settings.
For information about recent SupportAssist activity reported in the activity log, see SupportAssist Activity
Log. For information about open Dell Support cases reported to SupportAssist, see Display Support Case
Information.
NOTE: To remove a group from the list of monitored groups, the computer running the SAN HQ
GUI must have read/write permission to the log file directory and the network file share, if
applicable. On Windows Vista operating systems and later, if the group was added by a client
running as the administrator, then the group might require that you set Run as administrator to
enable administrative privileges to make any group modifications.
To remove a group from the list of monitored groups:
1. Select the group from the Servers and Groups tree. The Summary of Group window appears (Figure
4. SAN HQ GUI Navigation).
2. (Optional) To preserve logged data, create an archive. See Group Data Archives.
3. From the Group menu, select Remove From Monitoring List.
4. In the confirmation dialog box, select Remove the group and delete the log files.
You are also given the choice to temporarily stop monitoring the group and keep the log files, as
described in Stop Group Monitoring.
1. Right click on Default Server or a server you added and then select Sort By.
2. Select Name to display alphabetically ascending or descending. The up or down arrow indicates the
direction the next sort will take.
3. Select Online Status to display online groups at the top of the tree.
To sort the list of groups from the All Groups Monitored by Server window:
1. On the Servers and Groups tree, select Default Server or a server you added.
2. Right-click on the Group column head, and select any of the following options:
1. In the Servers and Groups tree, right-click the name of the group that you want to hide.
2. Select Hide Group Information.
You can also access the Edit Hidden Groups page by right-clicking in the Servers and Groups tree panel
and selecting Edit Hidden Groups.
The option to increase log file size can be selected only by a user with full access to the log location.
Increasing the log size might cause the amount of time required to load the logs to also increase due
to the extra levels of detail.
After you increase the log size, you cannot decrease the size without removing the group from SAN
HQ and deleting the existing logs.
To increase the log file size:
1. In the Servers and Groups tree, select a group, and then from the Group menu, select Increase Log
File Size. You can also right-click a specific group in the Servers and Groups tree, and then select
Increase Log File Size. The Increase Log File Size dialog box appears.
Alternatively, to change the log file size of several groups from one location, select a server or group
from the list in the Servers and Groups tree. From the SAN HQ menu bar, select Settings Group
Settings. The Group Settings page appears, showing all groups for that server.
a. In the list of group names, check the box to the left of the group for which you want to increase
the log file size.
b. In the Log File Size column, click Change Size. The Increase Log File dialog box appears.
2. In the Increase Log File dialog box, use the slider to increase the log file size.
3. Select the checkbox to acknowledge that you understand that you cannot later decrease the log file
size.
4. Click OK.
For related information about how data is compressed in log files, see How Data Is Compressed in Log
Files.
You can obtain the name by using the Group Manager GUI or CLI. For example, in the Group Manager
GUI, click Group Configuration, and then click SNMP.
NOTE: To change the SNMP community name for a group, the computer running the SAN HQ GUI
must have read/write permission to the log file directory and the network file share, if applicable.
See Log File Directory Requirements.
1. From Settings, select Group Settings. The Group Settings page appears.
NOTE: Do not use the special character # in the SNMP community name. If the page is
labeled read-only, you do not have the correct credentials for changing the SNMP community
name.
2. Select the groups SNMP community name and modify the name.
3. Click Apply.
Each new alert generates an email message. Subsequent alerts for the same problem are combined into a
single email message every 6 hours to limit redundant messages.
The SAN HQ Server also can construct combined email notifications for any alerts still active on a group.
To limit redundant messages, you can adjust the time between these combined email notifications from
the default of 6 hours. For information, see SAN Headquarters Server Configuration Settings.
NOTE: If an issue that generates a SAN HQ alert is resolved within a polling period, email notification
does not occur.
To set up or modify email notification for a monitored group, the computer running the SAN HQ GUI
must have read/write permission to the log file directory and the network file share, if applicable. See Log
File Directory Requirements.
Dell recommends that you use the PS Series group email notification feature as the primary notification
method for group events. The SAN HQ alert notification system augments the group email notification
feature and should be considered a supplementary method of providing notification.
1. From Settings, select EMail Settings. The EMail Settings page appears.
NOTE: If the page is labeled as read-only, you do not have the correct credentials for
configuring email notification.
2. In the row with the group name, select the checkbox in the Enabled column.
3. In the EMail Settings panel, select Send EMail Alerts.
4. Select Use group email settings if you want to populate the fields with the email notification settings
already set up in the group. Select Use these email settings if you want to use the email settings you
enter in this panel. Enter the following information:
IP address of the SMTP server that will deliver the email. Alternatively, you can enter the DNS
name for the server.
(Optional) A port for the SMTP server. The default port is 25.
To modify the existing email notification configuration for a group, in the EMail Settings Notification
dialog box, select the group in the top table and edit the data in the top table and click Apply, or edit the
fields in the EMail Settings panel and click Apply.
You can also change the alert selections in the EMail Settings Notification dialog box. Click Apply
when finished.
1. In the All Groups window, select the group in the Servers and Groups tree. The Summary of Group
window appears.
2. From the Group menu, select Stop Monitoring. Group data will still appear in the GUI, but the
monitoring status is disconnected. Note: If Stop Monitoring is not active in the Group menu, you do
not have the correct credentials to stop group monitoring.
3. To resume monitoring the group, from the Group menu select Start Monitoring.
The credentials you supply are encrypted and stored in the domain user account and used to
automatically log in to Group Manager from SAN HQ clients connected to this SAN HQ service. For more
information about requirements for single sign-on, see Single Sign-On Requirements. For information
about deleting login credentials for a group, see Delete Login Credentials for a Group on page lxxvi.
You can also configure single sign-on when adding the group to the monitoring list or at a later time (see
Enable Single Sign-On / Modify Login Credentials).
Single sign-on has a 5-minute timeout. If the system is idle for 5 minutes, SAN HQ displays a dialog box
indicating that unless you click Continue Session, the single sign-on session will close.
The user account running the SAN HQ GUI must be a domain user account with write privileges to
the log file directory.
NOTE: The saved login credentials for a group can be used only if you are running this domain
user account on the same computer.
The computer running the SAN HQ GUI must be running Java 1.5.0 or a later version. Dell
recommends using Java 1.6.0 Update 7.
1. From the Settings menu, select Group Settings. The Group Settings page appears.
2. Select the group.
3. Below Single Sign-On Settings, select Enable.
4. In the table row that contains the group, click Add to add login credentials or click Modify/Delete to
modify the existing login credentials.
5. In the Single Sign-On dialog box, enter or modify a group administration account name and
password and click Save.
To launch Group Manager and log in to the group, click Launch in the Group Settings window or click
Group Manager in the toolbar.
NOTE: If you enabled single sign-on and you are prompted for login credentials, be sure you
entered the correct account name and password. Also, make sure you are using the same domain
user account and the same computer that you used to configure single sign-on.
1. From the Settings menu, select Group Settings. The Group Settings page appears.
2. Select the group in the table.
3. Below Single Sign-On Settings, select Disable.
4. Click Apply.
When you want to re-enable single sign-on for the group:
a. From the Settings menu, select Group Settings. The Group Settings page appears.
b. Select the group in the table.
c. Below Single Sign-On Settings, select Enable.
d. Click Apply.
1. From the Settings menu, select Group Settings. The Group Settings page appears.
2. Select the group in the table.
3. Click Modify/Delete.
4. In the Single Sign-On dialog box, click Delete and confirm that you want to delete the credentials.
The Dell Storage Update Manager is deployed as a Java application and runs independently from the
Group Manager and SAN HQ. Java 6 and later are supported, with certain constraints. For information,
see "Usage Constraints in this Release" in the Dell Storage Update Manager Release Notes, available from
the Dell EqualLogic Support Site: https://eqlsupport.dell.com
The Dell Storage Update Manager targets groups running PS Series firmware 7.0 and later and FS Series
firmware version 3.0 and later, and updates the following components:
NOTE: For updating PS Series firmware prior to version 5.0, you must use either the Group Manager
or CLI. For updating FS Series firmware prior to version 3.0, you can only use the Group Manager
CLI.
The Dell Storage Update Manager supports both 32-bit and 64-bit browsers.
For complete information on using the Dell Storage Update Manager, see the Dell Storage Update
Manager User's Guide.
SAN HQ menubar
Servers and Groups tree
Context link bar options menu
Group settings for server panel
SAN HQ Menubar
Perform these steps to Launch Dell Storage Update Manager from the SAN HQ menubar:
1. Click the Settings button in the lower left-hand pane or use the Settings menubar option.
The General Settings panel displays.
2. Click the Group Settings link in the left-hand panel
3. For the desired group in the grid, click Launch in the Dell Storage Update Manager column.
SAN HQ links to the Dell EqualLogic support site and launches the Dell Storage Update Manager. If you
configured a standalone kit, SAN HQ launches the local copy of the Dell Storage Update Manager (for
more information, see xref Launch Dell Storage Update Manager for Standalone Firmware Updates).
1. Click the Settings link below the Servers and Groups tree or the select Settings from the menubar.
The General Settings page is displayed.
2. Select Server Settings.
The Dell Storage Update Manager panel displays on the bottom of the screen. Choose among the
following options:
Click Browse to locate and select the standalone Dell Storage Update Manager kit on a local
drive.
If you previously configured a local kit, optionally click Delete to remove it. Use this feature to
remove the older version. When you launch the Dell Storage Update Manager (see Options for
Launching Dell Storage Update Manager), you will be directed to the Dell Support to obtain the
latest kit.
Last Updated indicates the date of the downloaded standalone kit. If a newer kit is available from
Dell Support, download the newer version, delete the older version, and browse for the new local
kit.
For information about using the Dell Storage Update Manager, see the Dell Storage Update Manager User
Guide and Dell Storage Update Manager online help.
Use the Toolbar menu. This method applies the active window as a favorite.
Select a specific node as a favorite from the Servers and Groups tree. This method adds the view
associated with that node as a favorite.
The current view is then added to the list of favorites in the drop-down menu available from the Favorites
menu on the toolbar and under the Favorites node in the Group view.
The selected node is then added as a favorite at the group level. It is added to the list of favorites in the
drop-down menu available from the Favorites menu on the toolbar and under the Favorites node in the
Group view.
1. Navigate to the specific view that you want to set as the home page.
2. From the toolbar, select Favorites and then Make This Home Page.
The current view is then defined as the default window for SAN HQ.
To set a view as a home page using the Servers and Groups tree:
1. Navigate to the specific view that you want to set as the home page.
2. Right-click in the view and select Make This Home Page.
SAN HQ lets you monitor synchronous replication for SyncRep-enabled volumes, where volume data
exists simultaneously on two copies of the volume:
SyncActive The active copy of the volume to which iSCSI initiators connect when reading and
writing data.
SyncAlternate When data is written to the SyncActive volume, the group simultaneously writes the
same data to this copy of the volume in the alternate pool.
To switch the roles of the SyncActive and SyncAlternate copies of the volume, use Group Manager.
in sync SyncActive and SyncAlternate copies of a volume contain the same data.
out of sync SyncActive volume and SyncAlternate volume might not contain the same data; the
SyncActive volume contains the most recent data. A volume can become out of sync if synchronous
replication is paused, or if one of the volumes becomes unavailable or has no free space. The volume
can become out of sync when the snapshot reserve in the SyncAlternate volume is full, but only when
the snapshot space recovery policy sets volumes offline when the snapshot reserve is depleted.
paused Administrator has paused synchronous replication. While synchronous replication is paused,
the volume is still online, and initiators can connect to and write to the SyncActive volume. An
administrator might pause and later resume synchronous replication.
For complete information about requirements for and setting up synchronous replication, see the Dell
EqualLogic Group Manager Administrator's Guide.
4. Select a volume name listed under Volume Name in the table of SyncRep volumes. Information
about that single volume appears in the SyncRep for Volume window (Figure 18. SyncRep for Volume
Window.
When a SyncRep volume's state is in sync, the SyncRep summary indicates whether the SyncActive pool
or the SyncAlternate pool is the default. When the copies of the volume are out of sync, the SyncRep
Summary indicates read and write rates for the SyncAlternate pool, total changes made, and the
remaining changes to be made.
You can also view information about SyncRep volumes on groups or volume collections by selecting the
object on the context link bar.
You can easily view I/O data for volumes or virtual volumes by changing the target and selecting a
volume or virtual volume to monitor. Volume data shows average I/O size, average IOPS, average latency,
average I/O rate, and average queue depth for the specified polling interval and session length. You can
set the poll interval to 1 (default), 3, 5, 8, 10, or 15 seconds and session length to 1 (default), 2, 3, 5, or 10
minutes for volume and member. For VVols, you can set the poll interval to 35 (default) or 60 seconds
and session length to 1 (default), 2, 3, 5, or 10 minutes.
The Live View Sessions feature is always visible under I/O in the Servers and Groups tree, even if you have
not established a Live View session (see Prerequisites for Establishing a Live View Session). In that case,
the Live View window appears with no sessions. When you activate a Live View session, the Servers and
Groups tree displays the Live View session window for a given group member or volume.
The SAN HQ Server must be running and have the correct SNMP community name and IP address
specified for the group.
The TCP/IP communication infrastructure must be enabled on the SAN HQ Server with the
appropriate firewall rules established. If this infrastructure is not enabled, the SAN HQ Server is still
listed as Connected, showing a green circle icon next to the server name in the Servers and Groups
tree. However, the Live View feature is disabled. To correct this problem, rerun the SAN HQ
installation on your server in Modify/Repair mode. During the reinstallation process, the Installation
wizard presents a dialog box asking you to select Enable TCP/IP Communication, enter an IP address
(or server name), and enter the port number for communication.
The SAN HQ Client must not have a degraded connection to the server. For information about
rectifying a degraded connection, see Degraded Connection Status.
You must have approximately 100KB of disk space for each Live View session.
a. Select a group from the Servers and Groups tree, then select I/O. The I/O for Group window
appears.
b. Select an individual member, volume, or VVol from the I/O for Group window. For example,
selecting Members for Group A shows all members for Group A. Select an individual member
from the list in the bottom panel. I/O information for that member appears
c. c)In the upper-right corner of the I/O for Member (or I/O for Volume or VVol) window, click Live
View to start a Live View session.
Run a Live View session directly from the Live View Sessions for Group window:
a. From the Servers and Groups tree, select I/O then Live View Sessions. The Live View Sessions for
Group window appears.
b. Click the New Session button.
a. Select the group member in the Servers and Groups tree, then I/O, and then Live View Sessions.
Live View sessions appear in the main panel for the member.
b. Select the saved session. The session name includes the member, volume, or virtual volume
name, and the date and time the session was run. The saved session appears in the Live View
window, from which you can select a target and rerun Live View.
NOTE: If multiple clients initiate a Live View session on the same member or volume, the second
client is unable to initiate a session until after the first session completes.
Figure 19. Completed Live View Session for a Volume shows a completed Live View session. The session
length was 1 minute with a 1-second polling interval. The status message indicates 58 polls were
completed in the session timeframe.
The RAID policy reliability feature scores your current RAID policy for overall effectiveness based on Dell's
baseline recommendations for all members in a group. The RAID policy reliability meter provides a quick
indication of whether your current or possible RAID policy is poor, fair, or excellent, and helps you make
informed decisions about whether to reconfigure your RAID policy for the group (see RAID Policy
Reliability Reporting).
1. Select a group from the Server and Groups tree, then I/O, and then RAID Evaluator. The RAID
Evaluation for Group window appears, showing groupwide data by default (see ). You can also
restrict data to a specific pool or member in the group from the context link bar (see SAN HQ Client
The RAID policy reliability meter (see ) indicates the RAID policy reliability score (see RAID Policy
Reliability Score) for your current RAID policy (top meter). The bottom meter initially shows identical
information. When you evaluate a new RAID policy, the possible score (bottom meter) dynamically
changes the reliability score and (if necessary) the color to match to the new RAID policy rating (see
Evaluate Possible RAID Policy Improvements).
Poor (red, 5.00 to 0.00) Dell recommends changing the RAID policy for the group.
Fair (yellow, 0.00 to 1.18) The potential exists for a failure. You should consider changing your RAID
policy.
Excellent (green, 1.19 to 5.00) Your current RAID policy is reliable (the higher the reliability score,
the more resilient the policy).
Unknown (gray, 0.00) SAN HQ does not recognize the group's configuration (see Unknown RAID
Policy Reliability Status).
The View Best Practice link between the meters connects you to a Dell Support page with information to
help you remedy or improve your RAID policy reliability score. For a RAID policy reliability status of
unknown, a More Info link appears with information about possible causes for the unknown
configuration.
By default, information is displayed at the group level, while the RAID policy is set at the member level.
The RAID policy reliability meter also appears for all members in a group, an individual member, all pools
on the group, or an individual pool.
Figure 20. RAID Reliability Initial View shows the initial view when you first access the RAID Evaluator.
Callout Description
1 Access RAID Evaluator from the Servers and Groups tree.
2 The current RAID policy is indicated here. Initially, the possible RAID policy is the same as
the current policy. From the drop-down list, you can select and evaluate a different RAID
policy.
3 RAID policy reliability scores are indicated in the Possible Statistics table and in the reliability
meters.
4 The current RAID policy reliability score is indicated by numerical score, quality rating, and
graphically in the meter. In this example, an excellent rating is evaluated. The RAID policy
reliability score of 5.00 is above Dells recommended threshold of 1.19. The green color in the
meter reflects excellent reliability.
5 A link to Dell Support provides information for remedying a suboptimal RAID policy. In this
case, the RAID policy reliability score is high and no changes are needed.
6 The possible RAID policy reliability initially defaults to the current evaluation. Changing the
RAID policy in callout 2 would change the data in this meter (for more information, see
Evaluate Possible RAID Policy Improvements).
7 The data graph shows current and possible estimated maximum IOPS. In this example, the
maximum IOPS are the same.
For example, Figure 21. Possible RAID Reliability Improvements shows a RAID policy evaluation where the
policy is changed from mixed to RAID 6. A mixed RAID set yields only a fair reliability score of 0.86.
Changing the policy to RAID 6 results in an excellent reliability score of 5.00, significantly above Dell's
recommended threshold of 1.19.
The numbered items in Table 22. Callouts for Possible RAID Reliability Improvements Figure correspond
to the callouts in Figure 21. Possible RAID Reliability Improvements.
Callout Description
1 Select a new RAID policy from the drop-down list. In this example, RAID 6 is selected.
2 RAID policy reliability scores are indicated in the Possible Statistics table and in the
reliability meters. In this example, the Max IOPS Growth rate is negative with the new
policy, which means performance growth is less than before. Although negative, ample
headroom still exists, as indicated by the Possible Max IOPS rate in the graph (callout 5).
Capacity remains relatively the same.
3 The current RAID policy reliability score is rated fair because the reliability score is less
than 1.19 (in this example, 0.86). A negative reliability score would have generated a poor
rating.
4 The possible RAID reliability score is indicated by numerical score and rating in the RAID
policy reliability meter. In this example, an excellent rating is evaluated. The RAID policy
reliability score of 5.00 is above Dells recommended threshold of 1.19. The green color in
the meter indicates excellent reliability.
5 The data graph shows a lower Max IOPS rate, but ample headroom with the new RAID
policy.
6 In the table, only one member has a current RAID policy reliability score of fair. Current
scores and ratings for all members are displayed. Select a member name to obtain more
information (see Figure 21. Possible RAID Reliability Improvements).
Member HandsOnLabSumo01 shows a RAID policy reliability score of fair. The Max IOPS Growth rate is
even less at the member level than reported for the group (seeFigure 22. Possible RAID Reliability
Improvements on Member).
The numbered items in Table 23. Callouts for Possible RAID Reliability Improvements on Member Figure
correspond to the callouts in Figure 22. Possible RAID Reliability Improvements on Member.
Table 23. Callouts for Possible RAID Reliability Improvements on Member Figure
Callout Description
1 The Max IOPS Growth rate is lower at the member level.
2 The graph shows lower headroom for the Possible Max IOPS rate.
Changing to RAID 10 continues to show an excellent reliability rating and increases room for
performance growth (see Figure 23. Changed RAID Policy for Member Shows Improvements).
The numbered items in Table 24. Callouts for Changed RAID Policy for Member Figure correspond to the
callouts in Figure 23. Changed RAID Policy for Member Shows Improvements.
Table 24. Callouts for Changed RAID Policy for Member Figure
Callout Description
1 RAID 10 is selected for evaluation.
2 The table shows an improved Possible Max IOPS rate.
1. Click the Alerts tab at the bottom of the RAID policy reliability view.
2. Select the alert to display the RAID policy reliability score.
3. Click the alert's priority to view information about the object that generated the alert.
Occasionally, SNMP timeouts during RAID verification might cause critical information to be temporarily
unavailable and an unknown status to display.
In Figure 24. Unknown RAID Policy Reliability, the current RAID 10 policy shows a gray reliability meter
with a RAID policy reliability score of zero and status of unknown. Changing to RAID 6 does not alter the
result.
The numbered items in Figure 24. Unknown RAID Policy Reliability correspond to the callouts in Table 25.
Callouts for Unknown RAID Reliability Figure.
Callout Description
1 No RAID policy reliability scores are indicated in the Possible Statistics table.
2 The current RAID policy reliability score is rated as unknown and no reliability score can
be calculated.
3 The More Info link provides additional information about possible causes of the unknown
configuration.
4 Changing the RAID policy from RAID 10 to RAID 6 does not change the status of an
unknown configuration.
Question: Does RAID 6 take longer to rebuild than RAID 50? If so, why is the RAID policy reliability score
better for RAID 6?
Answer: RAID 6 does take longer to rebuild than RAID 50. However, the odds are lower for a RAID set
failure using RAID 6. RAID 6 has two parity drives per set, which allows it to survive a double failure,
thereby maintaining an advantage over other RAID policies that cannot survive a double fault.
Question: I have a PS6510E array with a RAID 50 configuration. However, my RAID policy reliability score
shows a poor rating. Why?
Question: With the same PS6510X array with 900GB in a RAID 50 configuration, the evaluated RAID 6
configuration shows that I will exceed my Estimated IOPS if I make this change. What should I do?
Answer: Before you make any RAID policy change, consider these aspects of your environment: the
overall I/O pattern throughout the day, the current response times and any existing I/O pressures, and
capacity. If all these elements align with the workload you are considering changing, then you might
need to move volumes to another pool or add an additional member to the current pool. This action will
help spread the load and provide a better overall utilization of the group.
Answer: The industry as a whole is finding that larger drives and configurations have increased the risk of
double disk failures per RAID set.
Dell recognizes this fact and uses RAID reliability analysis to help you understand whether your
configuration can withstand multiple hard disk failures.
Question: The term reliability typically refers to a single disk drive or other single component. How
should I understand it?
Answer: Dell uses this term as it relates to the pool and members RAID policy.
Question: How do you determine the RAID policy reliability score for a group with a mixed RAID policy,
especially when it contains one or more pools and members?
Answer: The RAID policy reliability score displayed for a mixed pool or group represents the lowest score
throughout the object. For example, in a two-member mixed pool, where one member with RAID 5 has a
reliability score of 0.8 and another member with RAID 6 has a reliability score of 1.3, the group and pool
RAID policy reliability score would be 0.8. Extending this scenario to a multipool group, where the second
pool (Pool 2) has a member with a RAID policy reliability score of 2.0, then the group's score would
continue to be 0.8 (0.8 for Pool 1 and 2.0 for Pool 2). Assume further that if Pool 2 has a score of 1.0
(instead of 2.0), then the group would have a RAID policy reliability score of 1.0.
1. Select a group from the Servers and Groups tree, then Hardware / Firmware, and then Disks.
The Disks on Group window is displayed.
2. From the drop-down list, select the type of data you want to view:
The following figures show the All Disks IOPS view at the group, pool, and member levels, followed by
the All Disks I/O Rate views.
All Disk IOPS at Member level. The graph in Figure 25. All Disks IOPS at Group Level shows IOPS for reads
and writes for all disks in the group.
NOTE: Firmware currently does not report IOPS at disk level so instead of IOPS this feature uses I/O
Rates. The total IOPS at disk level includes all the reads and writes to all disks. These reads/writes
can be due to either a direct read/write from the host or due to internal RAID realignments/
balancing and such. In the above screens at Group/Pool/Member level, IOPS are getting displayed
which is the sum of reads and writes (inbound and outbound).This cannot be computed at the
individual disk level.
The graph in Figure 26. All disks IOPS at the Pool levelshows IOPS for reads and writes for all disks in the
selected pool.
The graph in Figure 27. All Disks IOPS at the Member level shows IOPS for reads and writes for all disks in
the selected member.
The graph shown in Figure 28. All Disks I/O Rate at Group level shows reads (read I/O rate) and writes
(write I/O rate) for all HDDs and SSDs in the group.
The graph shown in Figure 29. All Disks I/O Rate at Pool level shows reads (read I/O rate) and writes (write
I/O rate) for all HDDs and SSDs in the selected pool.
The graph in Figure 30. All Disks I/O Rate at Member level shows reads (read I/O rate) and writes (write
I/O rate) for all HDDs and SSDs in the selected member.
Overview of the PS Series firmware Volume Undelete feature, including the types of volumes
supported (see About the PS Series Firmware Volume Undelete Feature)
How to access information for recoverable volumes using the SAN HQ GUI (see Display Recoverable
Volumes in SAN Headquarters)
For more information about managing deleted volumes from Group Manager, see the Dell EqualLogic
Group Manager Administrator's Guide.
The SAN HQ GUI visualizes these pending deleted volumes (shown as Recoverable Volumes in the
Servers and Groups tree) and displays the following information:
Manual purge operations are displayed in the audit log. Automatic purge operations performed by the
firmware are logged in the event log.
From the Servers and Groups tree, select a group, then Capacity Recoverable Volumes.
The Recoverable Volumes on Group window appears, showing the total space available for borrowing,
recoverable volume borrowing, and snapshot borrowing.
You can also view information about deleted volumes for all pools on the group by clicking Pools on the
context link bar (see SAN HQ Client GUI Navigation).
Figure 31. Recoverable Volumes Window shows information about recoverable volumes for a group.
In Figure 31. Recoverable Volumes Window, hover over the graph to display the amount of storage space
available for borrowing, the amount of space borrowed for recoverable volumes by the group, and the
amount of space borrowed for snapshots by the group. The pie chart shows estimated percentages.
Below the graph, the General data table shows the following information:
The Original data table shows information about the volume before it was deleted:
Reported Size Size of the volume as seen by iSCSI initiators before it was deleted. For volumes that
are not thin provisioned, the reported size is the same as the volume reserve. For thin-provisioned
volumes, the reported size can be greater than the volume reserve.
Type Original volume type, either Thin for a thin-provisioned volume or Standard for a fully
provisioned volume.
Due to Be Purged Date and time when the volume will be purged automatically, after which the
deleted volume can no longer be recovered. Note that if the group runs low on free space, the
deleted volume might be purged before this date.
Space Borrowing
SAN HQ provides snapshot space borrowing statistics for volumes and replication borrowing statistics for
replicas.
Volume snapshots can borrow from replica reserve and delegated space
Replicas can borrow from other replicas reserves, delegated space, and snapshot reserve
NOTE: The firmware can delete snapshots that are borrowing space when that space is needed for
other functions. Because of this potential deletion, you should always be aware of when snapshots
are borrowing space.
For complete information on snapshot space borrowing and replication borrowing, see the Dell
EqualLogic Group Manager Administrator's Guide.
From the Space Borrowing on Group window, view space borrowing statistics for volumes and
replicas
From the All Outboutn Replicas for Group window, view statistics for replica sets and partners.
From the SAN HQ Top 10 Report Across Groups and SAN HQ Top 10 Report
In Figure 32. Space Borrowing Statistics for a Group, hover over the graph to display the amount of
storage space available for borrowing, the amount of free space in the group, the snapshot reserve space,
and the time period for which the data applies.
The pie charts shows estimated percentages for free space borrowing and snapshot reserve borrowing:
Actual Free Amount of total group capacity that is not reserved, used, or being actively
borrowed
Borrowed Free Amount of unreserved group capacity that is temporarily being used for volumes
that are oversubscribed or in the volume recovery bin
Snapshot Reserve Borrowing
Actual Free Amount of snapshot reserve space that is not used or actively being borrowed
Borrowed Free Amount of snapshot reserve free space that is temporarily being used for
volumes or replicas that are oversubscribed or in the volume recovery bin
The information data table below the graph shows the following information:
For a description of the statistical information in the data table below the graph, see Outbound Replicas.
Compression Overview
Compression maximizes available space within the storage pool by targeting inactive snapshots and
replicas and compressing them to reduce their size. Compressed snapshots and replicas are
decompressed when the array needs to access that information.
Compression is disabled by default on the array. When enabled, the array periodically monitors activity to
determine which snapshots and replicas can be safely compressed and to provide access to
decompressed data.
The firmware presents three main compression statistics that SAN HQ presents for group, member, pool,
and volume views:
Expanded/uncompressed size Amount of snapshot data a volume can hold when fully
uncompressed.
Compressed size Amount of physical disk space consumed.
Space savings Space savings achieved by compression, actual and as a percentage of
uncompressed space.
For complete information about snapshot and replica compression, see the Dell EqualLogic Group
Manager Administrators Guide.
For complete information on activating and using snapshot and replica compression, see the Dell
EqualLogic Group Manager Administrators Guide.
The statistics reported depends on the view and context you selected. Table 26. SAN HQ Compression
Reporting lists the SAN HQ views and contexts showing compression statistics.
To access this information, from the Servers and Groups tree, select All Groups Summaries Volume
Capacity Summary.
Figure 33. Volume Capacity Summary View Showing Compression Statistics shows the All Groups
Volume Capacity Summary view with compression statistics in the Compressed Size, Expanded Size, and
Space Savings columns.
To access this information, select a group from the from the Servers and Groups tree. By default,
information displays on the group level.
NOTE: The compression statistics shown on the group level are calculated from the individual
member(s) in the pool and represent an approximation. The most accurate space saving statistics is
at the pool and member level.
Select the Pools context link bar option (see Context Link Bar Options) to view space savings statistics for
all pools.
Select the Members context link bar option to view space savings for all members.
To access this information, from the Servers and Groups tree, select a group then Capacity.
Select the Pools context link bar option to view space savings statistics for the capacity of all pools on the
group. Select a single pool to view more details: the compression state, compressed size, expanded size,
and space savings. Select an individual pool from the grid in the lower panel to see specific compression
statistics for that pool.
Select the Members context link bar option to view space savings statistics for the capacity of all
members in the group. Select an individual member from the grid in the lower panel to see specific
compression statistics for that member.
Select the Volumes context link bar option to view space savings statistics for the capacity of all volumes
on the group. Select an individual volume from the grid in the lower panel to see the status of
compression on the volumes pool.
Select the Pools context link bar option to view space savings statistics for all inbound replicas of all pools
of the group. Select a member from the context link bar Pools menu to see similar details for all inbound
replicas in that pool. Figure 36. Summary of Pool Statistics Panel Showing Compression Statistics shows
an example.
Select the Members context link bar option to view space savings statistics for all inbound replicas of all
members on the group. Select a pool from the context link bar Pools menu to see similar details for all
inbound replicas on that member.
Select the Volumes context link bar option to view space savings statistics for all inbound replicas of all
volumes on the group. Select an individual replica set from the grid in the lower panel to see the status of
compression on that replica sets pool.
The SAN HQ Server includes a syslog server. If you configure a group to log events to the computer
running the SAN HQ Server, the events appear in the Events tab at the bottom of the SAN HQ GUI and
also when you select Events/Audit Logs in the Servers and Groups tree.
Although data in the SAN HQ GUI is updated only when a successful poll occurs, events continue to be
logged to the syslog server on the computer running the SAN HQ Server, regardless of polling activity.
After messages consume all the free space in the event file, new messages overwrite the oldest
messages. You can change the default event log size only when you first add the group to the monitoring
list. The default size of the event log file is 5MB, the minimum size is 2MB, and the maximum size is
20MB.
If you are using the SAN HQ Server as a syslog server, do not run another syslog server on the
computer running the SAN HQ Server.
A firewall must not block UDP port 514 on the computer running the SAN HQ Server.
Events will appear in the SAN HQ GUI after the SAN HQ Server successfully polls the group. See Syslog
Events.
CAUTION: For groups enabled for SupportAssist, disabling syslog event logging from the group
can potentially disable part of the functionality for SupportAssist data collections. For more
information, see SupportAssist Requirements.
You must change the syslog server configuration using a domain user account on the computer that
is running the SAN HQ Server.
The domain user account must have write privileges to the log file directory.
You must restart the SAN HQ Server to apply the change.
Before changing the syslog configuration, back up or make a copy of the SyslogConfig.xml file that is
located in the log file directory.
To change the syslog server configuration and enable only specific network interfaces for use as listening
UDP sockets:
1. Edit the SyslogConfig.xml file that is located in the log file directory used by the SAN HQ Server. For
example:
<?xml version="1.0"?>
<SyslogConfig xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xmlns:xsd="http://www.w3.org/2001/XMLSchema">
<Enable>true</Enable>
<IPv4Interface>ANY</IPv4Interface>
<IPv6Interface>ANY</IPv6Interface>
<Port>514</Port>
</SyslogConfig>
2. In the <IPv4Interface> tag or the <IPv6Interface> tag, whichever applies to your network
configuration, specify either:
ANY Specifies that any network interface can be used as a listening UDP socket. This option is
the default.
If the SyslogConfig.xml file has a syntax error or some other problem, SAN HQ logs the event to the
Windows event log.
You must disable the SAN HQ syslog server using a domain user account on the computer that is
running the SAN HQ Server.
The user account must have write privileges to the log file directory.
You must restart the SAN HQ Server to apply the change.
1. Edit the SyslogConfig.xml file that is located in the log file directory used by the SAN HQ Server. For
example:
<?xml version="1.0"?>
<SyslogConfig xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xmlns:xsd="http://www.w3.org/2001/XMLSchema">
<Enable>true</Enable>
<IPv4Interface>ANY</IPv4Interface>
<IPv6Interface>ANY</IPv6Interface>
<Port>514</Port>
</SyslogConfig>
2. In the <Enable> tag, specify false:
<Enable>false</Enable>
3. Restart the SAN HQ Server, as described in Restart the SAN Headquarters Server.
NOTE: If the SyslogConfig.xml file has a syntax error or some other problem, SAN HQ logs the
event to the Windows event log.
SNMP Notifications
In SAN HQ version 3.1 and later, you can configure the SAN HQ Server to send SNMP notifications to an
SNMP management console.
This section provides the following information:
SNMP notifications are identified using a unique object identifier (OID). The SNMP manager looks up the
OID in the MIB to determine its placement in a logical hierarchy of all SNMP notifications.
For information about configuring SNMP access to a PS Series group, see the Group Manager online help.
To configure SNMP notifications, you specify the type of notification (trap or inform), the PS Series group
receiving the notification, and notification priorities.
SAN HQ version 3.1 and later provides a MIB file (eqlsanhq.mib) that describes the SNMP notifications.
This file is dependent on the eqlgroup.mib, which is part of the PS Series array MIBs.
The SAN HQ Server creates the AlertNotificationConfig.xml file in the log share path when it
starts. The Server attempts each SNMP notification three times. If all attempts fail for any SNMP
notification, the SNMP management console is ignored for the rest of the polling cycle. This behavior
prevents the retries from delaying other notifications and prevents interference with SNMP polling or alert
processing. Failures are logged in the SAN HQ Servers alertnotifications trace log.
If an alert condition continues to exist, the SAN HQ Server sends follow-up SNMP notifications using the
same interval as email notifications (by default, 6 hours). A single notification is sent per alert instance;
multiple alerts are not combined in a single notification.
Setting Description
Groups Specifies one or more PS Series groups for which SAN HQ Server will send
SNMP notifications.
Origin (Optional) Defines which IP address and port the SANHQ Server binds
locally to send SNMP notifications. Origin includes one address and one
port.
Notification Type (Optional) Defines the type of SNMP notification sent: SNMP-
Inform(default) or SNMP-Trap.
Community Specifies the SNMP community name, sent as part of the notification
message.
Enabled Enables the SNMP Alert Notification. If set to False, the feature is
disabled.
Alerts Includes one settings for each alert priority (Critical, Warning, Caution,
and Information):
All Sends notifications for all alert priorities
None Does not send notifications for any alerts of this priority.
UseEmailSettings Uses the email notification settings for each group
to determine whether or not an SNMP notification will be sent for alert
of this priority.
NOTE: If you choose to load the eqlSanHq.mib into your management console, you must also
load the PS Series firmware MIBs (eqlgroup mib).
1. Click the Settings button in the far-left panel or from the Settings menu.
Table 28. Callouts for SNMP Notifications Settings Figure describes the callouts in Figure 37. SNMP
Notifications Settings Page.
Callout Description
1 SNMP Notifications Settings link Displays the SNMP Notifications Settings page.
2 Send SNMP Notifications Select the checkbox and then define SNMP notification
type: informs or traps. Group selection in the right-hand pane becomes active.
Specify the SNMP community name configured in the PS Series group.
3 SNMP Notification Destinations Enter at least one IPv4/IPv6 address or the DNS
name for the SNMP notification destination. You can specify up to 5 destinations.
Destinations can be:
IPv4 or IPv6 address
hostname
5 Local Binding Optionally, enter a different port to override the source port. The
IPv4/IPv6 address or DNS name corresponds to an SNMP destination defined in
Callout 3. Use the local binding option if you want to force sending notifications to
a specific Ethernet NIC or to satisfy firewall requirements.
Click the download link to open the Windows File Explorer and select the SAN HQ
MIB (eqlSanHq.mib).
6 Groups Select to enable a group or groups; the SAN HQ Server will send SNMP
notifications to the group(s). Use the toggle button to select all groups or deselect
all groups.
1. Access the AlertNotificationConfig.xml file from the log file network share path.
2. Modify the settings for SNMP notifications described in SNMP Notifications Configuration File.
The changes will take effect when you restart the SAN HQ Server.
SNMP destination failed to respond after three attempts during the polling cycle
Connected host failed to respond after a period of time
On the SNMP Notifications Settings page, verify that the IP address or DNS name of the SNMP
destination is correct.
On the SNMP Notifications Settings page, verify that the PS Series group is enabled to receive SNMP
notiifications.
Verify that the PS Series group is in a connected state.
Storage container Storage made available for virtual volumes (VVols). One or more storage
containers can reside in a pool. Dell recommends that you use the Dell Virtual Storage Manager (VSM)
to create storage containers.
Virtual Volume (VVol) Virtual volumes you configure in a storage container. A storage container can
include one or more VVols.
Virtual machine (VM) A virtual system created by VMware vSphere. In EqualLogic environments,
VMs are managed by the Dell Virtual Storage Manager for VMware (VSM), which creates the VM
entities on the array. A virtual volume is mapped to a VM.
For information about Group Manager support for VMware virtual volumes, see the VMware Group
Access Panel chapter in the Dell EqualLogic Group Manager Administrator's Guide.
Table 29. Storage Container Information Provided in the Client GUI shows the location of this information
in the Client GUI.
All Groups Volume All Groups Summary Group/Volume column Shows VVols per
Capacity Summary Volume Capacity Summary group, type of VVol
All Groups Volume IO All Groups Summary Storage container column Shows storage
Summary Volume IO Summary containers in the associated group
Capacity of Group Group in the Servers and Storage Container Reserve panel Total GB
Groups tree and percentage amount of free and in-use
storage container reserve
Unused Space panel Storage container
reserve not used
Space Allocation chart Storage container
reserve allocated in percentage
Unused Space chart Storage container
reserve not in use
Capacity of All Storage Group in the Servers and Graph shows total storage container reserve and
Containers on Group Groups tree Capacity amount in use
Storage Container context
Table lists the storage containers, pool, number
link bar object of VVols, total physical space allocated, physical
space free, total logical space, and logical space
free
Capacity of Storage Group from Servers and Graph shows total capacity and amount of
Container Groups tree Capacity storage container reserve
Storage Container context Storage Container Physical Space panel
link bar object Select a Amount of storage container physical space
that is free and currently in use.
storage container
Storage Container Logical Space panel
Amount of storage container logical space
that is free and currently in use.
VVols and Snapshots Total number of
virtual volumes in the storage container and
snapshots
Combined Graphs of Group in the Servers and Combined information for network and I/O
All Storage Containers Groups tree Combined activity, capacity information, and total counts for
on Group Graphs Storage iSCSI connections, snapshots, and replicas for all
Containers context link bar storage containers on the group
object
Combined Graphs of Group in the Servers and Combined information for network and I/O
Storage Container Groups tree Combined activity, capacity information, and total counts for
Graphs Storage iSCSI connections, snapshots and replicas for the
Containers context link bar selected storage container
Table 30. Virtual Volume Information Provided in the Client GUI shows the location of this information in
the Client GUI.
All Groups Volume Capacity All Groups Summary Type column Shows type of VVol
Summary Volume Capacity Summary (Data,Swap, and Config)
All Groups Volume IO Summary All Groups Summary Type column Shows type of VVol
Volume IO Summary (Data,Swap, and Config)
Capacity of All VVols on Group Group in the Servers and Graph shows total storage container
Groups tree Capacity reserve and total VVol space used
VVols context link bar object
Table lists the virtual volumes, VVol
type, storage container in which it
resides, reported size of the VVol,
status (for example, available), and
number of snapshots of the VVol
Capacity of VVol Group in the Servers and Description panel Name and type
Groups tree Capacity of the VVol, storage container in
which it resides, and status
VVols context link bar object
Select a VVol Details panel Size of the virtual
volume and number of snapshots of
the VVol
Combined Graphs of All VVols Group in the Servers and Combined information for network
on Group Groups tree Combined and I/O activity, capacity
Graphs VVols context link information, and total counts for
bar object iSCSI connections, snapshots, and
replicas for all VVols on the group
Combined Graphs of VM Group in the Servers and Combined information for network
Groups tree Combined and I/O activity, capacity
information, and total counts for
Live View Session Group in the Servers and Specify VVol as target and an
Groups tree IO Live View individual VVol, and then run a Live
Sessions View session to obtain I/O data
captured in brief session intervals at
quick polling rates. For more
information on Live View sessions,
see Live View Data.
Name of the VM, guest operating system installed on the VM, total space allocated for the VM, and
description of the VM
List of connected VVols
Capacity information including reported size of the VM
Aggregate capacity information for mapped VVols
Aggregate I/O performance data for mapped VVols
Table 31. Virtual Machine Information Provided in the Client GUI shows the location of this information in
the Client GUI.
Capacity of VM Group in the Servers and Graph shows total storage container
Groups tree Capacity VMs reserve and total space used by the
VM
context link bar object
Select a VM Table shows the name of the VM
and associated VVols, type of VVol,
storage container in which the VVols
reside, status, number of snapshots,
reported size of the VM and
individual VVols
Combined Graphs of All VMs Group in the Servers and Combined information for network
on Group Groups tree Combined and I/O activity, capacity
Graphs VMs context link bar information, and total counts for
object
Combined Graphs of VM Group in the Servers and Combined information for network
Groups tree Combined and I/O activity, capacity
Graphs VMs context link bar information, and total counts for
object Select a VM iSCSI connections, snapshots and
replicas for the selected VM
SAN HQ reports the aggregate of delegated space from all pools on the replication partner to the primary
group. Figure 38. Delegated Space for Replication in Multiple Pools illustrates this concept.
Callout Description
1 Primary group with two pools has replication established to the secondary group
(replication partner). Replication is considered outbound from the perspective of the
primary group. The primary group must have enough free space for the local
replication reserve for each replicated volume. Local replication reserve is located in
the same pool as the volume.
2 Secondary group with two pools configured as the replication partner. The
secondary group administrator allocates space across both pools. Replication is
considered inbound from the perspective of the secondary group. The secondary
group must have enough free space to delegate to the primary group. For groups
running firmware v8.0 and later, delegated space can be across multiple pools. Both
primary and secondary groups must be running compatible firmware versions.
3 Primary group contains two pools with two volumes each.
4 Delegated space in the Default pool contains two replica sets.
5 Delegated space in Pool B contains four replica sets.
In SAN HQ 3.1 and later, the values shown for delegated space and unused space in the Capacity of
Group view show the sum of delegated space from all pools (see Delegated Space in Capacity of Group
View).
These values are also reflected in the delegated space statistics on the All Groups Capacity Summary view
(see Capacity Summary ).
The Space Allocation chart (see Figure 40. Space Allocation Chart) shows the aggregate delegated space
from multiple pools from their respective replication sites.
1. Create an archive of the latest group data. See Group Data Archives.
2. Either:
Stop monitoring the group but retain the group log files. You can still view the group data in the
SAN HQ GUI, but no new data will be collected for the group at the obsolete network address.
See Stopping Group Monitoring. Note that the data associated with the obsolete network address
still appears in the GUI under the group name; however, the monitoring status is disconnected. If
you try to resume monitoring the group using the obsolete network address, the operation will
fail.
Remove the group from the monitor list. The log files are deleted and the GUI no longer shows
group data. See Removing a Group from the Monitor List.
3. Add the group to the list of monitored groups, specifying the new network address. See Add a Group
to the Monitoring List from the SAN HQ GUI.
Cannot connect to a group Make sure an SNMP community name is configured in the group and
the same name is configured in the SAN HQ GUI for that group.
In addition, for each monitored group, the computer running the SAN HQ Server must have network
access to all the configured network interfaces on all the group members, the group IP address, and
the management address (if applicable). Use the ping command to determine if you can access all IP
addresses.
Cannot add a group Make sure the computer running the SAN HQ GUI has read and write access to
the log file directory.
Cannot set up email notification or change the SNMP community name Make sure the computer
running the SAN HQ GUI has read and write access to the log file directory.
Each set of log files can contain up to one years worth of data. After a year, the SAN HQ Server
overwrites the oldest data.
By compressing data over time, the SAN HQ Server can efficiently store a large amount of data and also
reduce data volatility (by moderating extreme values). However, the SAN HQ Server calculates a data
point by averaging data from consecutive polls; therefore, old data shown in the SAN HQ GUI might not
be as precise as new data.
NOTE: Using a log file size that is larger than the default size of 5MB enables you to store more-
precise data; however, a larger log file size might have a slightly negative impact on response time.
If you use a log file size that is smaller than the default size, data will be less precise but response
time might improve.
Table 33. Compression Impact on Performance Data shows how compression over time affects different
types of performance data. For example, some data is understated because intermittent idle time in
typical workloads decreases averages. See Performance and Capacity Terms for a description of the
terms used in .
For related information about increasing the size of the log file, see Increase the Log File Size.
SAN HQ updates the data in the GUI only if a polling operation succeeds. Unsuccessful SNMP polls can
result from network problems or the group workload. For example, if a group is busy processing I/O
operations, it might drop SNMP requests.
If the Monitor Service determines that a group is not responding to SNMP requests, it reviews the results
of each poll to determine if the polling interval needs to expand, contract, or remain the same. This
process uses historical evidence of a group's poll responsiveness to determine the optimal frequency for
the group and then makes adjustments accordingly. The Monitor Service can automatically shift among a
list of polling frequencies, ranging from 2 minutes to 24 hours.
The SAN HQ GUI displays the status of the group polling activity (Last Poll Status) at the top of the
window. See Polling Status.
You can determine the polling period used to obtain a data point by placing the cursor over a point in
time on a SAN HQ graph.
Depending on the SAN HQ Client GUI view you are using, this data can differ significantly.
For example, in Figure 44. TCP Packet Retransmissions in Multiserver Environment, Server A and Server B
are monitoring the same group and currently show the same TCP retransmit percentage. Both servers
can potentially acquire data every 15 minutes. At the first retransmit interval (T1), only Server B acquires
the data. Server A acquires data at T2, or 30 minutes since its last polling cycle. Server B then acquires the
data at T3, or 30 minutes since its last polling cycle. Server B acquires data again at the regular 15-minute
TCP retransmit interval (T4). Server A must wait until T5, which is 45 minutes since it last acquired data. In
this scenario, the retransmission rates are significantly different between Server A and Server B.
A sustained high TCP retransmit rate (greater than 1%) might indicate a network hardware failure,
insufficient server resources, or insufficient network bandwidth. Dell recommends that you evaluate your
network for any TCP retransmit rates greater than 1%.
NOTE: Dell recommends that you not use multiple SAN HQ servers to monitor the same group.
Performance and capacity dataTo analyze the group data in the GUI, you must understand the
terminology. See Performance and Capacity Terms.
Polling statusGetting a complete set of data points in the GUI depends on how frequently SAN HQ
can successfully poll the group. If polls are not successful, you should investigate the cause. See
Polling Status.
AlertsAlerts notify you when hardware or performance issues occur in a group so that you can take
action to prevent problems. See Reported Alerts.
Events and audit logsIf you configure a group to use the syslog server that is part of the SAN HQ
Server, you are notified when group events occur so that you can take action to prevent problems.
See Syslog Events and Audit Messages.
NOTE: When displaying group data, make sure you select the correct time range. By default, the
GUI displays data from the most recent 8-hour time period. Use the Zoom links above the
timeline to quickly set the value of the time-range selector and also to control the range of
dates seen in the timeline. For example, click Show Latest to show data up to the most recent
time. See SAN HQ Client GUI Navigation.
NOTE: Tooltips are associated with the data that appears in the SAN HQ GUI. To obtain information
about the data, move the pointer over a graph legend or the question mark icon next to a table title.
See the Dell EqualLogic Group Manager Administrators Guide for detailed information about the status
that appears in the SAN HQ GUI, in addition to information about PS Series group operation.
The SAN HQ GUI provides the following capacity and replication terms and space statistics:
Delegated spaceSpace on the secondary group reserved for storing all replica sets from the primary
group. If free delegated space is low, you might want to increase it to ensure space is available for
Local replication reserveSpace on the primary group reserved for storing volume changes during
replication and the failback snapshot.
Overall group capacityAvailable space. Capacity depends on a number of variables. For example,
group capacity depends on the number of members, the number and size of the disks installed in the
members, and each member's RAID policy.
Make sure free space in a pool does not fall below the lesser of these values:
5% of pool capacity
100GB times the number of pool members Otherwise, load balancing, member removal, and
replication operations do not perform optimally, and performance for thin-provisioned volumes
degrades.
Recoverable volumesPending deleted volumes that the firmware places in the recovery bin that can
be restored. SAN HQ shows the name of the volume, the pool in which it resides, the amount of
space borrowed by the recoverable volume, the date when the volume was deleted, the size of the
volume, the type of volume, and date and time when the volume will be purged from the recovery
bin.
Replication partnerList of all groups currently configured as an outbound replication partner.
Replica reservePortion of delegated space reserved for storing the replica set for a volume. After
replica reserve has been consumed, the oldest replicas are deleted to free space for new replicas. To
retain more replicas, increase the replica reserve percentage.
Reported volume sizeVolume size seen by iSCSI initiators.
Snapshot reserveSpace reserved for storing snapshots. After snapshot reserve has been consumed,
the oldest snapshots are deleted to free space for new snapshots. To retain more snapshots, increase
the snapshot reserve percentage.
Space borrowingAbility to temporarily increase the available snapshot space for a volume by
borrowing from the snapshot reserve of other volumes and from pool free space. This temporary
increase enables you to prevent the oldest snapshots in your collection from being deleted
automatically.
Thin provisioning statisticsNumber of volumes that are thin provisioned, the amount of unreserved
(unallocated) space for thin-provisioned volumes, and the percentage of group space required to
fulfill the maximum in-use space requirements for thin-provisioned volumes.
Volume reserveSpace allocated to a volume. For thin-provisioned volumes, the volume reserve is
based on usage patterns. As more data is written to the volume, more space is allocated to the
volume, up to the user-defined limit.
Volume typeType of volume: template, thin-provisioned, thin clone, or standard (a fully provisioned
volume).
Average I/O rate Average data transfer rate (also called I/O throughput). This rate is the average
amount of data that is transferred each second.
Usually, the I/O rate for reads and writes is not a significant indicator of performance. All storage
systems have a maximum throughput capacity. Because most I/O is random and not sequential,
storage systems rarely reach this threshold. If the threshold is reached, it indicates a sequential
workload.
NOTE: As data is compressed over time, the I/O rate in the GUI becomes less precise.
Average IOPS Average number of I/O operations processed each second. The GUI displays data for
all Ethernet activity, including iSCSI traffic and SAN HQ SNMP polling.
NOTE: As data is compressed over time, the IOPS values shown in the GUI become less precise.
Average latency Average time required to process and complete an I/O operation. Latency (also
called delay) is the best gauge for measuring the storage load and is the principal method for
determining if a group has reached its full capabilities.
In SAN HQ, latency is measured from the time the group acknowledges an I/O request to the time the
group completes the I/O operation. Latency is measured in milliseconds and is reported as an average
for the I/O operations in a polling period.
Latency that occurs in the server is not included in the SAN HQ data.
NOTE: Increasing the period of time over which latencies are averaged does not make the
latency data in the GUI less precise. While some volatility is lost, idle time does not affect the
average latency. Therefore, older latency data is still a good indicator of performance.
See Identifying Performance Problems for information about interpreting latency values.
Average I/O size Average I/O operation size. The size of the I/O operations can help you obtain a
better understanding of your applications and workload. For example, a workload that consists of
many small, random I/O operations will have different performance characteristics than a workload
with large, sequential I/O operations.
NOTE: Increasing the period of time over which I/O sizes are averaged does not make the data
in the GUI less precise. While some volatility is lost, idle time does not affect the average I/O size.
Therefore, older I/O size data is still a good indicator of the workload.
Distribution of read and write IOPS Percentage of IOPS that are reads and the percentage of IOPS
that are writes in the group. The read and write percentages are not an indicator of performance.
However, this information is important when sizing and configuring groups for specific workloads.
For example, certain RAID configurations perform better for a certain read/write distribution. In
general, RAID 50 and RAID 5 do not perform as well as RAID 10 for workloads consisting
predominantly of random writes. If latencies for a specific member suggest a performance problem,
and the member performs more than 70% of write I/O operations, moving the random write workload
to a pool with a RAID 10 member might solve the problem.
Network Terms
In general, network data measures all network traffic, including iSCSI operations, GUI operations, SNMP
requests, and replication operations.
The SAN HQ GUI provides the following network statistics:
Polling Status
The SAN HQ Server regularly uses SNMP to poll a group to obtain data.
By default, the polling periodthe time between consecutive polling operationsis 2 minutes. You can
capture data in shorter polling intervals, as brief as 1 second, by using the Live View feature. See Live View
Data.
If a group is busy processing I/O operations, it might drop SNMP requests from the SAN HQ Server. If the
server determines that a group is not responding to SNMP requests, it automatically adjusts the polling
frequency based on historical evidence for the group until a poll succeeds. See How Group Performance
Affects SNMP Polling.
The SAN HQ GUI displays the status of the group polling activity (Last Poll Status) at the top of the
window.
Status Description
Successful SNMP poll was successful.
Increasing polling Group performance or a network problem prevented the group from responding
period to SNMP requests in a timely manner. SAN HQ automatically adjusts the polling
frequency based on historical evidence for the group until a poll succeeds.
Failed SAN HQ cannot contact the group. In this case, an alert describing the problem
is generated.
Member rebooted A member rebooted. A successful SNMP poll is required to obtain up-to-date
group information.
Reported Alerts
Alerts enable you to be quickly informed of problems so you can diagnose and correct them.
SAN HQ displays two types of alerts:
Performance-related alerts detected by SAN HQ (for example, low free pool space or high latency)
Some alerts have an increasing priority, as the condition increases in severity. shows a list of SAN HQ
alerts.
Hardware alarms detected by the group (for example, high temperature or a failed control module)
Hardware alarms depend on the PS Series firmware version and also the member hardware. See the
Dell EqualLogic Group Manager Administrators Guide for a list of hardware alarms.
In some cases, a SAN HQ alert and a hardware alarm might be generated for the same group event.
Optionally, you can configure email notification for alerts. See Configure Email Notification for Group
Alerts.
Informational Normal, operational events in the group that do not require any administrator action.
Informational alerts are unique to SAN HQ.
Caution Low-level conditions that, if not addressed, might lead to performance issues or unwanted
results. Caution alerts are unique to SAN HQ.
Warning Conditions that will affect group operation unless addressed immediately. Warning alerts
correspond to Warning events and Warning alarms in the group.
Critical Serious problems that are currently affecting group operation. Critical alerts correspond to
Error events and Critical alarms in the group.
Alerts can have multiple priority levels, given the condition that triggered the alert. For example, the Thin
Provision Volume Space alert can generate a caution, warning, or critical alert, depending on how close
the group comes to exceeding the maximum in-use space threshold.
Further, when email notification is enabled for a group, SAN HQ by default sends email notification for
only certain priority levels. In the example of the Thin Provision Volume Space alert, email notification is
sent only for warning and critical severity levels. Email notification is not automatically sent for all alerts
generated on the monitored groups. You can manually set email notifications for other alert priority levels
from the Email Settings page. For information, see Configure Email Notification for Group Alerts.
NOTE: For some alerts (for example, the Snapshot Reserve alert), SAN HQ's default email
notification settings might be different in earlier versions of SAN HQ. The current version might not
generate the expected email notification unless you manually specify notification for that alert
priority level. If an issue that generates an alert is resolved within a data polling period, email
notification will not occur.
Display Alerts
Alerts appear in the bottom panel of the All Groups Monitored by Server window (active alerts only) or
under the Alerts tab at the bottom of the GUI windows.
To display alerts:
Column Description
Priority Shows the priority of the alert, either Informational, Caution, Warning or Critical
Alert ID An assigned numeric value based on different Alert classes implemented in SAN
HQ. This should display the same number as is present in the email configuration
settings tabs
Estimated Time The time stamp of estimated first occurrence of the alert when it was seen as
Detected Active for selected time range
Duration The duration for which the respective alert was Active in synchronization with the
Estimated Time Detected
Status The status of the alert for selected time range. If, at the end of selected time range,
the alert is active, it will show it as Active, else the status will be Cleared
The following sample data is extracted from the Alerts Panel and provides additional explanations and
examples of the information in the panel.
Example 1:
In Alert Example 1, the top alerts Estimated Time Detected is shown as 04/01/2016 15:41 and duration as
2 m 15s and status as Cleared. This implies that the alert was raised on 04/01/2016 15:41 and was active
for 2 minutes and 15 seconds before getting cleared.
Example 2:
In Alert Example 2 the top alerts Estimated Time Detected is shown as 03/30/2016 11:50 , the Duration as
2 d 4 hr 9 m, and the status as Active. This implies that this alert has been raised on 03/30/2016 11:50 and
has been active since 2 days 4 hours and 9 minutes ago.
The lower alerts Estimated Time Detected is shown as 03/31/2016 12:30, with a Duration of 1 d 3 hr 29 m
and the status as Active. This implies that this alert has been raised on 03/31/2016 12:30 and has been
active since 1 day 3 hours and 29 minutes ago.
Example 3:
In Alert Example 3 the Estimated Time Detected is shown as in Figure 49 04/05/2016 09:15 and duration
as 1 d 2hr 28m. Upon further observation, we see that its Estimated Time is the same as the beginning of
the selected time range (though it is showing a 4 min difference, which is explained below). This implies
that there might be two situations here; either this alert has been raised at the given estimated time, or it
has been raised before the selected time range.
NOTE: In Alert Example 3 the Estimated Time Detected [1d 2 hr 28 min] is more than the beginning
of the selected time range[1d 2 hr 24 min]; this is because the system was not polled exactly at the
beginning of the selected time range. Therefore, it went one poll before the beginning of the
selected time range and considers that time to be the start of that particular alert. This is an
expected behavior. The time difference between the Estimated Time Detected and the beginning of
the selected time range should not be more than the polling interval at the beginning of selected
time range.
Based on the SAN HQ existing framework, the alert list will display only the alerts which were active at
least one time in the selected time range. If any alerts were raised before the beginning time of the
selected time range, it will display the Estimated Time Detected of the alert as the same as the selected
time range (with the time difference as explained above in few cases). Therefore, if the Estimated Time
Detected of an alert is same as the beginning time of the selected time range, there might be either of
two situations.
Duration is the time period through which an alert was active in a selected time range, starting from its
estimated time detected.
There might be cases where an alert might have more than one presence in a selected time range
because it got raised, cleared at some poll, then raised again. This type of activity could occur multiple
times.
As shown in Alert Example 4, there are some alerts whose alert ID is seen as 14.x. These are systems alerts
which will not calculate duration. These alerts will eventually vanish from the list as the alert issue
resolves.
Some alert descriptions might contain some criteria message inside square brackets [] . This implies that
when the alert was last seen as active, it met the criteria given in the brackets. In Alert Example 5, this
pool has 37.1 % free space which met the criteria of <= 50% at the last poll when it was seen active. Note
that the criteria might be different on the first estimated time detected (or in between) during which the
alert was active. It will display the criteria [<= 50%] which was found on the last poll of it in an active state
as it displays its value (37.1 %) found on last poll as active.
The square brackets [] also imply that an alert has customizable threshold settings.
Example 5:
You might observe some alert description which will not include square brackets [] inside its description,
as seen in Alert Example 6. This implies that the alert does not have customizable threshold settings and
that the meeting criteria for this alert is fixed.
Export Alerts
You can export alerts to an .xls file.
To export alerts:
1. Click the Export Alerts icon at the top of the Alerts panel.
2. Enter a file name.
In the Alerts panel, click the Copy Selected Alerts to Clipboard icon or the Copy All Alerts to
Clipboard icon.
Table 36. SAN HQ Alerts shows a list of SAN HQs alerts. See the Dell EqualLogic Group Manager
Administrators Guide for a list of hardware alarms.
Connection failure 14.2 Warning A SAN HQ Server SNMP request to the group
failed.
Controller not 3.3 Warning A member control module has failed or is not
detected installed.
Critical member 2.5 Critical A member has a RAID policy reliability that is
RAID policy rated as poor. Dell recommends changing the
reliability RAID policy.
Critical pool RAID 3.31 Critical The pool has a RAID policy reliability that is rated
policy reliability as poor. Dell recommends changing the RAID
policy.
DNS resolution 14.8 Critical The DNS name for the group network address
failure cannot be resolved by the server running the
SAN HQ Server.
Disk not found 3.19 Warning A member is missing a disk drive, based on the
model number, the number of disk slots, and the
standard configurations (14, 16, 24 or 48 disk
drives).
Disk status 3.13 Warning A disk has a status other than online or spare and
requires administrator attention.
Group firmware 3.8 Critical A groups firmware is incompatible with the disks
and member disk firmware.
incompatibility
Group mixed 1.2 Informational, Group members are running different firmware
firmware Warning versions.
Group non-WKA 1.6 Warning A group has been added using the non-Well
address Known Address (WKA) address.
ICMP ping to group 14.11 Critical The SAN HQ Server was unable to contact a
failed group, so it canceled the poll.
Incomplete SNMP 14.3 Warning The SNMP request was not complete, making
poll the poll unusable.
Low free member 3.32 Caution, Warning, A members free space is less than the
space Critical recommended minimum.
Maximum replicas 4.8 Warning Replication failed because the partner reached
and snapshots the maximum number of replicas and snapshots.
Member added 1.4 Warning A member has been added to the group.
Member disk added 3.6 Caution A disk has been added to a member.
Member disk 3.11 Caution A disk has been removed from a member.
removed
Member disk 3.7 Critical A disk has been detected with out-of-date
firmware out of firmware.
date
Member disk RPM 3.12 Warning A member contains drives with mismatched
mismatch RPMs.
Member firmware 3.16 Informational A member has been upgraded with new
upgrade firmware.
Member firmware 3.17 Informational The firmware on a control module has been
upgrade reboot upgraded.1
pending
Member health 3.18 Warning, Critical A health condition exists, and is probably related
status to a hardware failure.
Member network 5.3 Warning, Critical Send, receive, or both send and receive traffic for
port load the network interface is approaching the caution
(80% load), warning (90% load), or critical
threshold (99% load).1
Member removed 1.5 Caution A member was removed from the group.
Member spare disk 3.25 Critical A member's spare drive is mismatched from
mismatch active drives.
Network down 14.10 Critical The network connected to the server running
the SAN HQ Server is down.
Partial SED pool 2.3 Warning A pool contains some members with self-
encrypting disks (SED) and another member with
non-self-encrypting disks.
Pool performance 2.2 Caution, Warning, A pool's performance and capacity (storage
and capacity alert Critical containers capacity, pool capacity) has
exceeded the defined threshold values. Dell
recommends that free pool space not fall below
the following value, whichever is smaller:
5% of the total pool space
100GB multiplied by the number of pool
members
Port at reduced 5.4 Informational A network interface, other than one dedicated to
speed a management network, is connected to a
network device with a speed of less than 1GB.
RAID set verifying 3.29 Caution A member's RAID set is verifying parity data.
Read-only account 1.14 Critical The credentials for the SupportAssist read-only
authorization account failed authorization.
failure for
SupportAssist
Replication partner 4.12 Warning The replication failed because the secondary
disallow group does not have downgrades disallowed.
downgrades
Replication partner 4.6 Warning The replication failed because the partner could
not available not be reached.
Sending SNMP 14.14 Warning The SAN HQ Server failed to send an SNMP
Notification error notification for the corresponding group.
SNMP poll 14.4 Informational A previously failed SNMP poll connection is now
connection successful.
restored
SNMP Notification 14.13 Warning The SAN HQ Server is unable to process the
Configuration SNMP notification configuration and could not
read the AlertNotificationConfig.xml file.
(All groups are notified.)
SNMP Notification 14.14 Warning The SAN HQ Server encountered a failure when
Failure sending an SNMP notification.
Storage Container 2.6 Caution, Warning, Low free storage container space was detected.
performance and Critical The alert will not generate until you provide a
capacity threshold value. Dell recommends that you set a
warning alert threshold value to provide
adequate time to allocate space for your storage
containers.
SupportAssist 1.17 Warning, Critical The most recent diagnostic data upload attempt
Uploads Failing has failed or is retrying over a period of at least
two weeks.
TCP retransmit 3.28 Caution, Warning, A member's ratio of TCP retransmits to sent
Critical packets is too high, indicating a network
problem. The threshold ranges from 1%
(caution), to 5% (warning), to 10% (critical). 1
Volume RAID 4.3 Caution, Warning A problem occurred with allocating volume
preference space according to the desired RAID preference.
For example:
A volume cannot be moved to a member
with the volumes preferred RAID level
because not enough member free space is
available.
Volume replication 4.9 Warning Replication failed because the partner is not
partner needs running the correct firmware and must be
upgrade upgraded.
Volume replication 4.11 Caution Replication of a volume was paused from the
paused primary group.
Volume replication 4.13 Caution Replication of a volume was paused from the
remote paused secondary group.
VVol performance 4.19 Caution, Warning, A virtual volumes performance and capacity
and capacity Alert Critical (Average Latency, Average IOPS, Average I/O
Size, Read IOPS, Write IOPS, and Estimated Read/
Write I/O Load) threshold exceeds the defined
threshold values.
Volume sync 4.17 Caution The SyncActive pool and SyncAlternate pool do
replication not contain the same volume data.
For certain performance alerts, SAN HQ lets you specify a value for threshold criteria. When a threshold
criteria is met, SAN HQ will generate an alert. Initially, default values for some threshold criteria appear
blank (see Table 37. Alerts With Threshold Criteria ) and the alert will appear unchecked (not enabled for
notification). If you specify a value for an alert criteria and later upgrade SAN HQ, these values are
retained. See the Dell EqualLogic SAN Headquarters Release Notes Version 3.2 for more information.
You can set values for alert threshold criteria at different alert priorities: Caution, Warning, and Critical.
When you set a value for a threshold criteria, be sure to check other priority levels for previously-set
values.
When a threshold is reached, SAN HQ generates an alert with a Caution, Warning, or Critical priority after
comparing the threshold value you set to the parameter (X) returned by the array, as follows:
For alerts where you set incremental values, such as Volume In-Use Space criteria (seeTable 37. Alerts
With Threshold Criteria) , SAN HQ generates the alert priority as follows:
For example, assume you set Low Free Member Space criteria with threshold values of 30% (caution),
20% (warning), and 10% (critical). SAN HQ generates alerts as follows:
Table 37. Alerts With Threshold Criteria shows the alerts with threshold criteria for which you can specify
a value. For threshold criteria that are blank by default, SAN HQ generates an alert only after you specify a
value and when the threshold criteria is met.
Snapshot reserve1, 2
Average latency1
Average IOPS1
Read IOPS1
Write IOPS1
Member Network Port Load 5.3 Network load percentage Default 80% (caution), 90%
(warning) 99% (critical)3
Pool performance and 2.2 Low free pool space Default 15% (caution), 10%
capacity (warning), 5% (critical)
1No default value for this criteria (threshold value displays as blank). You must enter a value for SAN HQ
to generate an alert when the threshold is met.
2 By default, the Group Manager uses a delete oldest snapshot policy for newly created volumes and
manages space accordingly. A 4.2 alert is not generated in this case. However, if the volume uses a "Set
volume offline" policy, a 4.2 alert will be generated for that volume when the snapshot reserve
threshold criteria is met.
To set threshold values for SAN HQ alerts, see Set Alert Thresholds. For a complete listing of all SAN HQ
alerts, see SAN Headquarters Alerts.
For example, Figure 52. Alert Configuration Tab for Volume Performance and Capacity Alert shows the
criteria for the Volume Performance and Capacity alert when expanded.
When you exit the SAN HQ Client and later restart it, the threshold criteria is preserved.
Syslog Events
The SAN HQ Server includes a syslog server. If you use Group Manager to configure a monitored PS
Series group to log events to the SAN HQ Server syslog server, the SAN HQ GUI displays the events.
NOTE: SAN HQ must successfully poll a group before the group can log events to the syslog server.
Event Priorities
Table 38. Event Priorities lists event priorities in order of lowest (least severe) to highest (most severe)
priority.
Priority Description
Info Informational message Indicates an operational or transitional event that requires no
action.
Warning Potential problem Can become an event with Error priority if administrator
intervention does not occur.
Error Serious failure Identify and correct the problem as soon as possible.
Display Events
You can display events that a group logs to the SAN HQ Servers syslog server. From the SAN HQ GUI,
either:
Click the Events tab at the bottom of the GUI window to open the Events panel (Figure 53. Events
Panel).
Click Events/Audit Logs in the Servers and Groups tree to display the Events and Audit Logs of Group
window (Figure 54. Events and Audit Logs Window Showing Events). By default, the Show All button is
selected. To show only event logs, select the Show Event Logs only button. To show only audit logs,
select Show Audit Logs only (see Display Audit Logs).
If no events appear, instructions are provided to verify that the group is properly configured to send
events and audit logs to the syslog server on the SAN HQ Server.
Click a column heading to sort according to the column data. (The Account column appears only when
audit logs are shown. See Display Audit Logs.)
Events that appear in the SAN HQ GUI can include events that occurred after the most recent poll or
while the group was not responding to SNMP requests. SAN HQ displays events that occurred within the
selected time period. To display the latest events, select Show Latest in the GUI window.
Search Events
You can search the event log for events that include a specific word, words, or text string. You can also
use the Filter Editor for advanced search capabilities.
To display events that include a specific word, words, or text string, in the Events panel or the Events
window:
1. Enter the text in the search field and click Search. Click Clear to return to the original event display.
2. For advanced search capabilities, click Filter Editor. The Filter Editor dialog box appears.
The Filter Editor enables you to set up a complex search algorithm:
Export Events
You can export the event log to an .xls file.
In the Events panel or Events window:
Audit Messages
Audit messages are syslog events about administrator actions. They provide a historical reference to
actions such as logging in, logging out, creating a volume, setting up replication, and so on.
The SAN HQ Server includes a syslog server. If you use Group Manager to configure a monitored PS
Series group to send audit logs to the SAN HQ Server syslog server, the SAN HQ GUI displays the
information.
Click the Audit Logs tab at the bottom of the GUI window to open the Audit Logs panel (Figure 55.
Audit Logs Panel).
Click Events/Audit Logs in the Servers and Groups tree to display the Events and Audit Logs of Group
window (Figure 56. Events and Audit Logs Window Showing Audit Logs). By default, the Show All
button is selected. To show only audit logs, select the Show Audit Logs only button. To show only
events, select Show Event Logs only (see Display Events).
If no audit logs are displayed, instructions are provided to verify that the group is properly configured to
send events and audit logs to the syslog server on the SAN HQ Server.
Events that appear in the SAN HQ GUI can include events that occurred after the most recent poll or
while the group was not responding to SNMP requests. SAN HQ displays events that occurred within the
selected time period. To display the latest events, select Show Latest in the GUI window.
Figure 56. Events and Audit Logs Window Showing Audit Logs
Search Audits
You can search the audit log for audit messages containing a specific word, words, or text string. You can
also use the Filter Editor for advanced search capabilities.
To display audits that include a specific word, words, or text string, in the Audit Logs panel or the Events
and Audit Logs window:
1. Enter the text in the search field and click Search. Click Clear to return the original audit display.
2. For advanced search capabilities, click Filter Editor. The Filter Editor dialog box appears.
The Filter Editor enables you to set up a complex search algorithm:
For an example of how the Filter Editor works, see Search Events.
1. Click the Export Audit Logs to Excel icon in the Audit Logs panel.
2. Enter a file name.
3. Click Save.
All audit logs for a group: click the Copy Current Audit Logs to Clipboard icon
Selected audit logs for a group: click the Copy Selected Audit Logs to Clipboard icon
How group data is compressed and therefore how it might be less precise as it ages. See How Data Is
Compressed in Log Files.
Basic performance terminology and the type of data collected by SAN HQ. See Performance and
Capacity Terms.
The areas in your environment that can be sources of performance problems and which areas SAN
HQ monitors. See Potential Sources of Performance Problems.
How your applications utilize group storage resources. See Application Storage Utilization.
NOTE: When displaying group data, make sure you select the correct time range. By default, the
GUI graphs display data from the most recent 8-hour time period and the GUI tables display
data from the most recent poll.
Use the Zoom links above the timeline to quickly set the value of the time-range selector and also to
control the range of dates seen in the timeline. For example, click Show Latest to show data up to the
most recent time. See Displaying Data from Different Times.
HardwarePoor performance can be the result of a hardware failure in the group (for example, a disk
failure), the network, or the server.
Network configurationAlthough network bandwidth is rarely fully utilized, the network can be a
source of performance problems. For example, some parts of the network might not be Gigabit
Ethernet, switches might not be properly configured, or interswitch links might not have sufficient
bandwidth.
Servers and applicationsServers that do not have sufficient resources (CPU, memory, bus) can
experience performance problems. Also, applications might not be properly configured.
However, SAN HQ tracks only a portion of the storage stack through which an I/O operation must pass,
starting with the application I/O request and ending with the data retrieved from the group. Latencies
reported by SAN HQ do not include latencies that occur in the server. To fully diagnose non-group
problems, you must use additional tools.
Different applications and workloads result in different performance profiles. Statistics that might indicate
a performance problem in one environment might indicate an efficient use of storage resources in
another.
For example, if your application normally generates large I/O operations, a relatively high average latency
might be expected and does not necessarily indicate poor performance.
To characterize how your applications utilize storage resources, you should understand:
Contact your PS Series support provider or your application support provider for more information about
characterizing your application storage utilization.
Figure 57. Analyzing SAN Headquarters Data describes the process for analyzing SAN HQ data.
Hardware Problems
Hardware failures are a common cause of performance problems and must be corrected immediately.
Always check for the following hardware-related issues in the SAN HQ GUI:
Hardware alertsCheck the Alerts panel for hardware problems that might affect performance, such
as a failed disk or a network connection that is not Gigabit Ethernet. See Reported Alerts for a list of
SAN HQ alerts.
Network retransmissionsA sustained high TCP retransmit rate (greater than 1%) might indicate a
network hardware failure, insufficient server resources, or insufficient network bandwidth.
RAID statusA degraded, reconstructing, or verifying RAID set might adversely affect performance. In
some cases, performance might return to normal after the operation completes.
Low pool capacityMake sure free space in each pool does not fall below the lesser of the following
values:
Otherwise, load balancing, member removal, and replication operations do not perform optimally.
Low free space will also negatively affect the performance of thin-provisioned volumes.
Correct hardware problems immediately. See Best Practices for Solving Performance Problems for more
information.
After you fix a hardware problem, allow time for SAN HQ to collect new data before analyzing the data.
Average I/O LatencyLatency must be considered along with the average I/O size, because large I/O
operations take longer to process than small I/O operations.
Consider the context of time for latency-sensitive applications when viewing metrics. For example,
high latency might occur without business impact during a large sequential batch program after
hours.
Consider the I/O load as well and correlate it to the observed latency. For example, latency over 20
ms during a 2-hour window on a volume doing less than small (8KB) block I/Os might not be relevant.
The following guidelines apply to I/O operations with an average size of 16KB or less:
Less than 20 msIn general, average latencies of less than 20 ms are acceptable.
20 ms to 50 msSustained average latencies between 20 ms and 50 ms should be monitored
closely. You might want to reduce the workload or add additional storage resources to handle the
load.
51 ms to 80 msSustained average latencies between 51 ms and 80 ms should be monitored
closely. Applications might experience problems and noticeable delays. You might want to reduce
the workload or add additional storage resources to handle the load.
Greater than 80 msA sustained average latency of more than 80 ms indicates a problem,
especially if this value is sustained over time. Most enterprise applications will experience problems
if latencies exceed 100 ms. You should reduce the workload or add additional storage resources
to handle the load.
If the average I/O operation size is greater than 16KB, the previous latency guidelines might not apply.
If latency statistics indicate a performance problem, examine the total IOPS in the pools. The storage
array configuration (disk drives and RAID level) determines the maximum number of random IOPS that
can be sustained. Dell Support or your channel partner can help size storage configurations for
specific workloads.
Also, check the latency on your servers. If the storage does not show a high latency, but the server
does, the source of the problem might be the server or network infrastructure.
Estimated I/O loadEstimated load, relative to the theoretical maximum capability of the group, pool,
or member. The estimated I/O load is based on latencies, IOPS, hardware, and the RAID configuration.
The load value is an estimate; use it only as a general indicator. The I/O load can be:
NOTE: The network is rarely a bottleneck in a SAN. Usually, network bandwidth is underutilized,
especially with random I/O workloads.
Network rateIn general, the network rate should be 100% to 200% of the I/O (iSCSI) traffic. A
network rate that is significantly more than 200% might indicate a problem.
Queue depthSAN HQ displays the queue depth (average number of outstanding I/O operations at
the start of each incoming I/O operation) for each disk drive (raw I/O), volumes (iSCSI traffic only),
groups, and pools. A queue depth of zero indicates no I/O activity. High or sustained queue depths
might indicate that the group is under a high load.
NOTE: A group must be running PS Series firmware version 4.2 or later to display iSCSI queue
depth for a volume.
Experimental analysis dataIn the Experimental Analysis windows, SAN HQ provides information
about how the group is performing, relative to a typical workload of small, random I/O operations.
This information is only an estimate and must not be used as the sole measure of group performance.
See Experimental Analysis.
Experimental Analysis
The Experimental Analysis window provides information about PS Series group performance, relative to a
specific workload. This information is only an estimate. However, it can help you more fully understand
group performance and also help you plan for storage expansion.
NOTE: Because the estimated information in the Experimental Analysis window is based on a
workload that might not resemble the actual group workload, the data should not be used as the
sole measure of group performance. The Experimental Analysis window also provides runtime
group performance data, so you can compare the estimates to actual data. Always consider latency
when examining estimated performance data.
The Experimental Analysis for Group window appears, as shown in Figure 58. Experimental Analysis Data
for a Group.
Current hardware configuration (including RAID level, controller type, and disk type)
Current distribution of reads and writes (that is, the percentage of IOPS that are reads and the
percentage of IOPS that are writes)
SAN HQ then calculates the performance estimates, based on the previous data and a workload with the
following IOPS characteristics:
Small (8KB)
Random
Estimated IOPS WorkloadFor the selected time range, the graph shows the percentage of how
much work (IOPS) the group is performing, based on the estimated maximum number of IOPS the
group can perform (estimated maximum IOPS) and the actual number of IOPS performed by the
group.The table to the left of the graph shows how much work (IOPS) the group is performing,
averaged over the time range. For example, if the estimate is 50%, the group is performing half the
work SAN HQ estimates that the group can perform. This estimate is based on a workload consisting
of small, random IOPS and the group hardware configuration and read/write distribution. The
Estimated IOPS Workload percentage is never more than 100%, even if the group is performing at
more than 100% of the estimated maximum number of IOPS.
Estimated Maximum IOPSFor the selected time range, the graph shows the estimated maximum
number of IOPS the group can perform, based on a workload consisting of small, random IOPS and
the group hardware configuration and read/write distribution. To compare the estimated data with
runtime data, the graph also shows the actual number of IOPS (reads and writes) performed by the
group. Because the estimated maximum IOPS data is based on the group hardware configuration and
read/write distribution, the estimated data usually tracks the actual number of I/O operations in the
group.
Every 7 days, the SAN HQ Server automatically updates the SAN local repository, including newly
supported disks and hardware, used for Estimated IOPS calculations. This local repository is also used
to verify member configuration and disk drive support, and can be updated manually.
Occasionally, the SAN HQ Client might show an estimated IOPS value of zero and display a
notification stating, SAN HQ Error: Unknown configuration is detected. Estimated max
IOPS cannot be calculated. This notification typically indicates that the SAN HQ version you are
running proceeds a newer drive type in your PS Series array. Dell recommends that you upgrade to
the latest release of SAN HQ (see Upgrade SAN Headquarters). Verify that the SAN HQ Server has
access to the Internet for upgrading the SAN HQ software or for updating the local repository.
Occasionally, SNMP timeouts during RAID verification might cause critical information to be
temporarily unavailable. In this case, Estimated IOPS cannot be calculated.
SAN HQ calculates the estimated maximum group IOPS when no disk drives in the group have failed
(orange line in the graph) and also when at least one RAID set is in a degraded state (brown line in the
graph). This information is useful for understanding the performance impact of a disk failure. The
degraded estimate is based on a drive failure in a RAID set that would result in the greatest
performance impact.
The table to the left of the graph shows the estimated maximum number of IOPS (under non-failure
and degraded RAID set conditions), averaged over the selected time range. The degraded estimate
does not include the performance impact that might occur during RAID reconstruction (for example,
when the array is reconstructing data from parity on a spare drive).
Estimated IOPS Workload Percentage is below 50% and no performance issues. If the runtime group
data does not indicate a performance problem (that is, latencies are low, applications complete on
time, and user response time is adequate), then you can assume that the group can handle an
increase in workload without a performance degradation.
Estimated IOPS Workload Percentage is more than 80% and no performance issues.
If the runtime group data does not indicate a performance problem (that is, latencies are low,
applications complete on time, and user response time is adequate), and your workload consists of
You reached the limit of the groupYou should immediately consider decreasing the load on the
group or adding additional hardware or arrays.
One or more members have degraded RAID setsReplace failed drives as soon as possible.
Network problems existCorrect the network problems immediately.
A server has reached its maximum capabilitiesConsider increasing the I/O capabilities of the
server (for example, install additional network interfaces and configure multipathing).
Member hardware problems existReplace any failed hardware and ensure that you configure all
the network interfaces on all group members.
As these examples show, estimated data must be used with runtime group data to obtain an accurate and
comprehensive understanding of group performance. For runtime group data examples, see
Experimental Analysis Data.
The examples in the following sections might help you better understand the relationship between
different types of performance data.
Table 39. Performance Data for Adequately Performing Group with Excess Capability
Data Description
Latency Good (less than 15 ms)
IOPS Good (performing approximately 50% of the estimated maximum
IOPS)
I/O Size Typical (approximately 50KB for reads and 6KB for writes)
I/O Rate Low (approximately 5 to 15 MB/sec)
Overall Assessment Adequate
Expansion Capability Might be able to increase workload by 25%
Queue Depth Typical range
Example 1 shows a group that is performing well and is within its capabilities. The latencies are all below
20 ms, which is good. The reported number of IOPS is about half the maximum IOPS (small, random) that
SAN HQ estimates the group can easily perform. The I/O sizes are typical, so you do not have to consider
any special circumstances (such as very large I/O sizes). The I/O rate (throughput) is low. The workload
appears to be reasonably static (and thus predictable).
Data Description
Latency Good (less than 4 ms)
IOPS Very low (approximately 15% of estimated maximum IOPS)
I/O Size Typical (approximately 40KB for reads and 8KB for writes)
I/O Rate Very low (average of 2.5MB/sec)
Example 2 shows a group that is mainly idle. The very low latency and low IOPS values indicate that this
group can handle a larger workload. However, because the current group workload is so low, it is difficult
to determine how large a workload increase the group can handle. Increase the workload gradually and
evaluate the group performance after each increase.
Table 41. Performance Data for a Group Possibly Near Full Capability
Data Description
Latency Good (less than 20 ms)
IOPS High (at least 100% of the estimated maximum IOPS)
I/O Size Typical (approximately 40KB for reads and 32KB for writes)
I/O Rate Low (average 30 to 40 MB/sec)
Overall Assessment Adequate
Expansion Capability Increase workload gradually and with caution
Queue Depth Moderate
Example 3 shows some contradictory information. The latencies are low (less than 20 ms). However, the
IOPS are at or above the maximum estimated IOPS (small, random) for the group. Because the I/O size is
small (less than 64KB), the workload might be sequential instead of random. Alternatively, the group
might be benefiting from a high level of control module cache hits.
Figure 65. I/O Window for a Busy Group Probably Near Full Capability
Table 42. Performance Data for a Busy Group Probably Near Full Capability
Data Description
Latency Cautionary (above 20 ms, but less than 50 ms)
IOPS High (approximately 100% of the estimated maximum IOPS)
I/O Size Smaller than typical (approximately 12KB for reads and 10KB for writes)
I/O Rate Low (less than 10MB/sec)
Network Load Low (less than 2%)
Overall Assessment Busy
Expansion Capability Increasing workload will probably result in performance degradation
Example 4 shows that a group can have a high I/O load and a low network load at the same time. The
network is rarely a bottleneck in a SAN. Usually, network bandwidth is underutilized, especially with
random I/O workloads.
This group has a heavy load, consisting of highly random, small reads and writes. Yet, only a fraction of
the network is being utilized. While this example is extreme, the concept is true for most groups.
Figure 68. I/O Window for a Group With High Latencies Probably Near Full Capability
Table 43. Performance Data for a Group With High Latencies Probably Near Full Capability
Data Description
Latency High (20 to 60 ms, sustained)
IOPS Very high (above 100% of the estimated maximum IOPS)
I/O Size Typical (approximately 62KB for reads and 18KB for writes)
I/O Rate Low (average 45MB/sec)
Overall Assessment Very busy
Expansion Capability Increasing workload will probably result in performance degradation
Queue Depth Moderate to high
Example 5 shows a busy group with no excess capacity for expansion. The high latencies, sustained over
an 8-hour period, indicate that group performance is troublesome. While brief peaks of high latency are
acceptable, high sustained latencies generally have a negative impact on application performance.
In addition, the number of IOPS in this example is two to three times the estimated maximum IOPS. This
increase is probably because the workload is sequential instead of random. Alternatively, the group might
be benefiting from a high level of control module cache hits.
Figure 70. I/O Window for Group With Many Small Writes But Some Large Reads
Data Description
Read/Write Distribution Approximately 32% reads and 68% writes
Read I/O Size Average 70KB
Write I/O Size Average 3KB
Read IOPS Average 170
Write IOPS Average 350
Read I/O Rate Average 12MB/sec
Write I/O Rate Average 1MB/sec
Queue Depth Moderate
Example 6 shows how the read/write distribution is important for understanding performance statistics.
The read/write distribution is the percentage of read IOPS and the percentage of write IOPS, based on the
overall number of IOPS.
The data in Example 6 indicates a workload that consists mainly of write operations (68% writes).
However, these writes are small (average of 3KB). Because the read operations are large (average of
70KB), compared to the writes, most of the I/O throughput is read data (average read I/O rate of 12MB/
sec).
Use network components that are recommended for a Dell EqualLogic iSCSI SAN.
Use switches and interswitch links that have sufficient bandwidth for the iSCSI I/O. Contact Dell
Support for details about correct sizing and configuration of iSCSI SAN switch infrastructure. Pay
careful attention to sizing interswitch links.
Make sure all member network interface connections are Gigabit Ethernet and make sure driver
settings are correct on the servers. SAN HQ generates an alert if it detects connections that are less
than 1GB.
Follow the network requirements and recommendations in the Dell EqualLogic PS Series Installation
and Setup Guide . In general:
If possible, do not use Spanning Tree Protocol (STP) functionality on switch ports that connect end
nodes.
Enable flow control on switches and NICs.
Disable unicast storm control on switches.
Enable jumbo frames.
Have sufficient server memory. Adding memory can decrease the read I/O load on the storage.
Use the latest drivers for your operating system.
Add high-performance network interfaces or host bus adapters to servers.
Consider using multipathing (MPIO), which provides a high-performance and highly available path
between a server and storage. Use the Dell EqualLogic Host Integration Tools/Microsoft for a
multipathing solution.
Make sure the server Ethernet ports, PCI bus, and CPU are rated for the workload.
If the server is a cluster node, isolate cluster network traffic from iSCSI storage traffic. Check for other
network traffic interference.
Configure the application properly so that I/O is performed efficiently.
Make sure all the network interfaces on the members are configured, functioning, and accessible. Use
the Group Manager CLI ping command to check accessibility. If you need more network bandwidth
for iSCSI I/O, consider not configuring a dedicated management network.
If pool capacity is low, make one or more of the following changes:
NOTE: The Read I/O Rate and the Write I/O Rate for members are retrieved from the MIB table. For
group and pool levels, these values are computed by adding these values from all members (for
pools) and by adding these values from all pools (for groups).
Case 1: A user wants to identify the difference in usage metrics for workloads on SSD and HDD disks at
the member level.
1. Select a group from the Servers and Groups tree, then Hardware / Firmware, and then Disks.
2. Select All Disks I/O Rate from the combo box.
3. In the menu bar, click on Members and select a specific member to monitor or all members in the
group.
The line chart containing SSD Reads, SSD Writes, HDD Reads and HDD Writes series displays at the
member level.
Case 2: A user wants to identify the difference in usage metrics for workloads on SSD and HDD disks at
the pool level.
1. Select a group from the Servers and Groups tree, then Hardware / Firmware, and then Disks.
2. Select All Disks I/O Rate from the combo box.
3. In the menu bar, click on Pools and select either All pools or the default pool to set the scope of the
data.
The line chart containing SSD Reads, SSD Writes, HDD Reads and HDD Writes series displays at the
member level.
1. Select a group from the Servers and Groups tree, then Hardware / Firmware, and then Disks.
2. Select All Disks I/O Rate from the combo box.
3. In the menu bar, click on Group.
The line chart containing SSD Reads, SSD Writes, HDD Reads and HDD Writes series displays at the
group level.
Case 4: A user wants to know which type of disks (HDD or SSD) are being used.
1. Select a group from the Servers and Groups tree, then Hardware / Firmware, and then Disks.
2. In the All Disks Summary window, refer to the Disk Type column. This information is available
regardless of the group, pool, or member selection.
Case 5: A user wants to know the percentage of lifetime remaining on the SSDs.
1. Select a group from the Servers and Groups tree, then Hardware / Firmware, and then Disks.
2. In the All Disks Summary window, refer to the Lifetime Remaining column. This information is
available regardless of the group, pool, or member selection. The Lifetime Remaining data is not
applicable to HDDs.
NOTE: Disks are considered to be of type SSD if they are SAS_SED, SAS_SSD, or SATA_SSD.
Otherwise they are considered HDD. These values are directly retrieved from the MIB table.
The Lifetime Remaining values are retrieved directly from the MIB table and are the same as the
manufacturing value.
You can preserve data at the current time or use a command line to perform the task.
Reports are generated as a background process, so you can continue to use SAN HQ.
You can also create reports from archived group data. Open the archive and follow the instructions for
creating a report.
SAN HQ GUIUsing the GUI is the easiest method of creating a report. You can display the report,
save the report to a file, or email the report. See Using the GUI to Create a Report.
Command lineUsing a command to create a report enables you to schedule the creation of reports
and regularly capture group data. See Use a Command to Create a Report.
Reports minimally require approximately 5MB per group per report, depending on the selected time
range.
Report Types
You can create the following report types for the selected groups:
1. From the Reports menu, either select the type of report you want to create or select Report Wizard,
which allows you to select the type of report.
2. A series of dialog boxes appears, prompting you for the information described in Information
Required for a Report. Enter the necessary information and click Next to continue.
3. In the final Report Wizard dialog box (Title and Generation Settings), select the Generate Report Now
option. SAN HQ generates the report in a background process.
The completed report (usually consisting of multiple pages) appears in the Report Viewer. Use the toolbar
at the top of the Report Viewer window to:
1. From the Reports menu, either select the type of report you want to create or select Report Wizard,
which allows you to select the report type.
2. A series of dialog boxes appears, prompting you for the information described in Information
Required for a Report. See Use the GUI to Create a Report for a description of the dialog boxes. Be
sure to select the Generate Command Line option in the Report Wizard Title and Generation
Settings dialog box. Optionally, you can automatically email the report after it is generated. In the
Report Wizard E-Mail Settings dialog box, select Automatically e-mail report as an attachment
after generation and enter the information in the fields. If you configured email notification for alerts,
the fields already contain data. You can edit the fields as needed.
3. In the Report Wizard Report Generation Ready dialog box, click Generate Command Line. The
Report Generator XML Viewer window appears (Figure 71. Report Generator XML Viewer), displaying
the contents of the XML file. You can edit this file, if necessary, as described in Modify an XML File for
Creating Reports.
4. When you are satisfied with the contents of the XML file, click Save in the Report Generator XML
Viewer dialog box and specify a file name for the XML file.
5. To generate the report, navigate to the SAN HQ installation directory and use the following
command format.
Be sure to specify the full path for the XML file. SANHQClient.exe -
ReportSettingsFile="xml_file_name"
For example: > SANHQClient.exe ReportSettingsFile="C:\SANReport_XML.xml"
The XML_file_name variable specifies the name of the XML file you created in Use a Command to Create
a Report.
<SanHQReport>
<ReportFile>
C:\Documents and Settings\sample_user\My Documents\SanHQ_Report.pdf
</ReportFile>
<Title>
SAN Headquarters Configuration Report
</Title>
<Type>
Configuration Report
</Type>
<Style>
Default
</Style>
<DataType>
Point
</DataType>
<GrabNearestPointWithin>
36000000000
</GrabNearestPointWithin>
<Time
span = "24"
/>
<Group>
magic4-grp
</Group>
</SanHQReport>
The following sample XML file creates a Capacity Report (in HTML format) from data averaged over a
specific 2-week time period for two groups:
<SanHQReport>
<ReportFile>
C:\sample_user\My Documents\Capacity_Report.html
</ReportFile>
<Title>
SAN Headquarters Capacity Report
</Title>
<Type>
Capacity Report
</Type>
<Style>
Default
</Style>
<DataType>
Point
</DataType>
<GrabNearestPointWithin>
36000000000
</GrabNearestPointWithin>
<Time
begin = "7/15/2013 2:35:31 PM"
The following sample XML file creates a Top 10 Report (in PDF format) from the latest data for a group
and then emails the report:
<SanHQReport>
<ReportFile>
C:\sample_user\My Documents\TOP_10_Report.pdf
</ReportFile>
<Title>
SAN Headquarters TOP 10 Report
</Title>
<Type>
Top 10 Report
</Type>
<Style>
Default
</Style>
<DataType>
Point
</DataType>
<GrabNearestPointWithin>
36000000000
</GrabNearestPointWithin>
<Time
span = "0"
/>
<Group>
MandaGrp
</Group>
<AutoEmailReport>
<SmtpServer>
10.20.30.40
</SmtpServer>
<Sender>
grouo@company.com
</Sender>
<SmtpPort>
25
</SmtpPort>
<SubjectLine>
SAN Headquarters Top 10 Report
</SubjectLine>
<SendToAddress>
me@company.com
</SendToAddress>
<SendToAddress>
you@group.com
</SendToAddress>
</AutoEmailReport>
</SanHQReport>
Table 45. Group Diagnostics Report Data lists and describes the data analyzed by the Group Diagnostics
Report.
Pool analysis Pools with high I/O loadInforms you to distribute the load to other pools
Pools with number of connections approaching 90% maximum (firmware
version 4.2 and later)
Pools with members offlineDisplays the number of members offline
Pools Total, Free, and In-Use capacity within 90% threshold of the collection
Pool delegated space at 80% threshold
Port analysis 10Gb Ethernet port operating below maximum link speed of 10Gb
1Gb Ethernet port operating below maximum link speed of 1Gb
Port with an admin status of not up (that is, a user-disabled port or a never-
enabled port)
Port with an admin status of up (that is, a user-enabled port) but operationally
down (that is, a disconnected port or unresponsive port)
It can be beneficial to regularly archive group data. SAN HQ maintains group performance data in log files
for up to 1 year. As data ages, the SAN HQ Server compresses the data in the log files, which can make
some older data less precise than newer data. You can periodically archive data to retain more-precise
data. See How Data is Compressed in Log Files.
You also might need to archive data if requested by your PS Series support provider.
Group archive data can also be included in a SupportAssist on-demand data collection. See On-Demand
Data Collection.
SAN HQ GUIUsing the GUI is the easiest method of archiving data. See Using the GUI to Create an
Archive File.
Command lineUsing a command to archive data enables you to schedule archive operations and
regularly capture group data. See Using a Command to Create an Archive File.
A group data archive minimally requires approximately 1.5 to 2 times the log size of the amount of data
selected.
1. From the SAN HQ menu, select Create Archive to initiate the Archive Groups wizard.
2. In the first dialog box of the wizard, click Next. The Group Selection dialog box appears.
Select the service (if multiple servers are configured) and any groups whose data you want to save.
You can archive groups from only one service at a time. Click Select all groups to archive data for all
the monitored groups.
3. Click Next. Select the time range for the data to include in the archive. By default, Time Range is
selected, allowing you to indicate the latest hour, latest 8 hours, latest 1 day (default), latest 7 days, or
latest 30 days for the time range, or enter a specific range of dates. Alternatively, select All Data to
include all data collected to date in the archive.
4. Click Next. The Archive Generation Type dialog box appears.
5. Select the Generate Now option.
6. Enter a file name for the .grpx archive file. The default name is SANHQ_Archive.grpx.
7. Optionally, check the Trace Files Only checkbox. By selecting this option, the resulting archive is a
compressed archive containing only debug trace files without any group log data. These compressed
debug trace files are designed for easy transmission of important diagnostic information.
8. Click Next. The Summary of Archive dialog box appears.
9. Click Archive now to create the archive.
1. From the SAN HQ menu, select Create Archive to initiate the Archive Groups wizard.
2. In the first dialog box of the wizard, click Next. The Group Selection dialog box appears.
Select the service (if multiple servers are configured) and any groups whose data you want to save.
You can archive groups from only one service at a time. Click Select all groups to archive data for all
the monitored groups.
3. Click Next. The Archive Generation Type dialog box appears.
4. Select the Generate Command line Archive Settings file option.
5. Enter a file name for the .grpx archive file. The default name is SANHQ_Archive.grpx.
6. Optionally, check the Trace Files Only checkbox. By selecting this option, the resulting archive is a
compressed archive containing only debug trace files without any group log data. These compressed
debug trace files are designed for easy transmission of important diagnostic information.
7. Click Next. The Summary of Archive dialog box appears.
8. Click Generate Command Line. The Create Archive XML Viewer dialog box appears (Figure 72.
Create Archive XML Viewer), displaying the XML file.
9. Examine the XML file. If necessary, you can edit the file as described in Modify an XML File for
Archiving Data.
10. Click Save if the file is acceptable and then enter the path and file name for the XML file. Click Cancel
to cancel the operation.
11. Navigate to the SAN HQ installation directory and use the following command format. Be sure to
specify the full path for the XML file.SANHQClient.exeArchiveSettingsFile="xml_file_name "
For example: > SANHQClient.exe ArchiveSettingsFile="C:\SAN_Archive_XML.xml"
<SanHQArchive>
<Path>
C:\Documents and Settings\sample_user\SANHQ_Archive.grpx
</Path>
<Group>
10.124.9.144
</Group>
<Group>
10.127.14.200
</Group>
</SanHQArchive>
If you want, you can create reports from archived data. You can also start the Group Manager GUI from
an archive.
You cannot export group data, create an archive, add a group, or change the SAN HQ settings for a group
though an archive.
1. Make the archive file available to the computer running the SAN HQ Client.
2. Start the SAN HQ GUI.
3. In the All Groups window (Starting the SAN Headquarters GUI), from the Monitor menu, select Open
archive.
4. Enter the archive file name and location and click Open.
A new SAN HQ GUI window appears, showing the archived group data. The window title bar shows
Archive, and the group Monitoring Status is disconnected.
It can be beneficial to regularly export group data. SAN HQ maintains group data in log files for up to 1
year. As the data ages, the SAN HQ Server compresses the data in the log files, which can make some
older data less precise than newer data. You might want to periodically export data to retain the precise
information.
SAN HQ enables you to select the date and time range for the exported data. You can choose to include
information about the entire group or more-detailed information about pools, members, volumes,
network ports, replicas, and disks.
SAN HQ GUI Using the GUI is the easiest method of exporting data. See Using the GUI to Export
Group Data.
Command line Using a command to export data enables you to schedule export operations and
regularly capture group data. See Using a Command to Export Group Data.
Exports minimally require approximately 50 MB per group, depending on the selected time range, data
object, and size of the groups log files.
1. In the Servers and Groups tree, right-click either All Groups Summary or a specific group and select
Export Group Data to initiate the Export Group Data wizard.
2. In the first dialog box of the wizard, click Next. The Group Selection dialog box appears.
3. In the Export Group Data dialog box, either:
Select the groups for which you want to export data.
Click Select all groups to export data from all the groups shown.
4. Click Next. The Time Range Selection dialog box appears.
5. Select the time range for the data in the log files that you want to export. Alternatively, select Specific
range to enter a range of dates. By default, Time range and latest 1 day are selected. The time range
ends with the most recent time. For example, if you change the default to Latest 7 days, the exported
data is from the most recent 7-day time period.
6. Click Next. The Data Selection dialog box appears.
7. Select the type of data to include. You can include summaries of group, pool, and member data or
information about volumes, volume collections, network interfaces, disks, inbound or outbound
replication, and replication partners.
8. Click Next. The Export Generation Type dialog box appears.
9. Make sure that the Generate Now option is selected.
10. Specify the path and identifier for the .csv file. The actual file name is generated automatically, based
on the file name you enter, the selected group names, and the time range.
11. Click Next. The Summary of Export dialog box appears.
12. Click Export Now to export the data. Click Cancel to cancel the operation.
The procedure for exporting group data by using a command requires using the SAN HQ GUI to create
an XML file that contains the groups and the type of data that you want to export. You then run the SAN
HQ executable, specifying the XML file as a parameter.
1. In the Servers and Groups tree, right-click either All Groups Summary or a specific group and select
Export Group Data to initiate the Export Group Data wizard.
2. In the first dialog box of the wizard, click Next. The Group Selection dialog box appears.
a. Examine the XML file. If necessary, you can edit the file, as described in Use a Command to Export
Group Data.
For example:
> SANHQClient.exe ExportSettingsFile="C:\SAN_Export_XML.xml"
The XML_file_name variable specifies the name of the XML file you created in Using a Command to
Export Group Data.
The following sample XML file exports all types of data gathered over the most recent 24-hour time
period for two groups:
<SanHQExport>
<Settings
Path="C:\SANHQ_Export.csv">
<Show
Group="True"
Pools="True"
Members="True"
Volumes="True"
HostedReplicas="True"
ReplicaSites="True"
Disks="True"
Ports="True"
Volume Collections="True"
Outbound Replicas="True"
/>
<Time
Span="24"
/>
</Settings>
<Group>
10.127.137.110
</Group>
<Group>
The following sample XML file exports all types of data, except for replication, gathered over a specific
date range for two groups:
<SanHQExport>
<Settings
Path="C:\SANHQ_Export_Range.csv">
<Show
Group="True"
Pools="True"
Members="True"
Volumes="True"
HostedReplicas="False"
ReplicaSites="False"
Disks="True"
Ports="True"
Volume Collections="True"
Outbound Replicas="False"
/>
<Time
Begin="1/01/2013 1:00:00PM"
End="1/12/2013 4:08:13PM"
/>
</Settings>
<Group>
10.117.127.120
</Group>
<Group>
10.117.141.140
</Group>
</SanHQExport>
Diagnostic data uploaded by the SAN HQ Server is organized such that support personnel can easily view
the data when working on a support case. Dell Support contacts you only regarding issues discovered
when analyzing the data collected by SupportAssist or when you specifically request support.
SupportAssist does not fully replace the E-Mail Home feature or the pssdiag script for diagnostic
reporting to Dell Support. Instead, it reduces the need to use these reporting methods on a frequent
basis. SupportAssist filters health condition event messages from the PS Series group and collects a
subset of critical events to upload to Dell Support. In contrast, various health condition events trigger the
E-Mail Home messages, which can be informational, errors, or warnings.
NOTE: If pssdiag is running, scheduled and event-driven data collections by SupportAssist cannot
override pssdiag data collections. For on-demand data collections, SupportAssist first informs you if
pssdiag is running and then lets you either cancel the SupportAssist data collection, or ignore
pssdiag and continue with the on-demand data collection.
The accumulated data collected by SupportAssist from many PS Series groups lets Dell Support
continually update its diagnostic analysis. illustrates this process.
Callout Description
1 PS Series group at Site A (blue) and Site B (green)
2 PS Series SAN arrays
3 SAN networks (orange)
4 LAN networks (purple)
5 SAN HQ servers
Fast and automatic data collectionDepending on the size of your group, a scheduled data collection
might take as little as 5 minutes for a typical two-to-four member configuration.
Parallel data collectionWhen you configure multiple groups to use SupportAssist, data collection
occurs in parallel, up to three groups at a time, across all groups.
Preventive diagnosticsBy gathering commonly used diagnostics and configuration information, then
making this information available to Dell Support, SupportAssist significantly reduces the time it takes
for the Dell Support organization to help customers who might be experiencing similar difficulties.
Dell Support can use historical data to determine if customers are at risk, without requiring customer
intervention.
SecuritySupportAssist uses a special read-only group account (by default, autopilot-ro, which you
can change) to access and gather only diagnostic data from the group. It cannot access user data
stored on the group. Diagnostic data is uploaded on an SSL-encrypted secure channel.
Consistency and accuracyBecause most data analysis occurs on the server-side software,
customers who enable SupportAssist functionality will benefit from consistent automated analysis,
free of human error.
ResiliencyIf SupportAssist cannot complete an upload due to problems with Internet connectivity or
the server being down, it defers the upload and attempts it at a later time.
Minimal performance impactFor the majority of groups, under normal conditions, using
SupportAssist adds only a 5% to 10% performance impact during the relatively short time when
SupportAssist collects data.
Password scrubbingAll user passwords are scrubbed before data is uploaded.
Convenient configuration managementFrom the SupportAssist Configuration Settings page, all
groups configured for SAN HQ monitoring are displayed in detailed dialog boxes from which you can
enable SupportAssist and set or modify configuration settings.
Communication with Dell SupportSupportAssist receives information about all support cases
created for your groups and conveniently displays the case status and progress on resolution, as well
as other critical customer support case information.
User-controllable features:
SupportAssist collects diagnostics according to a weekly schedule that you set up. You can also
perform an on-demand data collection at any time.
SupportAssist gathers diagnostic and configuration data for your own use and inspection, without
transmitting it to Dell (see On-Demand Data Collection).
SupportAssist monitors current and past activity via the SupportAssist activity log.
You can enable SupportAssist for any number of groups, up to the supported maximum of 50.
You can create a custom user name for the SupportAssist read-only account that accesses and
collects diagnostic data from the group.
SupportAssist collects diagnostics and configuration data from PS Series groups, including but not limited
to the following categories:
SupportAssist Requirements
This section describes the requirements for using SupportAssist:
Firmware PS Series array firmware version 7.0 or later release. If you are using a firmware version
that is more recent than the SAN HQ release, you might need to upgrade SAN HQ.
SAN HQ version 3.0 or later release.
Read-only account credentials You must supply grpadmin-level credentials for each monitored
group to SAN HQ during the initial setup.
You can rename this account when you configure a group for SupportAssist data collection, either
when adding a group using the Add Group wizard or, when configuring SupportAssist, using the
SupportAssist Configuration wizard. SupportAssist uses this read-only account to connect to the
sanmond daemon that gathers diagnostic data from the PS Series group.
The password must contain a minimum of 3 alphanumeric characters, which are case-sensitive. SAN
HQ uses these group access credentials one time for SupportAssist configuration; it does not store
You can manage the autopilot-ro account using the PS Series Group Manager GUI (or CLI), allowing
you to disable SupportAssist from Group Manager by changing the SupportAssist account password.
syslog notification One free IP address slot to configure syslog notification to the SAN HQ Server. A
group can support up to three IP addresses for syslog notification. If all addresses are in use, you must
free up an address and specify the SAN HQ Server instead. This address is used to monitor critical
events on the PS Series arrays.
SAN HQ automatically enables syslog event logging for fatal and error events under the Group
Manager GUIs Notifications tab.
You must manually enable these group management features:
SNMP Configure a read-only SNMP community name under the Group Manager GUIs SNMP
tab. If you are already monitoring your group with SAN HQ, you do not need to configure a name.
SSH Enable SSH under the Group Manager GUIs Administration tab. SSH is enabled by default.
If all the groups share a physical location, use a dedicated SAN HQ Server to monitor all groups,
installing multiple SAN HQ clients for the same server.
If groups are spread across different physical locations, use a dedicated SAN HQ Server for each
physical location. Install one or more SAN HQ clients, each connecting to multiple SAN HQ
servers, to see all the data from a single client.
You can have a single SAN HQ Server connected to multiple groups, running a single instance of
SupportAssist for each group.
Use a static IP address (see Network and Bandwidth Requirements).
The SAN HQ Servers IP address is added to the list of syslog servers for all SupportAssist-configured
groups. Use a static IP address for syslog notification, which prevents the need to update the list of
syslog servers for each group if the address changes dynamically.
The SAN HQ Server must be able to connect to the Dell Support data center via the Internet.
Table 47. Network and Bandwidth Considerations When Using SupportAssist presents approximate
data transfer rates for the SupportAssist infrastructure.
Configure SupportAssist
You can configure SupportAssist at various stages when using SAN HQ:
When adding a new group (see Configuring SupportAssist When Adding a New Group).
By using the SupportAssist Configuration wizard for an existing group (see Configuring SupportAssist
for an Existing Group).
From the SupportAssist Settings page, when enabling a group for SupportAssist or modifying an
existing configuration (see Modify the SupportAssist Configuration).
Use the configuration worksheet to record information needed to set up SupportAssist (see the
Configuration Worksheet).
Obtain the following data for every PS Series group that you intend to configure:
Configuration Worksheet
Use the worksheet shown in Table 48. Group and SupportAssist Configuration Worksheet to gather data
and make decisions about configuring SupportAssist.
Verify the configuration View the configuration settings Modify the SupportAssist Configuration
and configure SupportAssist
from the SupportAssist
Configuration Settings page.
Test the configuration Run an on-demand data On-Demand Data Collection
collection.
To prepare for adding a new group, gather the data as described in Configuration Worksheet and make
sure you configure the group as described in SupportAssist Requirements.
Make sure that your SAN HQ Server is fully functional and not in a degraded state.
1. Click SANHQ and then Add New Group to open the Add Group wizard.
2. Click Next to open the PS Series Group Information dialog box (Figure 75. Add Group Wizard
Group Information Dialog Box).
3. In the dialog box:
a. Type the Group IP address or DNS name.
b. Type the Group Community SNMP name.
c. Check the Enable SupportAssist functionality box. (You must accept the license in step 4.
Depending on whether the TCP/IP initialization is complete, the box might already be checked.)
4. Click Next to open the SupportAssist License Agreement dialog box and accept the license terms. If
you decline the license, you can continue to add the group for SAN HQ monitoring, but you cannot
enable SupportAssist.
5. Click Next. A message dialog box shows that SAN HQ is polling the group and loading group
information. When SAN HQ contacts the group, it opens the PS Series Group Login Credential and
Syslog Selection dialog box (see Figure 76. Result of Adding Group Login Credentials and Syslog
Server Selection). Note: If the groups configuration is invalid, or if any problems occur with the SAN
HQ Server or its network connections to the group, the procedure can fail at this point. If the
procedure fails, correct the problem and try again to add the group.
6. Enter the following information:
a. The user name and password as the group admin credentials.
b. A user name for the SupportAssist read-only account. You can accept the default autopilot-ro
account name or enter your own custom account name. User names follow the same
conventions as the PS Series firmware local account and can contain up to 16 alphanumeric
ASCII characters, including period, hyphen, and underscore. The first character must be a letter or
number and the last character cannot be a period.
c. A password for the SupportAssist read-only account Passwords must contain a minimum of 3 and
a maximum of 16 alphanumeric characters, which are case-sensitive. Dell recommends a
minimum of 6 characters for improved security.
Figure 76. Result of Adding Group Login Credentials and Syslog Server Selection
9. Click Next and type information in the Customer Contact Information dialog box (Figure 77.
SupportAssist Customer Contact Information).
Dell Support contacts you only if, when analyzing aggregated data, it discovers an issue that is
relevant to your PS Series groups and the service tag has ProSupport, or if you request support. To
make sure that Dell Support contacts the appropriate person, you must supply contact information
(for example, for your SAN administrator). For a description of these fields, see Update or Configure
Contact Data.
10. Click Next and enter your preferred data-collection and upload settings (see Figure 78. SupportAssist
Data Collection and Upload Settings Dialog Box):
a. Check the box to enable weekly data collection and set the day, time, retry interval, and the total
number of attempts, which includes the initial attempt.
b. Check the box to collect diagnostics data when critical events are detected in addition to
scheduled activity.
c. Check the box to allow automatic transmission of diagnostic data to Dell Support for review.
NOTE: If you do not check the box to allow automatic data transmissions, you will not get
the benefit of automatically uploading the data to Dell Support for analysis. Dell
recommends keeping this setting selected unless you are in a situation when you are not
always connected to the Internet and you want to store data locally. For more information,
see Use a Different Server to Upload Diagnostic Information. For a description of the fields
in Figure 78. SupportAssist Data Collection and Upload Settings Dialog Box, see Data
Collection and Upload Settings.
11. Click Next and, optionally, enter the settings for email notification:
Do not send email alerts (default)
Send email alerts using group settings
Send email alerts using these settings: SNMP server IP address, SNMP port (optional), subject line,
sent from, sent to Click Test E-mail to verify that the notification settings are valid.
12. Click Next and use the sliders to specify your preferences for the size of log files and of event and
audit log files.
13. Click Next and review your configuration settings in the Completing the Wizard dialog box (Figure
80. Successfully Added Group with SupportAssist Enabled).
14. Click Add Group if the settings are correct, or click Back to make changes.
SAN HQ connects to the group and starts to poll the group for data. Depending on the group size, this
operation can take several minutes. You can continue to add more groups or perform other monitoring
operations.
5. Check the boxes in the Select column to select one or more groups, provided that the SupportAssist
Status column indicates that a group is compatible with SupportAssist. You can also click the Select
all groups button to select or deselect every compatible group.
6. Click Next to open the PS Series Group Login Credentials dialog box (Figure 82. SupportAssist Group
Login Credentials Dialog Box) and enter the following information:
a. The group administration user name and password for the selected groups. To select multiple
groups, use Shift + Select.
b. A user name for the SupportAssist read-only account. You can accept the default autopilot-ro
account name or enter your own custom account name. User names follow the same
conventions as the PS Series firmware local account and can contain alphanumeric ASCII
characters, including period, hyphen, and underscore. The first character must be a letter or
number and the last character cannot be a period.
Optionally, click Test to verify the login credentials. If successful, SAN HQ displays "Credentials
Validation Succeeded." If unsuccessful, SAN HQ displays "One or more groups need validation to
continue."
7. Click Next to open the Syslog Server Configuration dialog box (Figure 83. Syslog Server
Configuration Dialog Box) and select the SAN HQ Servers IP address from the list of available syslog
servers. Dell recommends using a static IP address.
NOTE: PS Series groups support up to three syslog servers. You might need to reconfigure or
remove preexisting syslog servers to enable SupportAssist.
8. Click Next to open the Customer Contact Information dialog box and enter your contact
information.
Dell Support contacts you only if, when analyzing aggregated data, it discovers an issue that is
relevant to your PS Series groups and the service tag has ProSupport, or if you request support. To
make sure that Dell Support contacts the appropriate person, you must supply contact information
(for example, for your SAN administrator). For a description of these fields, see Update or Configure
Contact Data.
9. Click Next to open the Data Collection and Upload Settings dialog box and specify when you would
like data to be gathered and transmitted (or retained locally).
NOTE: If you do not check the box to allow automatic data transmissions, you will not get the
benefit of automatically uploading the data to Dell Support for analysis. Dell recommends
keeping this setting selected unless you are in a situation when you are not always connected
to the Internet and you want to store data locally. For more information, see Use a Different
Server to Upload Diagnostic Information. For a description of the fields in Figure 83. Syslog
Server Configuration Dialog Box, see Data Collection and Upload Settings.
10. Click Next to display the Ready to Configure SupportAssist screen (Figure 84. Ready to Configure
SupportAssist Screen) and verify that the information is correct. Click Back to make corrections.
11. Click Configure to configure the groups with SupportAssist.
A progress bar shows the configuration in process. It can take up to two minutes for each group. If all
groups are successfully configured for SupportAssist, the status indicates the configuration is finished and
SupportAssist is automatically deployed. If the configuration reveals any errors for a certain group, an
error message displays. In this case, rerun the wizard after correcting any problems. You need to rerun
the wizard only for groups that fail, after you correct any problems.
For information on the CLI command parameters, see SupportAssist Command Reference.
NOTE: The information presented on the SupportAssist Configuration Settings page is similar to the
configuration information presented when running the Configuration wizard (see Configuring
SupportAssist for an Existing Group).
To access the SupportAssist Configuration Settings page, either:
Click SupportAssist below the Servers and Groups tree (lower-left panel).
Click SupportAssist in the SAN HQ menu bar and then SupportAssist Settings.
The SupportAssist Configuration Settings window appears. The upper-left panel lets you navigate to the
activity log for recent SupportAssist activity (see SupportAssist Activity Log), to a list of customer support
cases reported back to SupportAssist (Display Support Case Information), or back to the SupportAssist
configuration page.
You can change the following SupportAssist configuration settings for a group:
Enabled Status Whether the group is configured and enabled or configured and disabled
Firmware PS Series firmware revision installed on the group
Group Name Alphabetically by group name, regardless of configured status
Server SAN HQ Server monitoring the group
When viewing the configuration settings for a large number of groups, use the Collapse All button to
condense the display. The name appears vertically in collapsed format and the button changes to Expand
All.
Configured groups are identified as Configured after the group name and display full configuration
information. Unconfigured groups are labeled "[Unconfigured]" and show only the firmware version and
server name.
A configured group (left panel in Figure 85. Groups Available for SupportAssist Panel) shows the
configuration information for the group. describes this information, along with available user actions. You
can enable the unconfigured group (right panel in ) and define the SupportAssist configuration settings.
describes the information and available user actions for an unconfigured group.
Configuration Describes the SupportAssist Click to enter or modify the group access
configuration status and enables credentials and SupportAssist read-only
you to access the credentials account credentials. The user name
defaults to autopilot-ro, which you can
change. User names follow the same
conventions as the PS Series firmware local
account and can contain up to 16
alphanumeric ASCII characters, including
period, hyphen, and underscore. The first
character must be a letter or number and
the last character cannot be a period.
Passwords must contain a minimum of 3
and a maximum of 16 alphanumeric
characters, which are case-sensitive.
Enable Weekly Data Checked when you configure Uncheck this box to temporarily disable
Collection SupportAssist SupportAssist.
Day of Week Lists the days of the week, Select a day for data collection (and
Monday to Sunday optional transmission) by SupportAssist.
Start Time Specifies the time in the format: Use the arrows (or type a value) for each
HH:MM AM/PM field to modify the time of day for data
collection (and optional transmission) by
SupportAssist.
Retry After Integer value in the range 10120 Specify the number of minutes to wait
before SupportAssist retries a failed data
collection.
Retry Count Integer value in the range 110 Select the maximum number of data
collection retries (including the initial
attempt) that SupportAssist makes before
timing out.
Collect Critical Data Indicates whether data collection Check this box to collect diagnostic data
should occur after a critical syslog after a critical syslog event, in addition to
event the scheduled weekly collection.
Person Icon GRAPHIC Current contact information View and change information for the
current contact. See Update or Configure
Contact Data.
Person Plus Icon Adds contact information Add new contact information. See Update
GRAPHIC or Configure Contact Data.
Receive SupportAssist Email authorization Authorize Dell Support to send you emails
Emails from Dell in the event that they need to contact you.
Save Saves modifications Click to save any changes that you made.
Discard Changes Removes modifications Click to remove any changes that you
made.
An unconfigured group shows only the group information described in Table 51. Group Not Configured
for SupportAssist.
Modify SupportAssist data collection settings for a group. See Modify SupportAssist Data Collection
Settings for a Group.
Click the SupportAssist link below the Servers and Groups tree and then click Activity Log in the
upper-left panel.
From the SAN HQ menu bar, select SupportAssist and then click Activity Log.
You can sort and filter the data using the sort keys for each column. Click any column header to display
the sort filter key icon. Click Custom to display the Custom Auto Filter dialog box. Right-click any column
header to display the Sort, Fit, and Filter options.
When a group's activity count exceeds the configured maximum, the group's activity log overwrites the
oldest entries in the log.
Table 53. SupportAssist Activity Typeslists potential activity types that you might encounter.
1. Navigate to the SupportAssist Configuration Settings page (see Modify the SupportAssist
Configuration).
2. Select the group.
3. Modify one or more of the following settings:
Day of week
Start time
Retry After (Minutes)
Retry Count
Collect critical data
4. Click Save and then click OK in the save confirmation dialog box.
NOTE: A maximum of three syslog servers are permitted on a group, one of which is the SAN HQ
Server. If you save a configuration and the maximum number of syslog servers already exists, the
configuration fails. To delete or modify the group syslog servers, launch Group Manager.
Table 54. Contact Information for SupportAssist describes the contact information fields and data.
Table 54. Contact Information for SupportAssist
Work Phone User-supplied Work (land line) telephone number for the contact person
Cell Phone User-supplied Cell telephone or alternate number for the contact person
Country Select from menu Location where the contact person resides
Timezone Select from menu Time zone where the contact person resides
Permanently delete the SupportAssist configuration for a group. See Delete the SupportAssist
Configuration for a Group.
Enable or disable automatic uploading of diagnostic data packages to Dell. See Enable and Disable
Automatic Upload for a Group.
1. Navigate to the group SupportAssist settings, as described in Modify the SupportAssist Configuration.
2. Select the group.
3. Click Delete to open the confirmation dialog box.
4. (Optional) Check Remove pending uploads to cancel any uploads that are in progress.
5. Click Delete to confirm or click Cancel to abort the deletion.
The group reverts to its unconfigured state.
Temporarily disable and then reenable the SupportAssist configuration for a group. See Enable and
Disable Weekly Data Collection.
Temporarily disable and reenable automatic uploading of diagnostic data packages to Dell. See Enable
and Disable Automatic Upload for a Group.
Use SupportAssist
This section explains how to use the SupportAssist on-demand data collection and data-decryption
options.
On-demand data collections are useful when you want to manually inspect the diagnostic data collection
before uploading the data to Dell Support. SAN HQ encrypts and stores local data packages in a folder on
your system under Auto-Pilot. For example: C:\ProgramData\EqualLogic\SAN HQ\Logs\Auto-Pilot
Before running an on-demand data collection, you should decide whether you want to include a SAN HQ
archive file in the data collection (see Group Data Archives) and whether you intend to transmit the data
file to Dell for analysis, or retain it as a local data package.
If you click the Cancel button on the Run Now dialog box while the progress bar is on step 2 or 3, the
diagnostic data uploads are not aborted. However, if you click Cancel in step 1, the data collections are
aborted and the uploads are not created.
After you run an on-demand collection, you can decrypt and analyze local data packages. See Decrypt
Local Data Packages.
If required for analysis and diagnosis, you can manually upload on-demand data packages to Dell if the
packages are requested by Dell Support.
Unlike scheduled data collections, an on-demand data collection does not automatically retry if it fails.
You must manually rerun the on-demand data collection.
1. Check one or more boxes to select PS Series groups, or click the Select all groups button.
2. Select one of the following data transmission options:
You might need the following information before decrypting a data package:
The name of one or more PS Series groups that are configured for SupportAssist and have existing
local data packages. (That is, you have run at least one data collection on one or more of your PS
Series groups.)
If you have many local data packages, you might need to know the date and time that you created a
particular local data package. You can view or search the SupportAssist activity log. See SupportAssist
Activity Log.
The decrypted file is a Unicode plain text file. You can temporarily change its file extension to .txt, and use
a text editor to view or search the data package content.
NOTE: Changing the file extension to .xml might enable you to review the file using an editor or
other tool capable of formatting XML for ease of reading.
If a package does not decrypt, verify that the file still exists in the log directory and that you have the
proper permissions to access the SAN HQ log directory.
Click the SupportAssist link below the Servers and Groups tree and then select Customer Support
Cases in the upper-left panel.
NOTE: The number that appears on the link indicates the total number of cases reported back to
SupportAssist. When you leave the page and return to other views, the number is cleared and
updated as new cases are logged.
From the SAN HQ menu bar, select SupportAssist and then Customer Support Cases.
The Customer Support Cases page appears (see Figure 86. Customer Support Cases). By default, the
Exclude closed cases checkbox is selected to only show active cases. shows the checkbox deselected,
with all cases displayed.
Case numberSupport request (SR) number for your support case opened at Dell
Case statusCurrent status of the case at Dell Support (for example, open, closed, and reopened)
Case titleTitle when referring to this case
ProgressCurrent disposition of this case
Service levelYour service level contract
Service contractYour service warranty status
Service tagPS Series array service tag for which this case was opened
Date openedDate the case was opened by Dell Support
Last updatedDate the case was last updated in the Dell Support system
Group namePS Series group for which the case was opened
Member nameMember name for the service tag
General Troubleshooting
Considerations When Enabling SupportAssist
Performance Impact When Using SupportAssist
SupportAssist Alerts
Display Groups Configured for SupportAssist
Use a Different Server to Upload Diagnostic Information
Multiple SAN HQ servers cannot monitor the same PS Series group. As such, you cannot set up
SupportAssist on multiple SAN HQ servers to collect diagnostic data from the same group.
Pending diagnostic data uploads are deleted when you delete a group configured for SupportAssist
from SAN HQ. Note the following scenarios:
If you remove a group from SAN HQ and choose to delete the log files, SupportAssist log files and
pending uploads are also deleted.
If you remove a group from SAN HQ and choose to keep the log files, SupportAssist log files are
kept but pending uploads are deleted.
SAN HQ creates a read-only user account (by default, autopilot-ro) on the PS Series array that it uses
to collect the diagnostic data. You can change the name of this account when enabling SupportAssist.
SAN HQ installs the sanmond daemon (if it does not already exist) on all group members to gather
diagnostic data. While sanmond is being installed, you can expect minimal or no performance impact
to your group.
SAN HQ configures the syslog settings on the PS Series array to receive the critical events.
The group configuration and diagnostic data collected by SAN HQ is saved locally after scrubbing the
user passwords. SAN HQ deletes this data from the local machine after successfully transmitting it to
the Dell data center.
If you consider any of the data collected by SupportAssist to be restricted (see Data Collected By
SupportAssist), access to SAN HQ should be limited to authorized personnel.
SupportAssist does not automatically adjust scheduled collections to account for daylight saving time
(DST). To update the data-collection schedule and account for daylight saving time, you must rerun
the SupportAssist Configuration wizard. Only new activity log entries will reflect the new time.
SupportAssist Alerts
SAN HQ generates specific alerts for SupportAssist-configured groups.
Table 55. Alerts Generated by SAN HQ for SupportAssist describes these alerts and trigger mechanisms.
For a description of all SAN HQ alerts, see SAN Headquarters Alerts.
Table 55. Alerts Generated by SAN HQ for SupportAssist
SupportAssist Uploads Warning The most recent diagnostic data The group has had a
Failing upload attempt has failed or is successful upload, the
retrying over a period of at least last upload has failed or
two weeks. is retrying, and the failure
has continued for at least
two weeks.
SNMP Notification Warning The SAN HQ Server is unable to A fatal error occurred
Configuration process the SNMP notification when attempting to read
configuration and could not the
read the AlertNotificationCo
AlertNotificationConfig.x nfig.xml file.
ml file. (All groups are notified.)
Sending SNMP Warning The SAN HQ Server failed to SNMP notification failure
Notification Error send an SNMP notification for
the corresponding group.
The Retry Count maximum is reached. This value indicates the number of times SupportAssist
attempts to collect diagnostic data from the PS Series group, after which an automatic upload can
occur.
The automatic upload cannot complete, after which SupportAssist stores the data before attempting
another upload six hours later. If the Internet is still unavailable, SupportAssist continues to upload the
data on six-hour intervals until successful. (If unsuccessful after 90 days, the data is deleted.)
The SupportAssist Configuration Settings page shows groups configured or unconfigured for
SupportAssist. Figure 87. Group Configured for SupportAssist shows a group configured to retry data
collection three times and perform an automatic upload.
NOTE: Optionally, you can disable automatic uploads for a group and retain all data collections
locally. Later, you can manually upload the data. For more information, see Enable and Disable
Automatic Upload for a Group.
You have SAN HQ v3.0 or later installed (see Chapter 2). Your SAN HQ Client and Server meet the
requirements in Requirements for Running the SAN Headquarters Client and Requirements for
Running the SAN Headquarters Server.
SNMPA read-only SNMP Community name is configured under the Group Manager GUI's SNMP
options. If already configured for SAN HQ, you do not have to create one.
SSHSSH is enabled in the Group Manager GUIs Administrators options.
1. Click SupportAssist in the lower left panel. The Support Assist Configuration Settings page displays.
2. In the upper left panel, click Activity Log.
The status of the activity will show Waiting for Retry, as in Figure 88. Activity Log Showing One
Pending Upload:
SupportAssist displays a message acknowledging the request to upload the data. When the operation
completes, the Activity Log shows a payload ID for the data collection (see Figure 90. Successful
Diagnostic Data Upload ).
After a successful upload, the Retry All Pending Uploads menu option appears grayed out, indicating that
there are no uploads with Waiting For Retry status.
To support this environment, you must set up two separate SAN HQ servers:
A SAN HQ "Data Collector" Server that will run the data collections on a regular, weekly schedule and
then save the output to the local file system. This server does not require Internet access; it does not
upload the diagnostic collections directly to Dell.
A SAN HQ "Data Uploader" Server that has Internet connectivity. This server does not need to collect
diagnostic data directly from the PS Series arrays.
The import process consists of adding the group if it does not already exist, creating upload job
definitions as needed (upload requests), and moving the data collection output directory (run directory)
to the C:\ProgramData\EqualLogic\SAN HQ\Logs\Auto-Pilot\ GroupUUID\Online directory.
This directory is considered complete when the data collection output directory has been copied and is
separate from the running of the upload jobs.
If the import process fails to import a run directory, it moves it to the failed import directory (typically, C:
\ProgramData\EqualLogic\SAN HQ\Logs\FailedImport) on the SAN HQ Data Uploader Server.
Import script:
You must write a script to move the data collection output directories that you want to upload
from the SAN HQ offline folder (typically, C:\ProgramData\EqualLogic\SAN HQ\Logs
\Auto-Pilot\ GroupUUID \Offline) on the SAN HQ Data Collector Server (the server
without Internet connectivity) to the C:\ProgramData\EqualLogic\SAN HQ\Logs\Inbox
folder on the SAN HQ Data Uploader Server (the server with Internet connectivity).
Use the following format for naming the data collection output directories: yyyymmddhhMMss_
runid guid.
After the files are copied to the output directory, the copy script must create a file named
importready.txt in the same directory. This file informs SAN HQ that the copy operation has
completed and that the run directory is ready for import.
On the SAN HQ Data Collector Server, you must configure the groups that collect data and not
automatically upload the data.
On the SAN HQ Data Uploader Server, if the group that has data to be uploaded already exists, you
must configure the group to automatically upload data to Dell Support.
You can import only those data collections that contain diagnostic data. Verify that a collection
contains data by confirming that the diagnostics.dg file exists in the run directory.
Day of Week day Accept the current day Select the day on which
or select a day from the SupportAssist collects data
menu (defaults to current weekday).
Start Time HH:MM AM/PM time Accept the current time Select the time at which
or modify the HH:MM SupportAssist collects data
and AM/PM fields (defaults to current clock time).
Retry After NNN minutes 10120 minute intervals Time interval in minutes that
SupportAssist must wait before
retrying a data transmission.
Best Fit Fits the column width to the data for the selected column
Best Fit All Columns Fits the column width to the data for all columns
Display SALA
Use the following command to display SupportAssist license agreement (SALA).
Syntax
Add SupportAssist
SupportAssist command referenceUse the -AddSupportAssist command parameter to configure and
enable SupportAssist.
Syntax
# SANHQClient.exe
-DefaultServer or -ServerName=server-name
-AddSupportAssist
-AcceptSALA
-UserName=host-user-name
-Password=host-password
IPAddress=group-IP-address
-GroupAdminUserName=group-admin-user-name
-GroupAdminPassword=group-admin-password
-ROUserName=read-only-user-name
-ROPassword=read-only-password
-ROPasswordVerify=read-only-password-verify
[-Schedule
-Day=day-of-week
-Hour=time-hour
-Minute=time-minutes
-MaxRetryCount=max-retry-count
-RetryAfter=retry-after-x-minutes]
[-CollectCriticalData]
[-IncludeArchive]
-SysLogServer=IP-address-of-sys-log-server
[-ContactId=default-email-address-to-contact]
[-AutoUpload]
[-UploadKeepCount=number-of-diagnostic-data-package-to-retain]
[-ReceiveEmail]
Table 59. Command Parameters for Enabling SupportAssist describes the command parameters for
enabling SupportAssist. All parameters are required unless indicated as optional.
License AcceptSALA Specifies that you have read and accept the
Agreement SupportAssist license agreement.
Examples
Add Contact
SupportAssist command referenceUse the following command to add customer contact information for
SupportAssist.
Syntax
# SANHQClient.exe
-DefaultServer or -ServerName=server-name
-AddContact
-UserName=host-user-name
-Password=host-password
IPAddress=group-IP-address
-ContactName=customer-contact-name
-EmailAddress=customer-email-address
-WorkPhone=customer-telephone-number
-CellPhone=Alternate-customer-telephone-number
-City=customer's-city
-State=customer's-state
-Country=customer's-country
-Timezone=customer's-timezone
-Language=customer's-local-language
-Company=company-name
-ContactTimes=preferred-contact-time
Table 60. Command Parameters for Adding a Contact describes the command parameters for adding
contact information. All parameters are required unless indicated as optional.
Example
capacity information
outbound replicas 39
E
summary 34 email alerts
chart display 67 hiding disabled 67
CLI email notification
adding a group using 58 configure 72
start SAN HQ GUI from 17 customizing alerts 73
combined graphs 41 requirements 72
command-line interface, EQLXPerf
See CLI assigning domain user name 22
compression event logging 107
overview 103 event priorities 144
prerequisites and considerations 103 events
reporting 104 copying to clipboard 135, 147
232
displaying 145 average I/O rate 128
exporting 134, 147 average I/O size 128
searching 146 average IOPS 128
experimental analysis average latency 128
examples 157, 160, 161, 164, 167, 169 estimated I/O load 128
experimental analysis data I/O load space distribution 128
examples 156 queue depth 128
replication IOPS 128
synchronous replication 128
F I/O information
favorites list experimental analysis 45
add and manage 77 I/O load
add node 78 space distribution 128
add view 78 inbound replicas 39
firewall rules 66 installation
full installation 24 change log file directory 66
configure single sign-on 25
considerations when updating 28
G directory location 65
displaying settings 65
group data log file directory requirements 21
preserving 174 post-installation tasks 25
group data exports pre-installation review 23
export from GUI 185 prerequisites 19
modifying XML file 187 required information 22
using command to export 185 restarting the SAN HQ Server 26
group diagnostics report 180 SAN HQ Client requirements 21
group manager SAN HQ Server requirements 20
launch with single sign-on 73 software location 15, 65
group monitoring update notifications 28
affect on SNMP polling 122 installation procedure
log file data compression 121 client and server 24
manage 69 client only 24
removing a group 69 installation type 65
resuming 73 IOPS
software and firmware dependencies 125 read-write distribution 128
sort list of groups 70 versus latency 128
stopping 73 IP address for group 57
TCP retransmissions 123 iSCSI connection information 46
troubleshooting 121 iSCSI connections view 124
group network address
manage changes 120
group performance 122 L
grpadmin credentials 192
latency 128
Live View
H data 81
prerequisites 82
hardware alarms running a session 82
displaying 130 local replication reserve 126
hardware/firmware information 41 Local Service
hardware/firmware summary 35 SAN HQ Server run as 22
HDD vs SSD performance 172 log file directory
change 66
I location 65
requirements 21
I/O data log files
233
access to 63
data compression 121
directory requirements 21
R
increasing size 71 RAID evaluator
modify 63 evaluate current RAID policy 85
retaining 23 evaluate RAID policy reliability 87
retaining data 26 RAID policy reliability
size 58 alerts 90
uploading backup data 27 FAQs 91
using after SAN HQ update 25 reporting 85
login credentials unknown status 90
deleting 75 recoverable volumes 98, 99, 126
replica compression 103
N replica reserve 101, 126
replication
network inbound replicas 39
infrastructure recommendation 171 replication borrowing 101
managing address changes 120 replication partner 126
requirements for SupportAssist 193 replication summary 36
network data reports,
active ports 129 creating from GUI 176
Ethernet port 129 See also SAN HQ reports
iSCSI connections 129 requirements
link speed 129 PS Series group 19
management network 129 SupportAssist 192
network load 129
network rate 129
TCP retransmissions 129 S
traffic 129
SAN HQ
network for group view 123
compression reporting 104
network information 47
features 11
network load information 48
installation procedure 23
network port information 48
obtaining software 22
network ports view 123
start from CLI 17
network rate 129
start from Windows 15
network rate information 48
SAN HQ alerts 135
notifications 26
SAN HQ Client
accessing log files 22
P cache directory 65
navigation 30
percentile reporting 95th 51 startup settings 66
performance data upgrade 62
capacity data 126 SAN HQ GUI
polling period adding a group 57
sampling period 55 circle graph 55
polling status 130 control appearance 65
post-installation tasks 25 control tooltips 66
PS Series group data description 126
adding from CLI 58 data table 55
requirements 19 display data from different times 53
summary information 33 group information categories 38
pssdiag 189 hiding groups 70
information categories 33
Q navigation 30
network for group view 123
queue depth 128
234
network ports view 123 storage container
summary views 32 information presented by SAN HQ 114
zoom links 54 SupportAssist
SAN HQ reports activity log 212
group data report 174 activity types 213
information required 175 alerts 220
modify XML file for archiving data 183 autopilot read only account 192
modifying XML file 177 command reference 226
types of 174 components and process 189
using a command to create 176 configuration settings 209
SAN HQ Server configuration steps for existing group 204
adding 61 configuration steps when adding new
changing the default server 61 group 197
configuration settings 67 configuration worksheet 195
identifying a running 65 configure contact data 214
pending restarts 69 configure from CLI 207
protocols and ports 20 configuring for an existing group 203
requirements 20 considerations when enabling 220
run as local service 22 customer contact information 200
syslog server 107 data collection 189
settings data collection settings 225
control GUI appearance 65 data sorting options 225
control temperature settings 67 decrypting local data packages 218
control tooltips 66 firmware requirements 192
email 72 general troubleshooting 220
installation 65 group configuration requirements 192
single sign-on network and bandwidth requirements 193
configuring 25 offline data upload prerequisites 222
delete login credentials 75 on-demand data collection 216
disabling 75 on-demand data collection procedure 217
enable 74 prerequisites for configuring 194
entering credentials 58 quick start information 196
launch group manager with 73 read-only account 192
modifying credentials 74 reference information 225
snapshot compression 103 troubleshooting 219
snapshot reserve upload settings 225
space borrowing 101 SupportAssist alerts 220
SNMP community name 19, 192 SupportAssist command reference
SNMP notification Display SALA 226
about 110 SupportAssist configuration settings
configuration file 110 reference information 208
configure from GUI 111 SupportAssist configuration wizard 204
modify configuration 113 SupportAssist data collection
requirements 111 run now 216
troubleshooting 114 Synchronous replication (SyncRep)
SNMP polling 122 about 79
space borrowing states 79
outbound replicas 103 SyncRep 79
statistics for group 101 SyncRep volumes
view statistics 101 monitoring 78
space utilization 126 syslog event logging
SSD usage statistics 92 disabling 108
SSD vs HDD performance 172 syslog notification 192
SSH 192 syslog server
statistics change configuration 108
thin provisioning 126 configure event logging 107
235
disabling 109 information presented by SAN HQ 116
requirements 108 support for 114
syslog server requirements 107 VM,
See virtual machine
T Volume I/O summary 37
volume reserve 126
TCP retransmissions 123 volumes
thin provisioning displaying recoverable 99
statistics 126 displaying SyncRep 79
time range selector 54 identifying by type 56
tooltips queue depth 128
controlling 66 recoverable 126
displaying 55 reported size 126
troubleshooting support for virtual 114
group monitoring problems 121 types 126
SupportAssist 220 undelete 98
VVol,
See virtual volume
U
update notifications 28
X
V XML file
modify to archive data 183
virtual machine modify to create a report 177
information presented by SAN HQ 117 modify to export data 187
virtual volume
236