IBM Network Attched Storage Concepts
IBM Network Attched Storage Concepts
Provides configuration
considerations
Mary Lovelace
Vincent Boucher
Shradha Radhakrishna Nayak
Curtis Neal
Lukasz Razmuk
John Sing
John Tarella
ibm.com/redbooks
International Technical Support Organization
November 2010
SG24-7874-00
Note: Before using this information and the product it supports, read the information in “Notices” on
page vii.
This edition applies to IBM Scale Out Network Attached Storage (SONAS) 1.1.1.
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
The team who wrote this book . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
Stay connected to IBM Redbooks publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xii
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
Contents v
vi IBM Scale Out Network Attached Storage Concepts
Notices
This information was developed for products and services offered in the U.S.A.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area.
Any reference to an IBM product, program, or service is not intended to state or imply that only that IBM
product, program, or service may be used. Any functionally equivalent product, program, or service that does
not infringe any IBM intellectual property right may be used instead. However, it is the user's responsibility to
evaluate and verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not give you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, Armonk, NY 10504-1785 U.S.A.
The following paragraph does not apply to the United Kingdom or any other country where such
provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION
PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR
IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT,
MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of
express or implied warranties in certain transactions, therefore, this statement may not apply to you.
This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.
Any references in this information to non-IBM websites are provided for convenience only and do not in any
manner serve as an endorsement of those websites. The materials at those websites are not part of the
materials for this IBM product and use of those websites is at your own risk.
IBM may use or distribute any of the information you supply in any way it believes appropriate without
incurring any obligation to you.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.
This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to the names and addresses used by an actual business
enterprise is entirely coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the sample
programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore,
cannot guarantee or imply reliability, serviceability, or function of these programs.
The following terms are trademarks of the International Business Machines Corporation in the United States,
other countries, or both:
AIX® HACMP™ Smarter Planet™
DB2® IBM® System p5®
Enterprise Storage Server® pSeries® Tivoli®
FlashCopy® Redbooks® xSeries®
GPFS™ Redbooks (logo) ® z/OS®
Snapshot, and the Network Appliance logo are trademarks or registered trademarks of Network Appliance,
Inc. in the U.S. and other countries.
Microsoft, Windows, and the Windows logo are trademarks of Microsoft Corporation in the United States,
other countries, or both.
UNIX is a registered trademark of The Open Group in the United States and other countries.
Linux is a trademark of Linus Torvalds in the United States, other countries, or both.
Other company, product, or service names may be trademarks or service marks of others.
IBM® Scale Out Network Attached Storage (IBM SONAS) is a Scale Out NAS offering
designed to manage vast repositories of information in enterprise environments requiring
very large capacities, high levels of performance, and high availability.
SONAS provides a range of reliable, scalable storage solutions for a variety of storage
requirements. These capabilities are achieved by using network access protocols such as
NFS, CIFS, HTTP, and FTP.
Utilizing built-in RAID technologies, all data is well protected with options to add additional
protection through mirroring, replication, snapshots, and backup. These storage systems are
also characterized by simple management interfaces that make installation, administration,
and troubleshooting uncomplicated and straightforward.
In this IBM Redbooks® publication, we discuss the marketplace requirements that led to the
SONAS stand-alone storage system. We introduce storage architectures and file systems.
We then introduce the reader to SONAS and describe the hardware and software that make
up the SONAS appliance.
Shradha Radhakrishna Nayak is a Staff Software Engineer working with IBM India software
Labs in Pune, India. She holds a Bachelor of Computer Science Engineering degree and has
over six years of experience. She has been working in the storage domain since and has
good expertise in Scale out File Service (SoFS) and Scale Out Network Attached Storage
(SONAS). Prior to this, she worked as a Level-3 developer for Distributed File Service (DFS)
and also worked for IBM AFS. Shradha is focusing on storage products and cloud storage
and is currently part of the Level-3 developers teams for SONAS. Being a part of the SONAS
developing and testing team, she has developed a thorough knowledge of SONAS with its
components and functions. In this book, she has mainly focused on installation, configuration,
and administration of SONAS. Shradha is also interested in social media and social
networking tools and methodologies.
John Sing is an Executive IT Consultant with IBM Systems and Technology Group. John has
specialties in large Scale Out NAS, in IT Strategy and Planning, and in IT High Availability and
Business Continuity. Since 2001, John has been an integral member of the IBM Systems and
Storage worldwide planning and support organizations. He started in the Storage area in
1994 while on assignment to IBM Hong Kong (S.A.R. of China), and IBM China. In 1998,
John joined the IBM Enterprise Storage Server® Planning team for PPRC, XRC, and IBM
FlashCopy®. He has been the marketing manager for these products, and in 2002, began
working in Business Continuity and IT Strategy and Planning. Since 2009, John has also
added focus on IT Competitive Advantage strategy, including Scale Out NAS and Cloud
Storage. John is the author of three Redbooks publications on these topics, and in 2007,
celebrated his 25th anniversary of joining IBM.
John Tarella is a Senior Consulting IT Specialist who works for IBM Global Services in Italy.
He has 25 years of experience in storage and performance management in mainframe and
distributed environments. He holds a degree in Seismic Structural Engineering from
Politecnico di Milano, Italy. His areas of expertise include IBM Tivoli Storage Manager and
storage infrastructure consulting, design, implementation services, open systems storage,
and storage performance monitoring and tuning. He is presently focusing on storage
solutions for business continuity, information lifecycle management, and infrastructure
simplification. He has written extensively on z/OS DFSMS, IBM Tivoli Storage Manager,
SANs, storage business continuity solutions, content management, and ILM solutions. John
is currently focusing on cloud storage delivery. He also has an interest in Web 2.0 and social
networking tools and methodologies.
Sven Oehme
Mark Taylor
Alexander Saupp
Mathias Dietz
Jason Auvenshine
Greg Kishi
Scott Fadden
Leonard Degallado
Todd Neville
Warren Saltzman
Wen Moy
Tom Beglin
Adam Childers
Find out more about the residency program, browse the residency index, and apply online at:
ibm.com/redbooks/residencies.html
Comments welcome
Your comments are important to us!
We want our books to be as helpful as possible. Send us your comments about this book or
other IBM Redbooks publications in one of the following ways:
Use the online Contact us review Redbooks publications form found at:
ibm.com/redbooks
Send your comments in an email to:
[email protected]
Mail your comments to:
IBM Corporation, International Technical Support Organization
Dept. HYTD Mail Station P099
2455 South Road
Poughkeepsie, NY 12601-5400
Preface xi
Stay connected to IBM Redbooks publications
Find us on Facebook:
http://www.facebook.com/IBMRedbooks
Follow us on twitter:
http://twitter.com/ibmredbooks
Look for us on LinkedIn:
http://www.linkedin.com/groups?home=&gid=2130806
Explore new Redbooks publications, residencies, and workshops with the IBM Redbooks
publications weekly newsletter:
https://www.redbooks.ibm.com/Redbooks.nsf/subscribe?OpenForm
Stay current on recent Redbooks publications with RSS Feeds:
http://www.redbooks.ibm.com/rss.html
In this chapter, we consider how a high-density, high-performance SONAS system can help
organizations consolidate and manage data affordably, reduce crowded floor space, and
reduce management expense associated with administering an excessive number of
disparate storage systems. With its advanced architecture, SONAS virtualizes and
consolidates multiple filers into a single, enterprise-wide file system, which can translate into
reduced total cost of ownership, reduced capital expenditure, and enhanced operational
efficiency.
There is an explosion in the amount of data, of course, but also there are shifts in the nature
of data (Figure 1-1). Formerly, virtually all the information available to be “processed” was
authored by someone. Now that kind of data is being overwhelmed by machine-generated
data, spewing out of sensors, RFID, meters, microphones, surveillance systems, GPS
systems, and all manner of animate and inanimate objects. With this expansion of the
sources of information comes a large variance in the complexion of the available data—very
noisy, lots of errors—and no time to cleanse it in a world of real-time decision making.
Also, consider that today’s economic times require corporations and governments to analyze
new information faster and make timely decisions for achieving business goals. As the
volume, variety, and velocity of information and decision making increases, this places a
larger burden on organizations to effectively and efficiently distribute the right information, at
the right time, to the people, processes, and applications that are reliant upon that information
to make better business decisions.
All of these situations are creating challenges, while providing an excellent opportunity for
driving an information-led transformation.
Innovative applications in business analytics, digital media, medical data, and cloud storage
are creating requirements for data access rates and response times to individual files that
were previously unique to high-performance computing environments, and all of this is driving
a continuing explosion of business data.
While many factors are contributing to data growth, these trends are significant contributors:
Digital representation of physical systems and processes
Capture of digital content from physical systems and sources
Deliveries of digital content to a global population
For example, suppose that Windows® 2000 formats a partition (or drive) and maps that
partition to your system. Every time you request to open data on that partition, your request is
processed by Windows 2000. Because there is a file system on the partition, it is accessed by
File I/O. Additionally, you cannot request access to just the last 10 KB of a file. You must open
the entire file, which is another reason that this method is referred to as File I/O.
Using File I/O is like having an accountant. Accountants are good at keeping up with your
money for you, but they charge you for that service. For your personal checkbook, you
probably want to avoid that cost. On the other hand, for a corporation where many kinds of
requests are made, having an accountant is a good idea, so that wrongful checks do not get
written.
A file I/O specifies the file. It also indicates an offset into the file (Figure 1-3). For instance, the
I/O might specify “Go to byte ‘1000’ in the file (as if the file were a set of contiguous bytes),
and read the next 256 bytes beginning at that position.” Unlike block I/O, there is no
awareness of a disk volume or disk sectors in a file I/O request. Inside the NAS appliance, the
operating system keeps track of where files are located on disk. It is the NAS OS which
issues a block I/O request to the disks to fulfill the client file I/O read and write requests that it
receives.
By default, a database application that is accessing a remote file located on a NAS device is
configured to run with File System I/O. It cannot utilize raw I/O to achieve improved
performance.
An example of this might be IBM DB2® using its tables to keep track of where data is located
rather than letting the OS do that job. We do not mean to say that DB2 cannot use the OS to
keep track of where files are stored. It is just more efficient for the database to bypass the cost
of requesting the OS to do that work.
One of the key differences in a NAS appliance, compared to direct attached storage (DAS) or
other network storage solutions such as SAN or iSCSI, is that all client I/O operations to the
NAS use file level I/O protocols. File I/O is a high level type of request that, in essence,
specifies only the file to be accessed, but does not directly address the storage device. This is
done later by other operating system functions in the remote NAS appliance.
By making storage systems LAN addressable, the storage is freed from its direct attachment
to a specific server, and any-to-any connectivity is facilitated using the LAN fabric. In
principle, any user running any operating system can access files on the remote storage
device. This is done by means of a common network access protocol; for example, NFS for
UNIX® servers and CIFS for Windows servers. In addition, a task such as backup to tape can
be performed across the LAN using software such as Tivoli Storage Manager, enabling
sharing of expensive hardware resources (for example, automated tape libraries) between
multiple servers.
A storage device cannot just attach to a LAN. It needs intelligence to manage the transfer and
the organization of data on the device. The intelligence is provided by a dedicated server to
which the common storage is attached. It is important to understand this concept. NAS
comprises a server, an operating system, and storage that is shared across the network by
many other servers and clients. So a NAS is a specialized server or appliance, rather than a
network infrastructure, and shared storage is attached to the NAS server.
This situation is compounded by the fact that at hundreds of terabytes or more, conventional
backup of such a large storage farm is difficult, if not impossible. There is also the issue that
even though one might be using incremental only backup, scanning hundreds of terabytes to
identify the changed files or changed blocks might in itself take to long, with too much
overhead.
More issues include the possibility that there might not be any way to apply file placement,
migration, deletion, and management policies automatically from one centrally managed,
centrally deployed control point. Doing manual management of tens or hundreds of filers was
proving to be neither timely nor cost-effective, and effectively prohibited any feasible way to
globally implement automated tiered storage.
Simply put, SONAS is a scale out storage system combined with high-speed interface nodes
interconnected with storage capacity and GPFS, which enables organizations to scale
performance alongside capacity in an integrated, highly-available system. The high-density,
high-performance SONAS can help your organization consolidate and manage data
affordably, reduce crowded floor space, and reduce management expense associated with
administering an excessive number of disparate storage systems.
Assuming 2 TB SATA disk drives, such a system has 14.4 petabytes (PB) of raw storage and
billions of files in a single large file system (up to 2 PB as standard support; larger is possible
by request to IBM). You can have as few as eight file systems in a fully configured 14.4 PB
SONAS system or as many as 256 file systems. It provides an automated policy-based file
management that controls backups and restores, snapshots, and remote replication. It also
provides:
A single global namespace with logical paths that do not change because of physical data
movement
Support for Serial Attached SCSI (SAS) and Serial Advanced Technology Attachment
(SATA) drives
Superior performance per price
High-availability and load-balancing
Centralized management
Centralized backup
An interconnected cluster of file-serving and network-interfacing nodes in a redundant
high-speed data network
Virtually no capacity limits
Virtually no scalability limits
IBM Call Home trouble reporting and IBM Tivoli Assist On Site (AOS) remote support
capabilities
Global namespace
SONAS provides a global namespace that enables your storage infrastructure to scale to
extreme amounts of data, from terabytes to petabytes. Within the solution, centralized
management, provisioning, control, and automated information life-cycle management (ILM)
are integrated as standard features to provide the foundation for a truly cloud storage enabled
solution.
Interface nodes
The high-performance interface nodes provide connectivity to your Internet Protocol (IP)
network for file access and support of both 1-gigabit Ethernet (GbE) and 10-GbE connection
speeds. Each interface node can connect to the IP network with up to eight separate
data-path connections. Performance and bandwidth scalability are achieved by adding
interface nodes, up to the maximum of 30 nodes, each of which has access to all files in all
file systems.
You can scale out to thirty interface nodes. Each interface node has its own cache memory,
so you increase caching memory and data paths in your file-serving capacity by adding an
interface node. Of course, you also increase file-serving processor capacity. If raw storage
capacity is the prime constraint in the current system, the SONAS system scales out to as
much as 14.4 petabytes (PB) with 2 terabytes (TB) SATA drives, with up to 256 file systems
that can each have up to 256 file-system snapshots. Most systems that a SONAS system
typically displaces cannot provide clients with access to so much storage from a single
file-serving head. Every interface node has access to all of the storage capacity in the
SONAS system.
SONAS is a scalable virtual file storage platform that grows as data grows. It meets
demanding performance requirements as new processors can be added independently or as
storage capacity is added, eliminating a choke point found in traditional scale-up systems.
SONAS is designed for high availability 24x7 environments with a clustered architecture that
is inherently available and, when combined with the global namespace, allows for much
higher utilization rates than found in scale-up environments.
Rich, sophisticated policies built into SONAS can transparently migrate data between pools
based on many characteristics, such as capacity threshold limits and age of the data. This
helps to address business critical performance requirements. Leveraging automated storage
tiering, users can finally realize the cost savings and business benefits of information lifecycle
management (ILM) at an immense scale.
In the event of data corruption or an unexpected disaster that might harm corporate data,
SONAS helps you to recover and quickly resume normal enterprise and data center
operations (Figure 1-9).
SONAS supports large enterprise requirements for remote replication, point-in-time copy
(file system-level snapshots), and scalable automated storage tiering, all managed as a single
instance within a global namespace. SONAS asynchronous replication is specifically
designed to cope with connections that provide low bandwidth, high latency, and low
reliability. The async scheduled process will pick up the updates on the source SONAS
system and write them to the target SONAS system, which can be thousands of miles away.
1.3.5 Summary
SONAS is a highly desired choice for organizations seeking to better manage their growing
demand for file-based storage. SONAS is designed to consolidate data that is scattered in
multiple storage locations and allows them to be efficiently shared and managed. The
solution helps improve productivity by providing automated ILM, automatic storage allocation,
user management by storage quota, and universal reporting and performance management.
We illustrate these benefits by following the lifecycle of a file from creation to deletion. We
demonstrate how files are written, centrally managed and migrated, backed up, archived,
remotely replicated, and finally deleted from the IBM SONAS file system.
In the following topics, the graphics show how SONAS physically manages files, without
changing the logical appearance to the users. We can see how files that logically appear to be
in the same directory might physically be in separate pools; how these files are created in
particular physical pools by a central SONAS policy engine; and how SONAS manages and
moves these files from creation to deletion, according to central policies.
Logical
Logical Diagram of IBM SONAS
/home/appl/data/web/important_big_spreadsheet.xls
/home/appl/data/web/important_big_spreadsheet.xls
/home
/home/appl/data/web/big_architecture_drawing.ppt
/ho me/appl/data/web/big_architecture_drawing.ppt
/app l
/data /home/appl/data/web/unstructured_big_video.mpg
/ho me/appl/data/web/unstructured_big_video.mpg
/web
Global Namespace
Policy Engine
Inte rfa c e
node s … Inter fac e
nodes … Interfa ce
node s … > scale
out
Stor age
node s
….. Stor age
node s
….. St ora ge
nodes
... > scale
out
Storage Pool (Tier 1): SAS drives Storage Pool (Tier 2) with 1TB SATA Storage Pool (Tier 3) with 2TB SATA
In the top half of this diagram, we see the logical file directory structure as seen by the users.
SONAS presents and preserves this same logical appearance to the users, no matter what
we do to physically manage these files (and all files in the SONAS system) from creation to
deletion. The users see only their global namespaces. As we expand and change the physical
data location and supporting physical infrastructure, the users will still have the unchanged
appearance of one single global namespace.
In the lower half of this diagram, we see a physical representation of the SONAS architectural
components. SONAS has interface nodes that serve data to and from the users over the
network. SONAS also has storage nodes, which service the storage for the SONAS clustered
file system.
All SONAS nodes are in a global cluster, connected by Infiniband. All interface nodes have full
read/write access to all storage nodes. All storage nodes have full read/write access to all
interface nodes. Each of the nodes runs a copy of IBM SONAS Software (5639-SN1), which
provides all the functions of SONAS, including a Cluster Manager that manages the cluster
and dispatches workload evenly across the cluster. Also included is the SONAS storage
policy engine, for managing the lifecyle of all files in a centrally deployed, centrally managed
way. The policy engine function is not tied to a particular node, it executes in a distributed
manner across all nodes. Not shown are the SONAS management nodes, which monitor the
health of the SONAS system.
Physical disk drives are allocated to SONAS logical storage pools. Typically, we might
allocate a high performance pool of storage (which uses the fastest disk drives), and a lower
tier of storage for capacity (less expensive, slower spinning drives). In the foregoing example,
we have allocated three logical storage pools.
An interface node is also selected to handle this client. The incoming workload is IP-balanced
equally by the external network Domain Name Server, across all SONAS interface nodes.
SONAS Software works in conjunction with the external Domain Name Server (DNS) to
allocate the interface node. As shown in Figure 2-3, we have selected an interface node.
Logical
/home/appl/data/web/big_architecture_drawing.ppt
/ho me/appl/data/web/big_architecture_drawing.ppt
/app l
/data /home/appl/data/web/unstructured_big_video.mpg
/ho me/appl/data/web/unstructured_big_video.mpg
/web
Storage Pool (Tier 1): SAS drives Storage Pool (Tier 2) with 1TB SATA Storage Pool (Tier 3) with 2TB SATA
The interface node takes the incoming create file request and passes the write request to all
appropriate storage nodes. The storage nodes, in parallel, perform a large data striped write
into the appropriate logical storage pool, across multiple disk drives.
SONAS data writes are done in a wide parallel data stripe write, across all disk drives in the
logical storage pool. SONAS Software architecture aggregates the file write and read
throughput of multiple disk drives, thus providing high performance. SONAS Software will
write the file in blocksize chunks, according to the blocksize specified at the file system level.
This wide data striping architecture determines where the data blocks must go, thus providing
SONAS Software with the ability to automatically tune and load balance all disk drives in the
storage pool. This process is illustrated in Figure 2-4.
Logical
/home/appl/data/web/big_architecture_drawing.ppt
/ho me/appl/data/web/big_architecture_drawing.ppt
/app l
/data /home/appl/data/web/unstructured_big_video.mpg
/ho me/appl/data/web/unstructured_big_video.mpg
/web
Global Namespace
Policy Engine
Inte rfa c e
node s … Inter fac e
nodes … Inte rfa ce
node s … > scale
out
Physica l
Stor age Stor age St ora ge
node s node s nodes
/home/a ppl/data/web/important_big_spr eadsheet.xls out
Storage Pool (Tier 1): SAS drives Storage Pool (Tier 2) with 1TB SATA Storage Pool (Tier 3) with 2TB SATA
Figure 2-4 Create and write file 1 - step 2 - wide parallel data stripe write
Now, let us write another file to SONAS. As shown in Figure 2-5, another interface node is
appropriately selected by the Domain Name Server, and the file is passed to that interface
node for writing.
Logical
/ho me/appl/data/web/big_architecture_drawing.ppt
/home/appl/data/web/ big_architecture_drawing.ppt
/app l
/home/appl/data/web/big_architecture_drawing.ppt
Global Namespace
Policy Engine
Inte rfa c e
node s … Inter fac e
nodes … Interfa ce
node s … > scale
out
Physica l
Stor age Stor age St ora ge
node s node s nodes
out
Storage Pool (Tier 1): SAS drives Storage Pool (Tier 2) with 1TB SATA Storage Pool (Tier 3) with 2TB SATA
Notice that another interface node has been chosen; this illustrates the automatic balancing
of the incoming workload across the interface nodes. The interface node is told by the policy
engine that this file is to be written to the 1 TB intermediate pool. In the same manner as
previously described, the file is written in a wide data stripe (Figure 2-6).
Logical
/home/appl/data/web/big_architecture_drawing.ppt
/ho me/appl/data/web/big_architecture_drawing.ppt
/app l
/data /home/appl/data/web/unstructured_big_video.mpg
/ho me/appl/data/web/unstructured_big_video.mpg
/web
Global Namespace
Policy Engine
Inte rfa c e
node s … Inter fac e
nodes … Inte rfa ce
node s … > scale
out
Storage Pool (Tier 1): SAS drives Storage Pool (Tier 2) with 1TB SATA Storage Pool (Tier 3) with 2TB SATA
Figure 2-6 Create and write file 2 - step 2 - wide parallel data stripe write
Logical
/home/appl/data/web/big_architecture_drawing.ppt
/ho me/appl/data/web/big_architecture_drawing.ppt
/app l
/data /home/appl/data/web/unstructured_big_video.mpg
/ho me/appl/data/web/unstructured_big_video.mpg
/web
/home/appl/data/web/unst ructured_big_video.mpg
Global Namespace
Policy Engine
Inte rfa c e
node s … Inter fac e
nodes … Inte rfa ce
node s … > scale
out
Physica l
Stor age Stor age St ora ge
node s node s nodes
out
Storage Pool (Tier 1): SAS drives Storage Pool (Tier 2) with 1TB SATA Storage Pool (Tier 3) with 2TB SATA
Logical
/hom
/home/appl/data/web/big_architecture_drawing.ppt
e/appl/data/web/big_architecture_drawing.ppt
/appl
/data /hom
/home/appl/data/web/unstructured_big_video.mpg
e/appl/data/web/unstructured_big_video.mpg
/web
Global Namespace
Policy Engine
Interface
nodes … Interface
nodes … Interface
nodes …> scale
out
Physica l
Storage
nodes
Storage
nodes
Storage
nodes scale out
/home/appl/data/web/unstructured_big_video.mpg
Storage Pool (Tier 1): SAS drives Storage Pool (Tier 2) with 1TB SATA Storage Pool (Tier 3) with 2TB SATA
Figure 2-8 Create and write file 2 - step 2 - wide parallel data stripe write
Logical
/home/appl/data/web/big_architecture_drawing.ppt
/ho me/appl/data/web/big_architecture_drawing.ppt
/app l
/data /home/appl/data/web/unstructured_big_video.mpg
/ho me/appl/data/web/unstructured_big_video.mpg
/web
Note: all three files,
in same directory,
but each allocated to
Workload auto- different physical
balanced across all storage pool
IBM Scale Out NAS
interface nodes
Global Namespace
Policy Engine Data striped across
Inte rfa c e
node s … Inter fac e
nodes … Inte rfa ce
node s … > scale
out
all disks in storage
pool.
High performance,
auto-tuning, auto-
load balancing
….. ….. ... > scale
Physica l
Stor age Stor age St ora ge
node s node s nodes
out
Storage Pool (Tier 1): SAS drives Storage Pool (Tier 2) with 1TB SATA Storage Pool (Tier 3) with 2TB SATA
In summary, SONAS workload is automatically balanced across all interface nodes. Data is
striped across all disks in the logical storage pool, providing high performance, automatic
tuning, and automatic load balancing. Notice that from the user’s perspective, these three
files all reside in the same logical path and directory. The user does not know that automatic
physical tiered storage has been implemented. The SONAS Software will continue to
maintain this same logical file structure and path, regardless of physical file location changes
due to automatic tiered storage data movement.
Logical
/appl
/data
/web
Global Namespace
Policy Engine
Inte rf ac e
node
… Inter fa ce
node
… Inte rf ac e
node
…> scale
out
Physical
Stora ge
node s
….. Stor age
node s
….. St or age
nodes
... > scale
out
/home/appl/data/web/im portant_bi g_spreadsheet.xls
Figure 2-10 Scale out more disk drives for more write performance
SONAS architecture thus provides the ability to scale out both the number of disk drives and
the number of storage nodes that can be applied to support a higher amount of parallel
physical data writes.
In this way, SONAS architecture is very flexible, providing the ability to expand the scale and
capacity of the system in any direction that is desired. Disk drives and storage nodes can be
added non-disruptively to SONAS. Immediately upon doing so, SONAS automatically starts
to auto-balance and auto-tune new workload onto the additional disks, and automatically
starts taking advantage of the additional resources.
Logical
/appl
/data
/web
Policy Engine
Interface
nodes … Interface
nodes … Interface
nodes …> scale
out
Physical
Storage
nodes
….. Storage
nodes
….. Storage
nodes
... > scale
out
/home/appl/data/web/important_big_spreadsheet.xls
Furthermore, the interface node is designed to utilize advanced algorithms that improve
read-ahead and write-behind file functions, and recognizes and does intelligent pre-fetch
caching of typical access patterns such as sequential, reverse sequential, and random
(Figure 2-12).
Logical
/app l
/data
/web
Interface node
performs read-ahead
caching, intelligent
pre-fetch
IBM Scale Out NAS
Global Namespace
Policy Engine
Inte rfa c e
node s … Inter face
nodes … Inte rfa ce
node s … > scale
out
Physica l
Stor age
node s
….. Stor age
node s
….. St ora ge
nodes
... > scale
out
Logical
/app l
Interface node
/data
performs read-ahead
/web caching, intelligent
pre-fetch
Physica l
Stor age
node s
….. Stor age
node s
….. S tora ge
nodes
... > scale
out
/ho me/app l/d ata /we b/impor ta nt _big_ s pr ea ds he et .xls\
Figure 2-13 Scale out more disk drives for read performance - parallel data stripe read
The scale-out architecture of SONAS provides superb parallel performance, especially for
larger data objects, and excellent performance for large aggregates of smaller objects.
As we have seen, SONAS is uniquely able to provide a scale out NAS architecture that can
grow to 30 interface nodes and 60 storage nodes, all in a global active-active
share-everything cluster. SONAS is based on the IBM General Parallel File System, which
has proven the ability to scale to thousands of nodes. IBM is scaling down that capability and
providing access to this capability in an appliance form factor, SONAS.
Logical
/home/appl/data/web/big_architecture_design.ppt
/app l
/data /home/appl/data/web/unstructured_big_video.mpg
/web
Physica l
Stor age
node s
….. Stor age
node s
….. S tora ge
nodes
... > scale
out
/home/appl/data/web/important_big_sprea dshee t.xls
The value of a scale out architecture is the ability to flexibly and dynamically add nodes as
needed to increase the amount of parallel concurrent users that can be supported. Each
individual node works in parallel to service clients, as shown in Figure 2-15.
Logical
/home/appl/data/web/big_architecture_design.ppt
/app l
/data /home/appl/data/web/unstructured_big_video.mpg
/web
Stor age
node s
….. Stor age
node s
….. S tora ge
nodes
... > scale
out
In the following topics, we examine how SONAS provides integrated, automated tools to help
accomplish these goals.
SONAS logical storage pools allow you to allocate physical storage hard drives to logical
storage pools within the SONAS file system. Using logical storage pools, you can create tiers
of storage by grouping physical disk storage based on performance, locality or reliability
characteristics. Logical storage pools can span storage nodes and storage pods.
/h ome /home/appl/data/web/file_in_storage_pool_1.xls
/ho me/a pp l /d ata /we b/im por tant _big_s pre ads hee t.xls
/ap pl
/home/appl/data/web/file_in_storage_pool_2.ppt
/h om e/ap p l/d ata/w eb /big_ ar chite ctur e _dra wing.ppt
/d ata
/w eb
/home/appl/data/web/file_in_storage_pool_3.mpg
/h om e/ap p l/d ata/w eb /unstr uctur ed_ big_ video.m pg
Global Namespace
Policy Engine scale
In e
t rfa c
e n od es
In e
t rfa c
e n od es
I nt er f a c
e n ode s
I nt er f ac
e no de s
I nt er f ac
e no de s
In e
t r f ac
e no des
In e
t rfa c
e n od es
In e
t rfa c
e n od es …> out
scale
out
Stor ag
e
nodes
Stor ag
e
nodes
t orag
S
e
nodes
St orag
e
nodes
St orag
e
nodes
St orag
e
nodes
St orag
e
nodes
St roag
e
nodes
Stor ag
e
nodes
Stor ag
e
nodes
Stor ag
e
nodes
Storag
e
nodes
St or g
e
nodes
a St orag
e
nodes
St orag
e
node s
St orag
e
nodes
St orag
e
nodes …>
scale
…> out
Logical storage pool #1 might be high performance SAS disks, and logical storage pool #2
might be more economical SATA 1 TB disk storage. Logical storage pool #3 might be a 2 TB
SATA pool with Hiearchical Storage Management, where the data is intended to be staged in
and out of SONAS, to external tape or data de-deduplication technology. Within the internal
SONAS logical storage pools, all of the data management, from creation to physical data
movement to deletion, is done by SONAS Software.
Attention: If all data movement is within a SONAS system, then an external Tivoli Storage
Manager server is not necessary.
SONAS Software provides the file management concept of a fileset. A fileset is a sub-tree of
the file system namespace and provides a way to partition the global namespace into smaller,
more manageable units. A fileset is basically a named collection of files that you want to
operate upon or maintain using a fileset name. Filesets provide an administrative boundary
that can be used to set quotas and be specified in a user defined policy, to control initial data
placement or data migration.
As shown in Figure 2-16 on page 27, data in a single fileset can reside in one or more logical
storage pools. As the data is physically migrated according to storage policy, the fileset
grouping continues. Where the file data actually resides, and how and when it is migrated, is
based on a set of rules in a user defined policy, and this action is performed by the central
policy engine, which is described next.
File placement policies determine which storage pool file the data is initially placed in. File
placement rules are determined by attributes known when a file is created, such as file name,
user, group, or the fileset. Here are a few examples:
‘place all files that end in .avi onto the silver storage pool’
‘place all files created by the performance critical applications in the gold storage pool’
‘place all files in the fileset ‘development’ in the copper pool’
Files written to SONAS are physically placed according to these rules, which are contained in
a SONAS storage policy. The SONAS administrator writes these rules, which are SQL-like
statements.
• rule 'cleangold' migrate from pool ‘TIER1' threshold (90,70) to pool ‘TIER2‘
• rule 'cleansilver' when day_of_week()=Monday migrate from pool 'silver' to pool 'bronze'
where access_age > 30 days
• rule 'purgebronze' when day_of_month()=1 delete from pool 'bronze' where access_age>365
days
When files exist in a file system, file management policies allow you to move, change the
replication status, or delete files. You can use file management policies to move data from one
pool to another, without changing the file’s location in the directory structure.
The rules are very flexible; as an example, you can write a rule that says: ‘replicate all files in
/database/payroll which have the extension *.dat and are greater than 1 MB in size to storage
pool 2’. In addition, file management policies allow you to prune the file system, deleting files
as defined by policy rules.
File management policies can use more attributes of a file than file placement policies,
because when a file exists, there is more known about the file. In addition to the file
placement attributes, the policies can now utilize attributes such as last access time, size of
the file or a mix of user and file size. This can result in policy statements such as: ‘delete all
files with a name ending in .temp that have not been accessed in 30 days’, ‘move all files that
are larger than 2 GB to pool2’, or ‘migrate all files owned by GroupID=Analytics that are larger
than 4 GB to the SATA storage pool’.
The threshold option comes with the ability to set high low and pre-migrate thresholds,
meaning that SONAS begins migrating data at the high threshold, until the low threshold is
reached. If a pre-migrate threshold is set, SONAS begins copying data until the pre-migrate
threshold is reached, thus allowing the data to continue to be accessed in the original pool
until it is quickly deleted to free up space the next time the high threshold is reached.
Policy rule syntax is based on the SQL 92 syntax standard and supports multiple complex
statements in a single rule enabling powerful policies. Multiple levels of rules can be applied
because the complete policy rule set is evaluated for each file when the policy engine
executes.
As the number of files continues to grow, the time required for this scan using the traditional
method of “walking the directory trees” has often become the major obstacle to effective
storage management. Shrinking backup and storage management windows require scan
times to stay small or even shrink, even as the file systems continue grow from hundreds of
terabytes to petabytes. It is increasingly clear that at the level of petabyte storage scalability,
which is becoming more commonplace every day, walking the directory trees to identify files
is no longer feasible; it simply takes too long.
The SONAS scan engine is an integrated part of the file system. Also integrated into the
SONAS file system is an internal database of file system metadata, which is specifically
designed for the integrated scan engine. The goal of these two functions is to provide the
ability to scan the file system very quickly, at any scale, extending to billions of files.
Scan Engine •Scan Engine reads internal SONAS file system metadata
•Does not need to read the file or directory tree
/home
•All nodes can participate in scan of file system
/app l
/data
/web
Central policy
engine starts scan
by reading policies
Figure 2-18 High performance scan engine - start scan by reading policies
The SONAS high performance scan engine is designed to utilize the multiple nodes of the
SONAS system in parallel, to scan the internal file system metadata. The multiple nodes
equally spread the policy engine rule evaluation, file scan identification, and subsequent data
movement responsibilities over the multiple nodes in the SONAS cluster.
Scan Engine •Scan Engine reads internal SONAS file system metadata
•Does not need to read the file or directory tree
/home
•All nodes can participate in scan of file system
/app l
/data
Global Namespace
Policy Engine
Inter fac e Inte r fac e Int er face
node
Interface nodes node node
Figure 2-19 High performance scan engine - parallel scan of metadata by all nodes
The results of the parallel scan are aggregated, and returned as the actionable list of
candidate files, as shown in Figure 2-20.
Scan Engine •Scan Engine reads internal SONAS file system metadata
•Does not need to read the file or directory tree
/home
•All nodes can participate in scan of file system
/app l
/data
Scan results
/web completed in much
shorter period of
time, compared to
traditional methods
IBM Scale Out NAS
Scale
Out
Stor age nodes Stor age nodes St or a ge node s S torage node s St ora ge node s S tora ge node s
Figure 2-20 High performance scan engine - return results of parallel scan
Now that we have used the scan engine to identify candidate files for automated storage
management, let us see how the parallel grid architecture of SONAS is used to scale out
physical data movement.
Figure 2-21 shows an example of physically moving a file from “Storage Pool 1” to ‘”Storage
Pool 2”.
/app l
/home/appl/data/web/big_architecture_drawing.ppt
/data
/home/appl/data/web/unstructured_big_video.mpg
/web
Scale
Out
Stor age nodes Stor age nodes St or a ge node s S torage node s St ora ge node s S tora ge node s
Figure 2-21 High performance parallel data movement for ILM - pool 1 to pool 2
All files remain online and fully accessible during this physical data movement; the logical
appearance of the file path and location to the user does not change. The user has no idea
that the physical location of their file has moved. This is one of the design objectives of
SONAS.
According to the results of the scan, SONAS continues with other physical file movement.
According to policy, data can be up-staged as well as down-staged, as shown in Figure 2-22.
/home/appl/data/web/important_big_spreadsheet.xls
/home
/app l
/home/appl/data/web/big_architecture_drawing.ppt
/data
/home/appl/data/web/unstructured_big_video.mpg
/web
Scale
Out
Stor age nodes Stor age nodes St or a ge node s S torage node s St ora ge node s S tora ge node s
Figure 2-22 High performance parallel data movement for ILM - pool 2 to pool 1
As the SONAS system grows in capacity over time, it is a straightforward matter to add
additional nodes to the parallel cluster, thus maintaining the ability to perform and complete
file system scans and physical data movement in a timely manner, even as the file system
grows into hundreds of terabytes and petabytes.
2.4.5 Summary
The SONAS parallel scan engine provides the following functions:
Reads policy engine policies
Identifies files to be moved within the physically tiered storage or sent to remote sites
Enables and makes feasible automated tiered storage at terabyte and petabyte scale
The SONAS high performance scan engine has the following characteristics:
Does not need to read the file or directory tree
Reads special metadata integrated and maintained by the file system
Is designed so that all nodes can participate in a parallel scan of the file system
Delivers very high performance scan with minimized impact on concurrent workloads
Can perform scans on a frequent basis due to low overhead
As long as the data movement is within the SONAS system, or between SONAS devices,
then all physical data movement is done solely through SONAS Software, and no involvement
of any external servers or external software is involved.
This combination of the internal file system metadata and the SONAS scale out parallel grid
software architecture enables SONAS to provide an architectural solution to scanning the file
systems quickly and efficiently, at the level of millions and billions of files in a short period of
time. The SONAS Software integrated high performance scan engine and data movement
engine work together to make feasible the management of automated tiered storage, with
physical data movement transparent to the users, at the level of hundreds of terabytes to
petabytes in the file systems.
Traditional software that performs these functions can accomplish these functions on SONAS,
by walking the directory trees, identifying candidate files through normal means, and
performing normal LAN I/O to do data movement. In this case, the normal parameters of file
system scan time will apply.
Tip: IBM has an extensive Independent Software Vendor (ISV) certification program for
SONAS. Enterprises use many ISV applications for their storage management to address
business requirements. IBM has done extensive testing and intends to continue to ensure
interoperability and compatibility of the leading ISV applications with SONAS to reduce
deployment risks. See 5.1.3, “ISV software supported” on page 144.
2.5.1 Requirements for high performance external HSM and backup restore
However, as file systems continue grow from hundreds of terabytes to petabytes, the following
requirements have arisen:
The elapsed time is becoming too long for traditional scans to identify files that have to be
moved for HSM or backup/restore purposes. This occurs because of the requirement of
walking the directory trees, and due to the scale of these searches, incurs a very large
amount of small block IOPs.
These long scan times are severely inhibiting the ability to manage a large amount of
storage. In many cases, the scan time alone can be longer than the backup window.
In order to make feasible automated tiered storage at large scale, we must have a way to
reduce this overly long scan time.
In addition, after we do identify the files, the large amount of data that this can represent
often drives the need for very high data rates in order to accomplish the needed amount of
Hierarchical Storage Management or backup/restore data movement, within a desired
(and continually shrinking) time window.
IBM SONAS has specific exploitation and integration with IBM Tivoli Storage Manager to
provide a solution to this problem. To address these requirements, SONAS provides specific
high performance capabilities for hierarchical storage management to external storage,
through integration of IBM Tivoli Storage Manager with the SONAS high performance scan
engine.
Specifically, Tivoli Storage Manager does not need to walk directory trees to identify files that
need to be moved to external storage, backed up, or restored. Instead, the SONAS high
performance scan engine is used to identify files to be migrated, and Tivoli Storage Manager
servers are architected to exploit multiple SONAS interface nodes for data movement in
parallel, as shown in Figure 2-23.
Migrate inactive
data to tape, tape lib,
or de-duplication device
Figure 2-23 Hiearchial Storage Management to external storage using Tivoli Storage Manager
In the HSM scenario, a stub file is left on disk. allowing the appearance of the file to be active
in the file system. Many operations, such as “list files,” are satisfied by the stub file without any
need for recall. You have flexible control over the HSM implementation, such as specifying the
size of the stub file, the minimum size of file in order to be eligible for migration, and so on.
If the file is requested to be accessed but is resident only on external storage, the file is
transparently auto-recalled from the external storage through the Tivoli Storage Manager
server. Data movement to and from external storage is done in parallel through as many
multiple SONAS interface nodes as desired, maximizing throughput through parallelism. Data
can be pre-migrated, re-staged, and de-staged, according to policy.
In this manner, SONAS provides the ability to support the storing of petabytes of data in the
online file system, yet staging only the desired portions of the file system on the actual
SONAS disk. The external storage can be any Tivoli Storage Manager-supported storage,
including external disk, tape, virtual tape libraries, or data de-duplication devices.
The same SONAS and Tivoli Storage Manager architecture is used for backup and restore
acceleration. The first step in backup / restore is to identify the files that need to be backed up
for Tivoli Storage Manager’s “incremental forever” method of operation. To identify changed
files, rules for backup are included in a SONAS policy engine scan of the file system. The high
performance scan engine locates files that need to be backed up, builds a list of these files,
and then passes the results to the Tivoli Storage Manager server, as shown in Figure 2-24.
•ProtectTier de-dup
•Virtual Tape Library
•Tape
Figure 2-24 Backup and restore acceleration using Tivoli Storage Manager
Let us now examine how the SONAS system and Tivoli Storage Manager exploitation works
in a little more detail.
The first step in either HSM or backup/restore is to identify the files that need to be migrated.
or backed up. We kick off the SONAS process to perform a central policy engine scan of the
file system, according to centrally managed rules written for the purpose of performing the
HSM or backup/restore. The high performance scan engine then passes these results to the
Tivoli Storage Manager server, as shown in Figure 2-25.
/home/appl/data/web/important_big_spreadsheet.xls
/home/appl/data/web/important_big_spreadsheet.xls
/home
/home/appl/data/web/big_architecture_drawing.ppt
/ho me/appl/data/web/big_architecture_drawing.ppt
/app l
/data /home/appl/data/web/unstructured_big_video.mpg
/ho me/appl/data/web/unstructured_big_video.mpg
/web
Policy Engine
Int er fac e
node … Inte rfa ce
node …
2. Pass list of changed
files to TSM
1. Scan to identify
changed files
St or a ge
node
… St ora ge
node
…
The results of the scan are divided up among multiple interface nodes. These multiple
interface nodes then work in parallel to with the Tivoli Storage Manager servers to initiate the
HSM or backup/restore data movement, creating parallel data streams. The Tivoli Storage
Manager software implements a “virtual node” function that allows the multiple SONAS
interface nodes to stream the data in parallel to a Tivoli Storage Manager server, as illustrated
in Figure 2-26.
/home/appl/data/web/important_big_spreadsheet.xls
/home/appl/data/web/important_big_spreadsheet.xls
/home
/home/appl/data/web/big_architecture_drawing.ppt
/ho me/appl/data/web/big_architecture_drawing.ppt
/app l
/data /home/appl/data/web/unstructured_big_video.mpg
/ho me/appl/data/web/unstructured_big_video.mpg
/web
Policy Engine
Int er fac e
node … Inte rfa ce
node …
2. Pass list of changed
files to TSM
1. Scan to identify
changed files
St or a ge
node
… St ora ge
node
…
Figure 2-26 Parallel data streams between SONAS and Tivoli Storage Manager
/app l
/home/appl/data/web/big_architecture_drawing.ppt
/data
/home/appl/data/web/unstructured_big_video.mpg
/web
TSM
IBM Scale Out NAS s er ve rs
TS M
ser v er s
5. Perform results of scan
Policy Engine
Int er fac e Inte r fac e Int er face
node Interface nodes node node
Scale
Out
Stor age nodes Stor age nodes St or a ge node s S torage node s St ora ge node s S tora ge node s
Figure 2-27 High performance parallel data movement at scale, from SONAS to external storage
SONAS scale out architecture combined with Tivoli Storage Manager can be applied to
maintain the desired time windows for automated tiered storage, hierarchical storage
management, and backup/restore, even as file systems grow into hundreds of terabytes to
petabytes.
2.5.3 Summary
In this section, we have seen the lifecycle of a file in SONAS:
Creation of files
Serving of files in a high performance manner
Automatic tiered storage management, effecting physical storage movement using central
policy engine
Migration of data to external storage for hierarchical storage management and for backup
We have also seen how SONAS provides a rich set of integrated tools to perform centralized
storage and data management at large levels of scalability. We reviewed specific exploition
with Tivoli Storage Manager to extend the scalability of SONAS scale out NAS data
management to external storage, including tape, virtual tape libraries, and data de-duplication
devices.
Existing Tivoli Storage Manager server infrastructure and skills can be used to provide these
Tivoli Storage Manager-based HSM and backup/restore acceleration capabilities to a SONAS
environment.
/app l
/home/appl/data/web/big_architecture_drawing.ppt
/data
/home/appl/data/web/unstructured_big_video.mpg
/web
Policy Engine
Int er fac e Inte r fac e Int er fac e
node Interface nodes node node
Scale
Out
Stor age nodes Stor age nodes St or a ge node s S tora ge node s St ora ge node s S torage node s
During all of these physical data movement and management operations, the user logical file
path and appearance remains untouched. The user does not have any idea that this large
scale, high performance physical data management is being automatically performed on their
behalf.
In the remainder of this book, we explore other aspects of SONAS in order to provide the
complete picture, starting with an overview of the SONAS hardware and software.
3.1.1 Nodes
The SONAS system consists of three types of server nodes:
A set of interface nodes that provide connectivity to your Internet Protocol (IP) network for
file services to external application machines running standard file access protocols such
as NFS or CIFS
A management node that provides a management interface to the SONAS configuration
Storage nodes that provide a gateway to the SONAS storage
The SONAS nodes and their roles are summarized in Figure 3-1.
Storage Nodes
• The storage node provides the connection to the InfiniBand cluster
interconnect and direct Fibre Channel attachment to the IBM Scale Out NAS
RAID controller.
• Storage nodes are configured in high-availability pairs, connected to one or
two IBM Scale Out NAS RAID controllers.
• Four PCIe adapters attach to the internal network and the Scale Out NAS
storage controller.
Management Node
• The management node provides the user interface for configuring,
administering and monitoring.
• There are two 1 GbE ports for health monitoring and configuration. There
are also two 1 GbE ports that connect to the customer network for the11/14/20 10
SONAS Management GUI and Command Line Interface (CLI) access.
The interface nodes, management nodes, and the storage nodes are connected through a
scalable, redundant infiniband fabric allowing data to be transferred between the interface
nodes providing access to the application and the storage nodes with direct attachments to
the storage. Infiniband was chosen for its low overhead and high speed - 20 Gbits/sec for
each port on the switches. The basic SONAS hardware structure is shown in Figure 3-2.
– Processor:
• Dual Quad Core Intel® Xeon® X5530
• 2.26 G Hz, 8 MB L2 cache, 80W
– Memory:
• 32 GB, 64 GB, or 128 G B RAM
– Network Interfaces:
• Four 1 Gbps NICs
- 2 Customer Network
- 2 Management Network
• Optional additional card with four more 1 GbE ports or
Optional 10 GbE NIC
- (Mutually exclusive)
Each SONAS interface node contains two redundant hot-swappable 300 GB 2.5-inch 10K
RPM SAS hard disk drives (HDDs) with mirroring between the two HDDs for high-availability.
The HDDs contain the SONAS system software product, containing the operating system and
all other software needed for an operational SONAS system.
The interface nodes can operate at up to 10 Gbit speeds with the optional 10 GbE Ethernet
adapters, providing extremely fast access to data. All nodes in the SONAS system are in a
global parallel cluster, and are interconnected and cross-communicate by a redundant high
speed InfiniBand data network.
Users can access files in the SONAS file system through each of the SONAS interface nodes.
In conjunction with an external Domain Name Server, incoming workload is IP-address
balanced across all interface nodes. If more interface node capacity is needed, you simply
add more interface nodes, which provides a highly scalable capability. Additional data access
performance can be obtained by adding interface nodes up to the maximum limits of SONAS.
At the time of writing of this book, SONAS supports a minimum of two interface nodes and a
maximum of 30 interface nodes. All interface nodes can access all files and all file systems in
the SONAS system (with proper authentication, of course). The interface nodes request
access to the SONAS file systems, which are resident on file systems served by the storage
nodes, which are gateways to SONAS integrated storage controllers and disks drawers. All
interface nodes can access all storage on all storage nodes. All storage nodes can send data
to any interface node. In 3.2, “Software architecture” on page 58, we see how the SONAS
Software exploits this parallel global clustering capability.
The specifications of the SONAS storage node are summarized in Figure 3-4.
Storage nodes are incorporated into “building blocks’” of physical SONAS disk storage called
storage pods. A storage pod consists of two storage nodes, which are configured as a
high-availability (HA) pair. The storage nodes are connected to one or two SONAS storage
controllers (2851-DR1). A SONAS storage controller can optionally attach a SONAS Disk
Expansion Unit (2851-DE1).
• Ultra-Dense
• 4U high 2851-DR1 storage controller includes 60
drives
• 4U high 2851-DE disk expansion Unit provides
additional 60 drives
• Flexible
• Intermix SAS and SATA
• A 60-drive drawer must be all SAS or all SATA
• Highly Reliable
• Active/Active Storage Controller Failover
• Data Integrity Validation
• RAID 5 (SAS) or RAID 6 (SATA)
two SONAS 2851-DR1 two SONAS 2851-DE1 • Redundancy Throughout
storage controllers disk expansion units • Battery Backed Cache
Figure 3-5 SONAS storage pod with two storage nodes, fully populated with 240 disk drives
The storage node contains two redundant hot-swappable 300 GB 2.5-inch 10K RPM SAS
HDDs with mirroring between them for high-availability. The hard disk drives contain the
SONAS System Software product (which hosts the SONAS operating system and all other
software needed for an operational SONAS system.
The global clustered architecture of SONAS means that all SONAS interface nodes can
access all storage on all storage nodes. All storage nodes can send data to any interface
node. A SONAS system contains a minimum of two storage nodes. The maximum number of
storage nodes depends on the size of the InfiniBand switches that are ordered - when using
96-port InfiniBand switches in the SONAS base rack, a maximum of 60 storage nodes is
possible. When using 36-port InfiniBand switches in the base rack, the maximum number of
storage nodes is 28.
The two onboard Ethernet ports in the storage node connect to the internal private SONAS
1 GbE management network, and two onboard Ethernet provide a NULL Ethernet connection
to the SONAS Disk Storage Controllers.
Management node
The Management node is a 2U server that provides a central point for the system
administrator to configure, monitor, and manage the operation of the SONAS cluster. The
management node supports both a browser-based graphical user interface (GUI) and a
command line interface (CLI). It also provides a System Health Center for monitoring the
overall health of the system. A single management node is required. The SONAS system will
continue to operate without the SONAS management node being active, but configuration
changes can only be performed from an active management node. The SONAS management
node specifications are shown in Figure 3-6.
The management node contains two hot-swappable 300 GB 2.5-inch 10K RPM SAS hard
disk drives with mirroring between the two HDDs for high-availability. The hard disk drives
contain the SONAS System Software product, containing the operating system and all other
software needed for an operational SONAS system. The third hot-swappable 300 GB 2.5-inch
10K RPM SAS hard disk drive stores the logging and trace information for the entire SONAS
system.
The management node comes with all of the cables that are required to connect it to the
switches within the base rack. The management node is assumed to be in the SONAS base
rack with the two InfiniBand switches.
3.1.3 Switches
The SONAS system contains internal InfiniBand, internal Ethernet switches and external
customer supplied external Ethernet switches.
Two redundant InfiniBand switches are included inside each SONAS system. These two
identical IB switches are configured when you order a SONAS system;- the two InfiniBand
switches are used in order to provide redundancy for each other. For small and medium
configurations, a 1U 36 port 4X DDR InfiniBand switch is available. For larger configurations,
a 7U 96-port 4X DDR InfiniBand switch is available.
Unfortunately, it is not possible to field upgrade the 36-port InfiniBand switch to the 96-port
switch. Therefore, at initial SONAS order time, consider your growth requirements, and order
the appropriate InfiniBand switch.
The SONAS InfiniBand network supports high bandwidth, low latency file system data and
control traffic among the nodes of the SONAS system. The data carried by the InfiniBand
fabric includes low level file system data, as well as the TCP/IP-based locking and
management messaging traffic for the parallel global cluster. The locking and management
traffic occurs on IP over InfiniBand.
Cabling within a SONAS rack is provided as part of the rack configuration and it is performed
by IBM at the plant of manufacture. You must order InfiniBand cable features for inter-rack
cabling after determining the layout of your multi-rack system.
The management network is not used for user data or file system data transfer, rather, it only
carries low bandwidth internal messages in support of managing, monitoring, and performing
configuration, health, and monitoring of the SONAS subsystem. All major components of a
SONAS system such as interface nodes, storage nodes, management nodes, and infiniband
switches are connected to internal Ethernet management network.
You can also connect the SONAS management network switches to your external network
However, these connections are only for your SONAS and IT management staff to be able to
securely access the SONAS management node for management GUI and Command Line
Interface. Your external user network infrastructure must support either or both 1 Gb/s or
10 Gb/s Ethernet links, according to the way that you have configured the 1 GbE or 10 GbE
ports on your SONAS interface nodes.
The external customer network and network switches and the cables attaching the SONAS
system to your customer network are not provided as part of the SONAS appliance.
Each interface node is capable of serving data to all users, you can control which users reach
which interface nodes, by the appropriate Domain Name Server definitions.
The manner in which this is done is described in “Software architecture” on page 58. Each
interface node is a balanced modular network building block, with sufficient main memory and
PCI bus capacity to provide full throughput with the Ethernet adapters configured to that
node.
Each interface node can be configured with either of the following possibilities:
One pair of active/passive 10 GbE Ethernet ports or
Two or six 1 GbE Ethernet ports, all of which can simultaneously be active.
You can attach one SONAS disk expansion unit (2851-DE1) to the first SONAS storage
controller. A storage pod can be expanded by additional second SONAS storage controller.
To that second storage controller, you can add one additional disk expansion unit. Each
storage controller has 60 drives, and each storage expansion unit has 60 drives. The 60
drives for any individual storage controller or storage expansion unit must all be the same
drive type. You can choose any drive type that you want for this group of 60 drives, as long as
they are all the same.
The storage nodes within each storage pod provides dual paths to all storage controllers, and
each storage controller or storage expansion unit has dual paths to all disks for reliability.
Each storage node and storage controller operate in parallel with each other. Multiple storage
pods operate in parallel with each other.
Note that there is no fixed relationship between the number of interface nodes and the
number of storage pods. Storage pods can be scaled to provide an ability to increase storage
capacity and bandwidth, independently of the interface nodes. Storage capacity can be
added to the SONAS system by either adding disks to existing storage pods, or adding
storage pods. Each storage pod supports a maximum of 240 hard disk drives.
The storage controller is configured to use RAID 5 with 450 GB or 600 GB SAS hard-disk
drives, and RAID 6 with 1 or 2 terabyte (TB) SATA hard-disk drives. All 60 disk drives in the
storage controller must be the same type, either six 10-packs of 7.2K RPM 1 TB SATA
hard-disk drives or six 10-packs of 2 TB SATA drives, or six 10-packs of 15K RPM 450 GB or
six 10-packs of 15K RPM 600 GB SAS drives. You cannot mix drive types or sizes within an
enclosure. The SONAS Storage Controller and attached disk expansion unit can contain
unique disk types. Each SONAS Storage Controller can attach one high-density SONAS disk
expansion unit (2851-DE1).
The purpose of the component connection determines the type of connection that is used.
There are four connection types that are used in SONAS:
External customer network:
The only customer supplied connection is the connection between the SONAS interface
nodes to the external customer network.
Customer IP Network
SONAS System Tap e
connects
Storage Node Storage Node
all SONAS
nodes for
internal
data transfer:
High Density High Density
RAID Controller RAID Controller Interface,
Fibre Storage,
Channel Management
connects High Density High Density
storage Disk Enclosure Disk Enclosure
nodes to
storage
Storage Pod
From Figure 3-7, you can see that the external network connections to the SONAS system
are provided by the customer. The internal SONAS connections can be summarized as
follows:
Each of the internal connections is fully duplexed and redundant. Specifically, each
internal SONAS component has dual connections, one connection each to pair of
switches. This provides provide high availability and automatic failover, in the event of
failure of any internal component connection.
For internal data transfer between all nodes in the SONAS cluster, InfiniBand is used. All
data transfer takes place over Infiniband cable connections to a pair of internal Infiniband
switches, taking advantage of Infiniband's very low latency and high 20 Gbits/second/port
data transfer rate.
Storage is Fibre Channel attached, directly to the Storage nodes. There is no internal
SAN, and there are no internal SAN switches inside the SONAS system.
Finally, all SONAS components are connected to the Management node by 1 Gbit/sec
Ethernet. Over this connection, the Management node monitors the health of the entire
cluster and all components, and manages the cluster. Because the nature of this
connection is commands and health information, 1 GbE is more than adequate.
SONAS features a very flexible, modular architectural approach, as discussed in the following
topics
Base rack
The SONAS system always includes a base rack that contains the management node, the
InfiniBand switches, a minimum of two interface nodes, and a keyboard, video, and mouse
(KVM) unit. The capacity of the SONAS system that you order affects the number of racks in
your system and the configuration of the base rack.
With the feature code 9003 rack, you are configuring the 36 Infiniband port switches. Note
that the InfiniBand switches cannot be expanded or exchanged. The feature code 9003
SONAS based rack can have up to 14 interface nodes. You can attach one additional
interface expansion rack and as many storage expansion racks as you see fit, up to the
maximum attachability of the 36 port InfiniBand switches.
The purpose of the feature code 9004 rack is to allow you to scale to maximum scalability,
because it has the large 96-port InfiniBand switches. You can attach one additional interface
expansion rack and as many storage expansion racks as you see fit, up to the maximum
attachability of the 96 port InfiniBand switches.
The purpose of the feature code 9005 rack is to support a combination of the interface nodes
and storage nodes within the base rack. This rack can have up to 6 interface nodes and one
storage pod. A fully populated storage pod in this base rack has 240 disk drives (configured in
groups of 60 as previously discussed). You can attach one additional interface expansion rack
and as many storage expansion racks as you see fit, up to the maximum attachability of the
36 port InfiniBand switches.
There is a SAS verses SATA configuration consideration when configuring SONAS storage
expansion racks.
Note that SAS drives consume more power than SATA drives. The power supplies in the
SONAS storage expansion rack can support a maximum of the following drives:
960 SATA drives (in groups of 60)
720 SAS drives, because that amount can consume the power supply capability of the
storage expansion rack
An enclosure holds six 10-packs (60 drives), and all drives within an enclosure must be of the
same type (you cannot mix drive types or sizes within an enclosure). The SONAS storage
controller uses RAID 5 with 450 GB and 600 GB SAS hard-disk drives, and RAID 6 with 1 TB
and 2 TB SATA hard-disk drives. Table 3-1 shows a summary of possible configurations.,
which are preconfigured and cannot be changed.
SATA 1 TB or RAID 6 60 48 12 0
2 TB
SONAS supports drives intermix at the level of the enclosure (a group of 60 drives). Each of
the four possible enclosures within a storage pod might be another drive type. Thus, it is
possible to have within the same storage pod, or attached to the same, a storage controller,
an enclosure with high performance SAS disks and another storage enclosure with high
capacity SATA disks.
As we can see in 3.2, “Software architecture” on page 58, SONAS Software provides
automated tiered storage, so that Hierarchical Storage Management can be performed on
physical data and storage between the various types of physical drives. Using this capability,
data can be automatically migrated according to centrally managed and deployed storage
policies within SONAS from faster, but smaller and more expensive SAS disks, to slower, but
larger and less expensive SATA disks, and even migrated out to external storage.
After you have decided on which base SONAS rack you will start with, you can then expand
the SONAS system independently in either interface node expansion rack or storage node
expansion racks as shown in Figure 3-11.
Using these expansion racks, you can build a SONAS system to the level of scalability that
you want. Figure 3-12 shows the mid-point of the SONAS expansion capability. The example
SONAS system has 3360 disks, out of the 7200 maximum possible.
Example SONAS with 30 interface nodes, 7 disk expansion racks, 3360 disks
Sw ti ch es
S w ti ch es S w ti c hes S w i t che s S wic
t he s S w i t che s S wic
t he s S w i t che s S w i t che s SwSw
i t iche
t chses Sw i t che s Sw i t che s Sw i t ch es Sw i t che s Sw i t ch es Sw i t ch es S w ti c hes
In e
t r fa ce N od e S t orag e N ode St o a
r ge N od e S t or a ge N od e St o a
r ge N od e S t or a ge N od e St o a
r ge N od e St o a
r ge N od e S tS
or to
a ge
r agNeod
Noede S to r age No de St o r age N o de S to rag e No de St o r age N o de S to r ag e No de S to rage No de S t or a ge N ode
S t orag e N ode St o a
r ge N od e S t or a ge N od e St o a
r ge N od e S t or a ge N od e St o a
r ge N od e St o a
r ge N od e S tS
or to
a ge
r agNeod
Noede S to r age No de St o r age N o de S to rag e No de St o r age N o de S to r ag e No de S to rage No de S t or a ge N ode
In e
t r fa ce N od e M g m t No de
In e
t r fa ce N od e
60 D i sks 60 D i sks 60 D i sks 60 Di sk s 60 D i sks 60 Di sk s 60 Di sk s 6 0
6 0
Di D
skissks 6 0 Di sk s 6 0 Di sk s 60 D i sks 6 0 Di sk s 6 0 D isk s 6 0 D isk s 60 D i sks
In e
t r fa ce N od e I n if n b
i an d sw i c
t h
In e
t r fa ce N od e
60 D i sks 60 D i sks 60 D i sks 60 Di sk s 60 D i sks 60 Di sk s 60 Di sk s 6 0
6 0
Di D
skissks 6 0 Di sk s 6 0 Di sk s 60 D i sks 6 0 Di sk s 6 0 D isk s 6 0 D isk s 60 D i sks
In e
t r fa ce N od e
I n if n b
i an d sw i c
t h
In e
t r fa ce N od e
60 D i sks 60 D i sks 60 D i sks 60 Di sk s 60 D i sks 60 Di sk s 60 Di sk s 6 0
6 0
Di D
skissks 6 0 Di sk s 6 0 Di sk s 60 D i sks 6 0 Di sk s 6 0 D isk s 6 0 D isk s 60 D i sks
In e
t r fa ce N od e
In e
t r fa ce N od e
60 D i sks 60 D i sks 60 D i sks 60 Di sk s 60 D i sks 60 Di sk s 60 Di sk s 6 0
6 0
Di D
skissks 6 0 Di sk s 6 0 Di sk s 60 D i sks 6 0 Di sk s 6 0 D isk s 6 0 D isk s 60 D i sks
In e
t r fa ce N od e M o ni t or / K VM
In e
t r fa ce N od e In e
t r fa ce N od e So
t r ag e No de S t ora ge N ode S t ora ge N ode S t or a ge N od e S t ora ge N ode S t or a ge N od e S t or a ge N ode S t St
or a
o ge
a
r ge
N od
N od
e e St o a
r ge N od e St o a
r ge N od e S to r age No de St o a
r ge N od e St o a
r ge N od e St o a
r ge N od e S to rag e No de
So
t r ag e No de S t ora ge N ode S t ora ge N ode S t or a ge N od e S t ora ge N ode S t or a ge N od e S t or a ge N ode S t St
or a
o ge
a
r ge
N od
N od
e e St o a
r ge N od e St o a
r ge N od e S to r age No de St o a
r ge N od e St o a
r ge N od e St o a
r ge N od e S to rag e No de
In e
t r fa ce N od e In e
t r fa ce N od e
In e
t r fa ce N od e
In e
t r fa ce N od e
60 D i sks 60 D i sks 60 D i sks 60 Di sk s 60 D i sks 60 Di sk s 60 Di sk s 6 0
6 0
Di D
skissks 6 0 Di sk s 6 0 Di sk s 60 D i sks 6 0 Di sk s 6 0 D isk s 6 0 D isk s 60 D i sks
In e
t r fa ce N od e
In e
t r fa ce N od e
In e
t r fa ce N od e In e
t r fa ce N od e
60 D i sks 60 D i sks 60 D i sks 60 Di sk s 60 D i sks 60 Di sk s 60 Di sk s 6 0
6 0
Di D
skissks 6 0 Di sk s 6 0 Di sk s 60 D i sks 6 0 Di sk s 6 0 D isk s 6 0 D isk s 60 D i sks
In e
t r fa ce N od e In e
t r fa ce N od e
In e
t r fa ce N od e In e
t r fa ce N od e
In e
t r fa ce N od e
In e
t r fa ce N od e
60 D i sks 60 D i sks 60 D i sks 60 Di sk s 60 D i sks 60 Di sk s 60 Di sk s 6 0
6 0
Di D
skissks 6 0 Di sk s 6 0 Di sk s 60 D i sks 6 0 Di sk s 6 0 D isk s 6 0 D isk s 60 D i sks
In e
t r fa ce N od e In e
t r fa ce N od e
Maximum SONAS with 30 interface nodes, 15 disk expansion racks, 7200 disks
Figure 3-13 SONAS maximum configuration
As we can see, the SONAS hardware architecture provides the foundation for petabyte level
scale out storage.
The functionality of IBM SONAS is provided by the IBM SONAS Software licensed program
product (5639-SN1). Each node of the IBM SONAS system is licensed and pre-installed with
one copy of SONAS Software. According to the role of the node (interface node, storage
node, or management node), the appropriate functions are called upon out of the common
software code load. Each node and each copy of SONAS Software together in a parallel grid
cluster architecture, working in parallel to provide the functions of SONAS.
Enterprise Linux
IBM Servers
Network
Storage
Users
Ethernet
Enterprise Linux
IBM Servers
The file access protocols that are supported by SONAS today are CIFS, NFS, FTP, and
HTTPS. Following this section, we then discuss the role the SONAS Cluster Manager plays in
concurrent access to a file from multiple platforms (that is, concurrently access a file from
both CIFS and NFS, for example).
The base SONAS file system is a POSIX-compliant UNIX/Linux-style file system. SONAS
communicates with Microsoft Windows clients by emulating Windows file system behavior
over this POSIX-compliant SONAS file system. The SONAS Cluster Manager (see “SONAS
Cluster Manager” on page 63) provides the mapping of the CIFS semantics onto the
UNIX/Linux-based SONAS file system. The SONAS Cluster Manager also maps UNIX/Linux
Access Control Lists (ACLs) to Windows security semantics.
CIFS locks set by Windows clients are stored both in the SONAS Cluster Manager interface
node cluster-wide database, and by mapping to POSIX locks in the SONAS file system. This
mapping ensures that NFS clients see relevant CIFS locks as POSIX advisory locks, and
NFS clients honor these locks.
Note that the SONAS Software file system implements NFSv4 Access Control Lists (ACLs)
for security, regardless of the actual network storage protocol used. This provides the
strength of the NFSv4 ACLs even to clients that access SONAS by the NFSv2, NFSv3, CIFS,
FTP, and HTTPS protocols.
The Apache daemon provides HTTPS access to the SONAS file system. SONAS supports
secure access only, so that the credentials will always be SSL encrypted. SONAS uses HTTP
aliases as a vehicle to emulate the share concept. For example, share XYZ will be accessible
by https://server.domain/XYZ.
Note that WebDAV1 and the REST2 API are not supported at this time in SONAS; they are
known requirements.
Enterprise Linux
IBM Servers
1
WebDAV is “Web-based Distributed Authoring and Versioning”
2 REST API is “Representational State Transfer”
The SONAS Cluster Manager also provides advanced functions such as byte-range capable
locking available in the SONAS parallel file system; managing the interface nodes and
multiple protocols to work in conjunction with the SONAS parallel file system base technology
for concurrent access; and parallel read and write access, for multiple protocols and multiple
platforms, across multiple SONAS interface nodes. It is able to do all of these tasks with full
data integrity to all files, anywhere within the file system.
HTTP
HTTP NFS
NFS CIFS
CIFS FTP
FTP Other
Other
Clients
Clients Clients
Clients Clients
Clients Clients
Clients Clients
Clients
IP Network
Global Namespace
Management
Management
Node
Node
Interface
Interface
Node
Node ... Interface
Interface
Node
Node .... Interface
Interface
Node
Node
Storage
controller &
disk
Storage
controller &
disk
... Storage
controller &
disk
Storage
controller &
disk
Node types
There are three types of nodes in a SONAS system. The nodes are divided and configured
according to one of three roles. All nodes are in a global cluster, and a copy of SONAS
Software runs on each of the nodes. A node performs only one of the following three roles:
Interface node: This node provides the connections to the customer IP network for file
serving. These nodes establish and maintain the connections to CIFS, NFS, FTP, or HTTP
users, and serve the file requests. All four of these protocols can and do co-exist on the
same interface node. Each interface node can talk to any of the storage nodes.
Storage node: This node acts as a storage servers, and reads and writes data to and from
the actual storage controllers and disks. Each storage node can talk to any of the interface
nodes. A storage node serves file and data requests from any requesting interface node.
SONAS Software writes data in a wide stripe across multiple disk drives in a logical
storage pool. If the logical storage pool is spans multiple storage nodes and storage pods,
the data striping will also span storage nodes and storage pods.
Management node: This node monitors and manages the SONAS global cluster of nodes,
and provides Command Line Interface management and GUI interface for administration.
Command Line Interface commands come into the SONAS system through the
Management node.
Notice that SONAS is a two-level architecture, there are multiple clustered interface nodes,
and there are multiple clustered storage nodes. This is an important aspect of the design, as
it allows independent scalability of interface nodes (user throughput) from storage pods and
storage nodes (storage capacity).
SONAS Software is designed with the understanding that over time, various generations and
speeds of System x servers will be used in the global SONAS cluster. SONAS Software
understands this and is able to distribute workload equitably, among various speed interface
nodes and storage nodes within the cluster.
Clusters usually cannot outperform a stand-alone server to a single client, due to cluster
overhead. At the same time, clusters can outperform stand-alone servers in aggregate
throughput to many clients, and clusters can provide superior high availability.
SONAS is a hybrid design that provides the best of both of these approaches.
SONAS.virtual.com SONAS.virtual.com
DNS Server
(name resolution)
SONAS.virtual.com
10.0.0.10
10.0.0.11
10.0.0.10 10.0.0.12 10.0.0.14 10.0.0.12
10.0.0.11 10.0.0.13 10.0.0.15 10.0.0.13
10.0.0.14
10.0.0.15
SONAS allocates a single user network client to a single interface node, to minimize cluster
overhead. SONAS Software does not rotate a single client’s workload across interface nodes.
That is not only unsupported by DNS or CIFS, but can also decrease performance, because
caching and read-ahead is done in the SONAS interface node, and it is for this reason that
any one individual client is going to be assigned, for the duration of their session, to one
interface node at the time they authenticate and access the SONAS.
At the same time, the workload from multiple users (numbering into the thousands or more)
is equitably spread across as many SONAS interface nodes as are available. If more user
network capacity is required, you simply add more interface nodes. SONAS scale out
architecture thus provides linear scalability as the numbers of users grow.
Agnostically to the application or the interface nodes, SONAS Software will always stripe data
across disks, storage RAID controllers, and storage pods, thus providing wide data striping
performance and parallelism to any file serving requests, by any interface node. This is
shown in Figure 3-19.
Parallelism
for high
performance
agnostic to
application
Figure 3-19 SONAS interface node workload allocation - parallelism at storage level
SONAS provides a single high performance NFS, CIFS, FTP, or HTTPS connection for any
one individual network clients. In aggregate, multiple users are IP-balanced equitably across
all the interface nodes, thus providing scale out capability; the more interface nodes, the more
user capacity that is available.
SONAS was designed to make the connection a standard CIFS, NFS, FTP, or HTTP
connection, in order to allow attachability by as wide a range of standard clients as possible,
and to make unnecessary the installation of any client side code.
When an interface node is removed from the cluster, or if there is an interface node failure,
healthy interface nodes take over the load of the failed node, and the SONAS Software
Cluster Manager will automatically perform the following actions:
Terminate old network connections and move the network connections to a healthy
interface node. IP addresses are automatically re-allocated to a healthy interface node.
– Session and state information that was kept in the Cluster Manager is used to support
re-establishment of the session and maintaining IP addresses and ports.
SONAS.virtual.com SONAS.virtual.com
DNS Server
(name resolution)
SONAS.virtual.com
10.0.0.10
10.0.0.13 10.0.0.12 10.0.0.11
10.0.0.10 10.0.0.14 10.0.0.12
10.0.0.11 10.0.0.15 10.0.0.13
10.0.0.14
10.0.0.15
At the time of the failover of the node, if the session or application is not actively in a
connection transferring data, the failover can usually be transparent to the client. If the client
is transferring data, depending on the protocol and application, the application service failover
might be transparent to the client, depending on nature of the application, and depending on
what is occurring at the time of the failover.
In particular, if the client application, in response to the SONAS failover and SONAS
notifications, automatically does a retry of the network connection, then it is possible that the
user will not see an interruption of service. Examples of software that does this can include
many NFS-based applications, as well as Windows applications that do retries of the network
connection, such as the Windows XCOPY utility.
If the application does not do automatic network connection retries, or the protocol in question
is stateful (for example, CIFS), then a client side reconnection might be necessary to
re-establish the session. Unfortunately for most CIFS connections, this will be the likely case.
If a storage node fails, the remaining healthy storage node in the storage pod takes over the
load of the failed storage node. An individual storage node is very high in capacity and
throughput, to allow good operation in the event of a failed storage node.
Logical storage pools in SONAS typically are defined as spanning disks, storage RAID
controllers, and storage pods. Furthermore, the data striping means that files are spread in
blocksize “chunks” across these components in order to achieve parallel performance and
equalized utilization of storage hardware.
One of the purposes of this dispersion of SONAS data is to mitigate the effect of a failed
storage node. Files in SONAS are spread across multiple storage nodes and storage pods,
with the intent that only a small portion of any file is affected by a storage node failure.
Just as with interface nodes, storage nodes can be dynamically removed and re-inserted into
a cluster. Similar to the interface node methodology, which is the same, the method of
upgrade or repair of a storage node is to take the storage node out of the cluster. The
remaining storage node in the storage pod dynamically assumes the workload of the pod.
The offline storage node can then be upgraded or repaired, and then re-inserted into the
cluster. When this is done, workload will then be automatically rebalanced across the storage
nodes in the storage pod.
During all of these actions, the file system and file access stays online and available to the
users.
3.4.5 Summary
We have seen that the IBM SONAS provides equitable workload allocation to a global cluster
of interface nodes, including high availability through clustered auto-failover:
All SONAS nodes operate in a global cluster.
Workload allocation to the interface nodes is done in conjunction with external Domain
Name Servers.
The global SONAS cluster offers dynamic failover/failback, and if the application supports
network connection retries, can provide transparent failover of the interface nodes.
Normal upgrade and maintenance for SONAS nodes is by dynamic removal and insertion
of nodes into and out of the cluster.
Let us now discuss in more detail the SONAS Software components that provide the
functionality to support these capabilities.
SONAS Cluster
Manager
concurrent_access_to_a_file
Figure 3-21 All file accesses traverse the SONAS Cluster Manager including concurrent accesses
The SONAS Cluster Manager is logically positioned in the file access path-length to provide
the mapping of the multiple protocols onto the SONAS parallel file system, and to manage the
necessary locking and data integrity across all interface nodes. The SONAS Cluster Manager
works together with the SONAS parallel file system to provide concurrent file access from
multiple platforms in the following way:
SONAS Cluster Manager: Provides the mapping and concurrent access control across
multiple interface nodes, and when multiple protocols access the same file, and provides
locking across users across all interface nodes.
SONAS parallel file system: Provides the file system concurrent access control at the level
of the physical file management, provides ability to manage and perform parallel access,
provides the NFSv4 access control list security, and provides the foundational file system
data integrity capabilities.
SONAS Software Cluster Manager functionality supports multiple exports or shares of the file
system, over multiple interface nodes, by providing distributed lock, share, and lease support.
An individual interface node can simultaneously be serving CIFS, NFS, FTP, and HTTPS;
the multiple file access protocols are not limited or exclusive to a particular interface node.
The SONAS Cluster Manager is transparent to the NFS, CIFS, FTP, and HTTP clients; these
clients do not need to know that the SONAS Cluster Manager is servicing and managing
multiple protocols concurrently.
The base SONAS file system is a full POSIX-compliant file system, a UNIX/Linux-style file
system. SONAS communicates with Microsoft Windows clients by emulating Windows file
system behavior over this POSIX-compliant SONAS file system.
01:00
12
9 3
6
01:01
12
9 3
6 Other
users see
files are
deleted
Figure 3-22 SONAS concurrent access to a shared directory from multiple users
Logical
/app l
Unix-Linux/home/appl/data/web/writing_reading_the_file.dat
/data
Any/home/appl/data/web/writing_reading_the_file.dat
/web
CIFS NFS
IBM Scale Out NAS SONAS file system
provides ability multiple
concurrent readers/writers
Global Namespace from multiple platforms
Policy Engine
Inter fa ce
node s … Int er fac e
nodes … Inter face
nodes …> scale
out
Physica l
Stor age
node s
….. Stor age
node s
….. S tora ge
nodes
... > scale
out
Figure 3-23 SONAS Software and file system provides advanced concurrent access and locking
SONAS Software has multiple facilities to provides scalability. These include the distributed
ability for multiple nodes to act as token managers for a single file system. SONAS Software
also provides scalable metadata management by providing for a distributed metadata
management architecture, thus allowing all nodes of the cluster to dynamically share in
performing file metadata operations while accessing the file system.
This capability distinguishes SONAS from other cluster NAS filer architectures, which might
have a centralized metadata server handling fixed regions of the file namespace. A
centralized metadata server can often become a performance bottleneck for metadata
intensive operations and can represent a scalability limitation and single point of failure.
SONAS solves this problem by managing metadata at the node that is using the file, or in the
case of parallel access to the file, at a dynamically selected node which is using the file.
The SONAS Cluster Manager provides these capabilities, and provides full active/active
interface node clustering, with the appearance of a single server global namespace instance
spanning all SONAS interface nodes. This clustering occurs transparently to all applications.
The authentication server is external to SONAS and must have a proper connectivity to
SONAS. The authentication server is configured externally to the SONAS, using normal
authentication server skills.
All of the SONAS nodes must have access to a network time protocol server that is
synchronized in time to be the same as an authentication server such as Active Directory
Server and/or Kerberos KDC server. An Active Directory domain controller can be used as a
NTP time source.
Only the SONAS interface nodes are configured for authentication by users. SONAS storage
nodes are not part of this configuration. In the current SONAS release, the number of groups
per user is limited to approximately 1000.
Note that the external authentication server is an essential component in the SONAS data
access flow. If the authentication server is unavailable, then the data access is not available.
To make both worlds work together in SONAS and provide full concurrent access from both
platforms, a mapping between Windows Security Identifiers (SIDs) and UNIX userids /
groupids (UID/GID) is performed.
When accessing SONAS data using UNIX or Linux systems (that is, through NFS), there are
no issues because the user UID/GID directly maps to the underlying SONAS file system UID
and GID.
However, when Windows clients access SONAS, the SONAS Cluster Manager provides the
mapping between the Windows user SID (Security identifier) and the internal file system UID
to identify users. Depending on the type of authentication used, various methods are applied
to solve this UID to SID mapping requirement.
Windows AD
uses Linux
User- / Groupname
SONAS file System (GPFS
UID / GID
NFS Shared
Id map db
UID / GID
In the SONAS Software, the parallel file system, which includes the central policy engine and
the high performance scan engine, are at the heart of SONAS Software functionality
(Figure 3-25).
Enterprise Linux
IBM Servers
Figure 3-25 SONAS Software - parallel file system, policy engine, scan engine
The SONAS file system is more than just a file system; it is the core foundation for an end to
end NAS file management infrastructure within SONAS. IBM utilizes IBM GPFS technology to
provide a proven high performance parallel grid file system architecture, with clustered high
reliability, high scalability.
In addition to providing file storage capabilities, the SONAS file system also provides storage
management and information life cycle management tools, centralized administration and
facilities that, in conjunction with the SONAS Cluster Manager, allows for shared high
performance access from multiple NAS protocols simultaneously.
IBM SONAS was designed to take advantage of the long history of IBM GPFS as a high
performance parallel file system, supporting many types of applications ranging from
relational databases, to digital media, to high performance analytics, to scalable file serving.
The core GPFS technology is installed today in across many industries including financial,
retail, and government applications. GPFS has been tested in very demanding large
environments for over 15 years, making GPFS a solid foundation for use within the SONAS
as the central parallel file system.
SONAS utilizes the fact that GPFS was designed from the beginning to support extremely
large, extremely challenging high performance computing environments. Today, SONAS uses
that technology to supports building a single global namespace and a single file system over
the entire 14.4 PB current maximum size of a physical SONAS system.
The theoretical limit to the SONAS file system size is (2 to the 99th power) bytes. The
maximum SONAS file size is (2 to the 64th power) bytes. The current maximum SONAS
number of files per file system is (2 to the 31st power) -1 (approximately 2 billion). You can
have up to 256 file systems in a SONAS system. Architecturally, a SONAS file system or a
single SONAS file can be as large as the entire SONAS system.
On the interface nodes, the SONAS file system code serves as “file system storage
requesters.” On the storage nodes, the SONAS file system code serves as “file system
storage servers.” All SONAS nodes (interface, management, and storage) are a peer-to-peer
global cluster.
There is no limit placed upon the number of simultaneously opened files within a single file
system. SONAS utilizes the experience from worldwide current GPFS customers who are
using single file systems from 10 PB to 20 PB in size, and growing. Other GPFS user file
systems contain hundreds of millions of files.
Let us see how SONAS scalability is achieved using the expandability of the SONAS building
block approach.
Figure 3-26 shows a SONAS interface node, performing a small single read or write to or from
a disk. Because this read or write is small in size, only the resources of one path, one storage
node, and one RAID controller / disk are sufficient to handle the I/O operation.
InfiniBand network
Storage Pod
Storage
Storage Node
Node Storage
Storage Node
Node
(NSD
(NSD Server)
Server) (NSD
(NSD Server)
Server)
RAID RAID
Controller Controller
Raid
disk
Figure 3-26 A small single read or write in the SONAS file system
The power of the SONAS file system, however, is in its ability to read or write files in parallel
(in blocksize “chunks”), across multiple disks, controllers, and storage nodes (Figure 3-27).
Interface
Interface Node
Node
InfiniBand network
Storage Pod
Storage
Storage Node
Node Storage
Storage Node
Node
(NSD
(NSD Server)
Server) (NSD
(NSD Server)
Server)
Raid Raid Raid Raid Raid Raid Raid Raid Raid Raid Raid Raid
(NSD) (NSD) (NSD) (NSD) (NSD) (NSD) (NSD) (NSD) (NSD) (NSD) (NSD) (NSD)
Figure 3-27 A high parallel read or write in the SONAS file system; can be one file
Interface
Interface Node
Node
InfiniBand network
Storage
Storage Storage
Storage
Node
Node Node
Node
High Density
Storage Array
High Density
Storage Array
Storage Pod
Figure 3-28 SONAS file system parallel read / write capability to one storage pod
If the file is big enough, or if the aggregate workload is big enough, the SONAS file system
easily expands to multiple storage pods in parallel (Figure 3-29).
Interface
Interface Node
Node
InfiniBand network
Storage
Storage Storage
Storage Storage
Storage Storage
Storage Storage
Storage Storage
Storage Storage
Storage Storage
Storage
Node
Node Node
Node Node
Node Node
Node Node
Node Node
Node Node
Node Node
Node
Figure 3-29 SONAS file system parallel read/ write capability to multiple storage pods
The SONAS file system recognizes typical access patterns such as sequential, reverse
sequential, and random, and optimizes I/O access for these patterns.
The SONAS file system has implemented a sophisticated distributed metadata server
function, in which multiple nodes act, share, acquire, and relinquish roles as token managers
for a single file system. This distributed architecture avoids metadata server bottlenecks, and
has been proven to scale to very large file systems.
Along with distributed token management, SONAS file system provides scalable metadata
management by allowing all nodes of the cluster accessing the file system to perform file
metadata operations. This key and unique feature distinguishes SONAS from other cluster
file systems which have a centralized metadata server handling fixed regions of the file
namespace. SONAS file system is specifically designed to avoid a centralized metadata
serve, to avoid problems where there is a performance bottleneck for metadata intensive
operations. This also improves availability, as the distributed metadata server function
provides additional insulation against a metadata server single point of failure.
SONAS implements the GPFS technology that solves this problem by managing metadata at
the node which is using the file or in the case of parallel access to the file, at a dynamically
selected node which is using the file.
These functions support cluster management and other standard file system administration
functions such as user quotas, snapshots and extended access control lists.
The SONAS file system provides functions that simplify cluster-wide tasks. A single SONAS
command or GUI command can perform a file system function across the entire SONAS file
system cluster.
The distributed SONAS file system architecture facilitates a rolling upgrade methodology, to
allow you to upgrade individual SONAS odes in the cluster while the file system remains
online. The SONAS file system also supports a mix of nodes running at current and new
release levels, to enable dynamic SONAS Software upgrades.
The SONAS file system implements quotas to enable the administrator to control and monitor
file system usage by users and groups across the cluster. The SONAS file system also
provides commands to generate quota reports including user, group, and fileset inode and
data block usage.
In addition, SONAS snapshots provide an online backup capability that allows files to be
recovered easily from common problems such as accidental deletion of a file. It is a known
requirement to increase the snapshot granularity to include filesets, directories, and individual
files, and IBM intends to address these requirements in a future SONaS release.
Access control
SONAS provides NFSv4 enhanced access control at the file system level, to allow all SONAS
users, regardless of the NAS protocol by which they access the SONAS, to be able to take
advantage of this robust level of central security and access control to protect directories and
files. SONAS file system implements NFSv4 access control lists (ACLs) in addition to
traditional ACL support.
SONAS ACLs are based on the POSIX model. Access Control Lists (ACLs) extend the base
permissions, or standard file access modes, of read (r), write (w), and execute (x) beyond the
three categories of file owner, file group, and other users, to allow the definition of additional
users and user groups. In addition, SONAS introduces a fourth access mode, control (c),
which can be used to govern who can manage the ACL itself.
The SONAS Cluster Manager function works in conjunction with the SONAS file system to
provide clustered NFS, clustered CIFS, clustered FTP, and clustered HTTPS. With the
SONAS Cluster Manager, SONAS provides a super-scalable, high performance file system
capability with simultaneous access to a common set of data from multiple interface nodes,
The SONAS file system works in conjunction with the rest of the SONAS Software to provide
a comprehensive, integrated set of storage management tools including monitoring of file
services, load balancing and IP address fail over.
The SONAS file system continuously monitors the health of the file system components. If
failures are detected, appropriate recovery action is taken automatically. Extensive logging
and recovery capabilities are provided which maintain metadata consistency when nodes
holding locks or performing services fail.
Internal data replication (SONAS Software RAID-1 mirroring) is optional and can be
implemented to provide further protection over and above the SONAS hardware storage
redundancy and RAID. In addition, the SONAS file system automatically self-replicates and
internally mirrors file system journal logs and metadata to assure hot failover and redundancy
and continuous operation of the rest of the file system, even if all paths to a disk or storage
pod fail.
Let us now examine the SONAS data management services in more detail.
Enterprise Linux
IBM Servers
We also discuss the role and use of external Tivoli Storage Manager servers to provide
accelerated backup and restore, and tor provide HSM to external storage pools. Finally, we
describe local data resiliency using Snapshots, and remote resiliency using asynchronous
replication.
3.7.1 SONAS: Using the central policy engine and automatic tiered storage
SONAS uses policies to control the lifecycle of files that it manages and consequently control
the costs of storing data by automatically aligning data to the appropriate storage tier based
on the policy rules setup in by the SONAS administrator. Figure 3-31 illustrates a tiered
storage environment that contains multiple storage tiers and each tier has specific
performance characteristics and associated costs, for example, the poolfast contains fast and
expensive disk, whereas pooltape contains relatively inexpensive tapes.
The SONAS policy implementation is based on and uses the GPFS policy implementation.
Files reside in SONAS storage pools, and policies are assigned to files and control the
placement and movement of files between storage pools.
A SONAS policy consists of a collection of rule, and the rules control what actions are
executed and against what files the actions are performed. So the smallest entity controlled
by a rule is a file. SONAS has three types of rules:
Initial file placement These rules control the placement of newly created files.
File management These rules control movement of existing files between storage pools
and the deletion of old files. Migration policies are used to transfer data
between the SONAS storage pools and to the external HSM storage
pool.
Restore of file data These rules control what happens when data gets restored to a
SONAS file system.
Each SONAS filesystem is mapped to storage pools. The default pool for a filesystem is the
system pool, also called pool1. A file system can have one or more additional storage pools
after the system pool.
The SONAS administrator assigns physical disk LUNs to a logical SONAS storage pool.
SONAS also manages external storage pools. An external storage pool is a mechanism for
SONAS to store data in an external HSM storage manager such as Tivoli Storage Manager.
Policies control the location of files among storage pools in the same filesystem. Figure 3-32
shows a conceptual representation of a filesystem, pools, and physical disk LUNs.
To simplify implementation of HSM and storage pooling, SONAS provides templates for
standard usage cases. Customized cases can be created from the default templates by using
the SONAS CLI. The standard usage cases, also called ILM profiles, are shown in the
diagram in Figure 3-34.
Peered pools:
Placement policies only
new file pool1 pool2
Tiered pools:
Files placed in pool1 and
new file pool1 pool2 then moved to pool2
The standard ILM policy profiles are based on the assumption that pool1 is the fastest pool
using the fastest storage devices such as SAS disks, and pool2 is based on less expensive
disks such as SATA. The SONAS GPFS metadata must always reside in the fastest storage
pool, pool1 in our examples, because it is the data that has the highest IO requirements when
SONAS GPFS file system scan operations are performed.
3.7.2 Using and configuring Tivoli Storage Manager HSM with SONAS basics
The use of SONAS HSM provides the following advantages:
It frees administrators and users from manual file system pruning tasks, and defers the
need to purchase additional disk storage.
It allows the Tivoli Storage Manager HSM to extend the SONAS disk space and
automates the movement of seldom-used files to and from external near line storage.
It allows pre-migration, a method that sends a copy of the file to be migrated to the Tivoli
Storage.
It manages the server prior to migration, allowing threshold migration to quickly provide
space by simply stubbing the premigrated files.
It allows you to pre-define the stub file size, which will eliminate the recall of the entire file
for programs that browse only the first part of a file.
To use the Tivoli Storage Manager HSM client, you must provide a Tivoli Storage Manager
server external to the SONAS system, and the server is accessed through the Ethernet
connections on the interface nodes.
The current version of SONAS requires that HSM be configured and managed using the CLI
as at the time of writing GUI support is not present for HSM. HSM migration work might cause
additional overhead on the SONAS interface nodes and so take care when planning the
timing and frequency of migration jobs.
When using HSM space management on a filesystem, each file in the filesystem can be in
one of three states:
Resident when the file resides on disk in the SONAS appliance
Premigrated when the file resides both on the disk in the SONAS and in Tivoli Storage
Manager HSM
Migrated when the file resides only in Tivoli Storage Manager
Files are created and modified on the SONAS filesystem and when they are present they are
said to be in the resident state.
Files in an HSM managed filesystem can be migrated to Tivoli Storage Manager HSM storage
for a variety of reasons, such as when a predefined file system utilization threshold is
exceeded. Migrated files are copied to Tivoli Storage Manager and replaced by a stub file that
has a preset size. Using a stub file leaves a specified amount of file data at the front of the file
on the SONAS disk, allowing it to be read without triggering a recall.
In a SONAS GPFS environment, a small file that is less than the 1/32 of the filesystem
blocksize, or one subblock, can become larger after an HSM migration because SONAS
GPFS adds meta information to the file during the migration. Because another block on the
file system is allocated for the meta information, this increases the space allocated for the file.
If a file system is filled to its maximum capacity with many small files, it is possible that the file
system can run out of space during the file migration.
A recall is triggered when the first byte of storage not on the SONAS disk is accessed. When
a migrated file is accessed, it is recalled from the external Tivoli Storage Manager storage into
the SONAS storage. If you have files with headers that will be periodically accessed and do
not want to trigger recalls on those header accesses, use the appropriate stub file size to
ensure that an appropriate amount of file header stays on the SONAS disk.
As data is accessed by CIFS or NFS, when a migrated file is opened and a byte of data that is
not in the SONAS cache is accessed, that access triggers a Data Management API (DMAPI)
event in the SONAS. That event is sent to the primary Tivoli Storage Manager client, that
resides on one of the interface nodes, and it triggers a recall. If the primary Tivoli Storage
Manager client is not overloaded, it issues the recall itself, otherwise it sends the recall to
another Tivoli Storage Manager client node. In practice most recalls will be performed by the
primary Tivoli Storage Manager client interface node.
Because a recall from physical tape requires waiting for cartridge fetching, tape drive loading
and tape movement to the desired file, physical tape recalls can take significant numbers of
seconds to start, so the application needs to plan for this delay.
HSM can be added to a filesystem at the time of filesystem creation or at a later time.
SONAS Snapshot technology makes efficient use of storage by storing only block-level
changes between each successive Snapshot. Only the changes made to the original file
system consume additional physical storage, thus minimizing physical space requirements
and maximizing recoverability.
At the current release level, SONAS Snapshot is a read-only, point-in-time consistent version
of an entire SONAS file system, frozen at a point in time:
Each SONAS file system can maintain up to 256 Snapshots concurrently.
Snapshots only consume space when the file system changes.
Snapshots use no additional disk space when first taken.
Snapshots are enforced to be consistent across the file system to a single point in time.
Snapshots can be taken manually or automatically on a schedule.
For CIFS users, SONAS Snapshots are readily accessible by Microsoft Volume Shadow
Services (VSS) integration into the Windows Explorer interface.
Filesystem /fs1/file1
/fs1/file2
Before /fs1/subdir1/file3
Read-only copy of
/fs1/subdir1/file4
Snapshot /fs1/subdir2/file5 directory
structure and files
Figure 3-35 SONAS Snapshot appears as a special directory in the file system
Snapshots of a SONAS file system are read-only; changes are made only to the active
(that is, normal, non-snapshot) files and directories. Snapshots are only made of active file
systems;, you cannot make a Snapshot of an existing Snapshot.
Individual files, groups of files, or entire directories can be restored or copied back from
Snapshots.
Use these
buttons
copy or
restore the
snapshot
Figure 3-36 SONAS Snapshots are accessible for Windows CIFS users by Windows Explorer
SONAS Snapshots are intended as a point in time copy of an entire SONAS file system, and
preserve the contents of the file system at a single point in time. The Snapshot function allows
a backup or mirror program to run concurrently with user updates and still obtain a consistent
copy of the file system as of the time that the snapshot was created. SONAS Snapshots also
provide an online backup capability that allows easy recovery from common problems such
as accidental deletion of a file, and comparison with older versions of a file.
SONAS asynchronous replication can also be useful for the idea of “backup-less backup”
disaster recovery, in other words, using direct disk to disk incremental change replication to a
disaster recovery remote site. This is particularly important when the raw amount of data for
backup/restore for large amounts of storage, is so large that a tape restore at a disaster
recovery site might be unfeasible from a time-to-restore standpoint.
In this section, we discuss the SONAS asynchronous replication capability that is designed to
address these requirements.
Async replication
begins by
executing a policy
IBM Scale Out NAS
1. Read policies
Global Namespace R em ote S ca le
Out NA S
Policy Engine
Inte rf ac e
node … Inter fa ce
node …
Target file system 1 snapshot
St or age
node
… Stor age
node
…
Target file system 2 snapshot
Figure 3-37 SONAS async replication step 1 - execute a policy, make snapshot at local and remote
The next step is to make a mathematical hash of the source and target snapshots, and
compare them, as shown in Figure 3-38.
Hash compare:
determines
incremental
changes to send
IBM Scale Out NAS
Global Namespace
R em ote Sc ale
Policy Engine Out NA S
Inte rf ac e
node … Inter fa ce
node … hash
hash hash
Target file system 1 snapshot
Figure 3-38 SONAS async replication step 2 - compare mathematical hash of snapshots
The final step is to exploit the parallel data transfer capabilities of SONAS by having multiple
nodes participate in the transfer of the async replication changed blocks to the target remote
file systems, as shown in Figure 3-39.
Global Namespace
R em ote Sc ale
Policy Engine Out NA S
Inte rf ac e
node … Inter fa ce
node …
Target file system 1 snapshot
St or age
node
… Stor age
node
…
Target file system 2 snapshot
File system 1 snapshot File system 2 snapshot
Figure 3-39 SONAS async replication step 3 - transfer data using multiple interface nodes
The internal snapshot at the source side assures that data being transmitted is in data
integrity and consistency, and is at a single point in time. The internal snapshot at the target is
there to provide a backout point in time capability, if for any reason the drain of the changes
from source to target fails before it is complete.
Let us review a few more details about the SONAS asynchronous replication.
SONAS asynchronous replication is designed to cope with connections that provide low
bandwidth, high latency and low reliability. The basic philosophy of SONAS asynchronous
replication proceeds as follows:
1. Take a snapshot of both the local and remote file systems. This ensures first that we are
replicating a frozen and consistent state of the source file system.
2. Collect a file path list with corresponding stat information, by comparing the two with a
mathematical hash, in order to identify changed blocks
3. Distribute the changed file list to a specified list of source interface nodes
4. Run a scheduled process that performs rsync operations on the set of interface nodes, for
a given file list, to the destination SONAS. Rsync is a well-understood open source utility,
that will pick-up the changed blocks on the source SONAS file system, and stream those
changes in parallel to the remote, and write them to the target SONAS file system.
5. The snapshot at the remote SONAS system ensures that a safety fallback point is
available if there is a failure in the drain of the new updates.
6. When the drain is complete, then the remote file system is ready for use.
7. Both snapshots are automatically deleted after a successful replication run.
The target SONAS system is an independent SONAS cluster that can be thousands of miles
away.
At the current release level SONAS R1.1.1, asynchronous replication is available for
replicating incremental changes at the file system level to one other site, meaning SONAS
machines are paired on a 1:1 basis.
Multiple file systems can be replicated between this pair of SONAS systems. As long as they
are unique pairs of file systems, a given SONAS can have certain file systems be sources for
async replication, and other file systems can be the targets.
The replication schedule is configured thru the SONAS GUI or by Command Line Interface.
Depending on the number of files included in the replication; the minimal interval will vary
depending on the amount of data and files to be sent.
Figure 3-40 shows a diagram of the SONAS and Tivoli Storage Manager configuration.
TSM client
code is pre- SONAS to TSM
installed is Ethernet only
Storage
Pod
IBM If using SONAS HSM to
SONAS external storage, an external
TSM server is used
Storage
Existing Tivoli Storage Manager customers can make use of existing Tivoli Storage Manager
servers to backup SONAS. SONAS is designed to perform Tivoli Storage Manager functions
on SONAS with only a few commands that need to be issued on the Tivoli Storage Manager
server and the SONAS system to make use of Tivoli Storage Manager with SONAS, and to
perform scheduled backups or HSM data movements.
In SONAS, Tivoli Storage Manager backup is performed over LAN through the interface
nodes to the Tivoli Storage Manager servers. It is not possible to do LAN-free backup at this
time from SONAS to the Tivoli Storage Manager server.
If you use Tivoli Storage Manager to archive files to Tivoli Storage Manager storage, those
files are removed from your local file system, and if needed later, you retrieve them from Tivoli
Storage Manager storage.
To users, the files appear to be online in the file system. If the migrated file is accessed, Tivoli
Storage Manager HSM will automatically initiate a recall of the full files from their migration
location in external Tivoli Storage Manager-attached storage. The effect to the user will
simply be elongated response time while the file is being recalled and reloaded into internal
SONAS storage. You can also initiate recalls proactively if desired.
Note that the Tivoli Storage Manager backup is not using the classical Tivoli Storage Manager
backup process to traverse the filesystem, query the server and identify the changes.
Instead, the SONAS Software is called to use the high performance scan engine and the
policy engine to identify changes in the filesystem, and to generate a list of files that need to
be expired, and a list of files that need to be backed up.
Scripts are provided with the SONAS Software to define the interface nodes involved in the
backup, the relationship of which filesystem needs to be backed up to which Tivoli Storage
Manager server, and to start/stop backups or restores.
Do not consider the use of SONAS HSM with Tivoli Storage Manager as a replacement for
backup/restore. Rather, consider HSM as an external storage extension of local SONAS disk
storage.
When a file is migrated using the Tivoli Storage Manager HSM server to the external Tivoli
Storage Manager HSM storage, there is still only one copy of the file available, because the
original is deleted on the SONAS file system itself, and replaced by the Tivoli Storage
Manager/HSM stub file. Also, HSM with Tivoli Storage Manager maintains only the last copy
of the file, giving no opportunity to store multiple versions. In comparison, Tivoli Storage
Manager backup/archive (or typically any backup/archive software) gives you the full ability to
storage multiple backup versions of a file, and to track and manage these backup copies in an
automated way.
According to Tivoli Storage Manager best practices, you must back up / archive a file before
the file has been migrated by Tivoli Storage Manager HSM to external storage. With proper
configuration, you can specify in Tivoli Storage Manager management classes that a file is
not eligible for HSM migration unless a backup has been made first with the Tivoli Storage
Manager backup-archive capability. This is an extremely useful function.
Another good reason for using the same Tivoli Storage Manager server for both backup
restore and HSM, is that if the file is migrated and the same Tivoli Storage Manager server
destination is used, the Tivoli Storage Manager server, outboard of the SONAS and without
any involvement from the SONAS, can copy the file from the Tivoli Storage Manager HSM
migration storage pool to the backup pool without recalling the file, which is another very
useful capability.
As of the writing of this book, the Tivoli Storage Manager client version 6.1.3 is used internally
in SONAS Software. There is no need to order it or install it separately on the interface nodes.
If you are not using Tivoli Storage Manager functions, there is no charge for the fact that this
Tivoli Storage Manager client code is present in the SONAS Software.
Note that an IBM Tivoli Storage Manager server talks to two unique types of Tivoli Storage
Manager clients. One client is backup/archive client for backup, restore, archive and retrieve
operations, and the other is a separate Tivoli Storage Manager HSM client for migrating and
recalling files. Both of these Tivoli Storage Manager client software are pre-installed on the
SONAS Software at the factory and are present on every SONAS system; you do not need to
order them or install them yourself. If you are not using Tivoli Storage Manager software
functions, there is of course no charge. If you are using Tivoli Storage Manager software
functions, then there is the normal Tivoli Storage Manager licensing charge for the
appropriate Tivoli Storage Manager client software, however, this is just on the interface
nodes that you are attaching to the Tivoli Storage Manager server.
If you use Tivoli Storage Manager functions, then you are required to pay a license charge for
the appropriate Tivoli Storage Manager client code. You only have to pay the license charge
for the interface nodes that are attached to Tivoli Storage Manager servers. The licensing is
for the processor value units of the SONAS interface node for a Linux server. Note that this
Tivoli Storage Manager client licensing is not for the terabytes of storage that might be on the
SONAS.
As for the Tivoli Storage Manager server, this Tivoli Storage Manager Server licensing (and if
necessary, the associated servers that the Tivoli Storage Manager Server will run on) needs
to be licensed and setup independently of SONAS.
If you are an existing Tivoli Storage Manager server user, then you can, of course, use that
existing Tivoli Storage Manager infrastructure and licenses with IBM SONAS.
Various Redbooks publications and white papers provide further information about Tivoli
Storage Manager and its sizing guidelines, performance optimizations, and tuning knobs.
SONAS Management GUI runs on the SONAS Management Node, and is web-based (https).
It provides role-based authorization for users, and enables the administrator to maintain the
SONAS cluster. These roles are used to segregate GUI administrator users according to their
working scope within the Management GUI. These roles are as follows:
Administrator: This role has access to all features and functions provided by the GUI, and
is the only role that can manage GUI users and roles.
Operator: The operator can do the following tasks:
– Check the health of the cluster.
– View the cluster configuration.
– Verify the system and file system utilization.
– Manage to set thresholds and notifications settings.
Export administrator: The export administrator is allowed to create and manage shares,
plus perform the tasks the operator can execute.
Storage administrator: The storage administrator is allowed to manage disks and storage
pools, plus perform the tasks the operator can execute.
System administrator: The system administrator is allowed to manage nodes and tasks,
plus perform the tasks the operator can execute.
SONAS has a central database that stores configuration information and events. This
information is partly pulled by the management node and partly pushed to the management
node by the other nodes in the cluster.
Panels are available for most of the major functions, as shown in Figure 3-41.
Figure 3-41 SONAS Management GUI has panels for most aspects of SONAS
External network
send/receive
average MB/sec
File systems
status
Internal data
network
performance
Storage node
status and
performance
Each of the icons is clickable, and will expand to show status of an individual components.
The SONAS Management GUI is the focal point for extended monitoring facilities and the
SONAS Health Center.
SONAS tracks historical performance and utilization information, and provides the ability to
graphically display the current and historical trends.
Figure 3-43 SONAS Health Center historical system utilization graphical reports
The length of time that can be reported is determined by the amount of log space set aside to
capture data.
The SONAS CLI is designed to be familiar to the standard UNIX, Windows, and NAS
administrator.
SONAS Software provides this functionality in one license, which includes all of the
components shown in Figure 3-44.
Enterprise Linux
IBM Servers
The SONAS Software provides the ability for central management of storage, providing the
functionality for a highly automated, extremely flexible, and highly scalable self-managing
system. You can start with a small SONAS system with less than 100 TB, and continue to
seamlessly grow and linearly scale and increase performance, using the SONAS Software to
manage scalability at petabytes.
/a ppl /hom
/ho e/ap
me/a pppl/data
l/d ata/w
/weeb /big_arch
b/big_ architectur e_ drawing.ppt
itect ur e_drawing.ppt
/d ata
/ho
/homme/a
e/appppl/data /weeb
l/d ata/w b/unstr
/uns tructur ed_ big_
uc tured_b ig_vide o. mpg
video.m pg
/web Logical
I nt er f ac I t erfac
n I nt er f ac In e
t rf ac I t er fac
n I nt er f ac
In t er fa c I nt er f ac In e
t rf ac I nt er f ac
…>
e no des e n od es e n od es e no des e n od es en o des e no des e n od es en o des e no d es
In t er fac I nt er f ac In e
t rf ac I nt er fa c I nt er f ac I n t erf ac
I nt er f ac In t er fac I nt erf ac In e
t rf ac
e n od es en o des e no des e n od es en o des e no d es e n od es e no des e no d es e n od es
… … … … … …
… … …
Stor ag Stor a
g Stora
g S tor a
g St o
r ag Stor ag St o
r ag Stor ag St o
r ag Stor a
g Stor ag Stor ag St orag St o
r ag St roag St roag St orag Stor ag St roag Stor a
g
… > scale
… > out
e e e e e e e e e e e e e e e e e e e e
nodes nodes nodes nodes nodes nodes odes
n nodes nodes nodes nodes nodes oe
n ds odes
n nodes nodes oe
n ds nodes nodes nodes
St roag Stor ag St o
r ag Stor ag S tor a
g S toa
rg St orag St o
r ag S tor g
a St o
r ag St orag Stor ag Stor ag Stor a
g Stor a
g S tor g
a S toa
rg St orag Stor a
g St o
r ag
…>
e e e e e e e e e e e e e e e e e e e e
odes
n nodes nodes nodes nodes node s oe
n ds odes
n nodes nodes oe
n ds nodes nodes nodes nodes nodes node s oe
n ds nodes nodes
Stor a
g St o
r ag S tor a
g St o
r ag Stor ag Stor ag Stor ag Stor a
g Stor a
g S tor a
g Stor ag St orag St o
r ag St roag St roag Stor ag Stor ag Stor ag Stor a
g Stor a
g
e e e e e e e e e e e e e e e e e e e e
nodes odes
n nodes nodes nodes nodes nodes nodes nodes nodes nodes oe
n ds odes
n nodes nodes nodes nodes nodes nodes nodes
…
… >
> scale
…
… >> out
etc….. etc….. etc…..
Figure 3-45 SONAS Software manages all aspects of a maximum size SONAS
As storage needs continue to grow over time, the SONAS Software is designed to continue to
scale out and support even larger configurations, while still maintaining all the storage
management and high performance storage characteristics that we have discussed in this
chapter.
In the remaining chapters of this book, we review typical SONAS usage cases and ISV
software integration.
A much more detailed discussion of SONAS hardware and software architecture can be
found in the companion Redbooks publication, SONAS Architecture, Planning, and Basic
Implementation, SG24-7875.
We can make the following general comments about these types of NAS.
Classical NAS storage:
This class of storage is general purpose file storage, and usually has one or two
controllers in an active-active or active standby configuration. The large growth of
classical NAS storage has been due to it’s ability to provide cost-effective, easy-to-use
storage for the small to mid-range NAS market space
Scalable NAS:
This class of NAS storage has arisen with features that are designed to resolve
considerations and restrictions in classical NAS storage, in terms of scalability, user
manageability, and expandability. The focus is on resolving the usability and management
from a user perspective.
Scale out NAS:
This final class of storage has arisen to address limitations and restrictions in scalable
NAS architectures. In this case, a parallel clustered file system is implemented, with global
distribution of workload management, workload balancing, and centralized management
and deployment capabilities.
Each NAS filer presents a namespace to the network as shown in Figure 4-2.
Scale Name
D i sk Enc los ure
Scale Name
D isk En closu re
Scale Name
D i sk Enc los ure
Up Space
D isk En clo su re Up Space
Di sk Enc los ur e Up Space
D i sk Enc los ure
#1
D isk En clo su re
D isk En clo su re
#2
Di sk Enc los ur e
Di sk Enc los ur e
#3
D i sk Enc los ure
With today’s highly powerful microprocessors and large commodity storage, classical NAS
filers have provided and continue to provide a cost-effective solution for users.
However, as the storage continues to grow in size, considerations appear when considering
large scale and using classical NAS storage servers:
Each NAS filer is an independent namespace. There is no sharing of resources with
another NAS filer or another namespace.
Failover occurs between primary and secondary filer controllers within the filer, however,
there is no ability to failover to any of the other NAS filers.
Ability to scale is limited to the controller capacity. When the controller is at capacity, then
another NAS filer must be implemented; each of these NAS filers must be managed
independently.
Balancing of workload between NAS filers must be done manually, and when data is
moved between filers, the corresponding changes need to be made in the domain name
servers, authentication servers, and the user paths and directory definitions. As the
number of NAS filers increases, this can increasingly become a management problem.
The characteristics and considerations when considering implementing large scale NAS
storage using Scalable NAS architectures are:
Scalable NAS typically provides a software re-direction layer. providing the appears to all
users of a single (virtual) global namespace. This software re-direction layer can be
implemented in multiple ways; it can reside in any of the following layers:
– NAS controllers
– Network switches and routers
– Servers which front-end the network
Regardless of where the software re-direction layer is, the goal is to provide the
appearance of a single virtual global namespace to the users, thus minimizing or
eliminating changes to the user environment, The software layer redirects the virtual
global namespace among the actual physical namespaces that reside on separate
physical NAS storage controllers.
However, at the physical level, each NAS file and controller set continues to be
independent. Data might have limitations in how widely it can be spread across multiple
controllers (if it can be spread at all), and failover might be limited to a certain subset of the
controllers.
From a physical implementation standpoint, the same physical controllers that were in the
classical NAS type of storage are used. At a physical level, there might not be any change
to the ability to failover to other controller pairs, and physical tuning might continue to
required manual monitoring and adjustment.
If data is physically moved, then coordination is required between the virtual software
redirection layer, and the physical controllers and their associated physical namespaces.
Certainly, scalable NAS implementations have value and can address a few of the many user
namespace management concerns. However, these implementations have not addressed
D isk En clo su re
Logical Storage
Di sk Enc los ure
Tier Gold Di sk Enc los ur e
Di sk Enc lo sur e
Logical Storage Tier Copper
D isk Encl osure Di sk Enc los ur e
Figure 4-4 Example of scale out NAS storage architecture as implemented in IBM SONAS
We define a true scale out NAS architecture as one where the physical implementation of the
filer has been re-architected, to provide the following capabilities:
Provide a true parallel global clustered environment for network serving, data serving,
any-to-any failover, global workload balancing, global centralized policy management,
global logical storage pools, integrated tiered storage
Provision of a true physical global namespace that can scale to petabyte levels, that can
remain constant and unchanging regardless of scale, for all users to see/access.
Utilize commercial off the shelf (COTS) enterprise servers, in order to take advantage of
today’s fast pace of technology advancement. Utilize these servers in a grid, where all
nodes share workload equally, and all nodes of a particular role can dynamically failover to
any other node with the same role
Provide integrated, advanced features such central policy-managed automated tiered
storage, cross-platform (UNIX, Linux, Windows) support of concurrent reads and writes to
the same file from within the global namespace, wide parallel data stripings, and high
performance scan engines
IBM SONAS is designed as a true scale out NAS architecture. SONAS is designed for NAS
environments where large scale, high performance and capacity requirements for file serving
are present, that cannot be adequately served by classic NAS or scalable NAS architectures.
Each of the features above are implemented within SONAS.
Medical
Advanced Financial Imaging
Search Analytics
3D On-Line
Infotainment
Video
Surveillance
Figure 4-5 New demands on IT for large scale, real-time data analysis and prediction
Let us examine the requirements that these applications are placing on the IT infrastructure.
These requirements are the driving factors behind the IBM approach to SONAS, and are the
reason for the IBM SONAS design and implementation.
However, each classical NAS filer has to be individually managed, and as the scale of today’s
IT systems have dramatically grown, the following issues and requirements have occurred:
Tremendous growth of “unstructured data” in recent years
That must be managed with the same (or less) administrative staff
Limitations as described earlier in scalability in both performance and capacity
No parallel access to data, yet today’s data rate requirements are continually climbing
Multiple filers have to be managed individually
These issues, and the value that IBM SONAS is designed to provide, can be summarized as
shown in Figure 4-6.
Figure 4-6 SONAS Usage case - large scale NAS management and administration
IBM Scale Out NAS is designed to provide a solution to these large scale requirements, by
providing a super-scalable, scale out NAS cluster that is centrally managed, centrally
deployed, and provides integrated, automated tiered storage, with centralized management of
logical storage tiers (instead of individual disks and filers).
$ $ $ $
$
Figure 4-7 SONAS Usage case - large scale NAS storage utilization, tiered storage
On the left, we see a typical case in the classical NAS environment, where:
Capacity must be managed individually by each NAS filer.
Average storage utilization is not as high as we prefer; it many environments, it can easily
be less than 50%.
All files are online in the file system, but large amounts of these files (perhaps as many as
80% of the files) have not been accessed during the last 6 months, legitimately can be
migrated to lower classes of storage, or to external tape storage.
Database
End Users
Application
File Servers
Servers
SONAS can transform the foregoing environment into a globally virtualized file server
environment that can be seen in Figure 4-9.
Global
Database
virtual
file
SONAS
server Global Global namespace
Virtual
File High performance
Server Protocols Management petabyte
CIFS Central Availability
Clustered NFS Administration Data Migration scale
Auto-failover HTTP Monitoring Replication
FTP File Mgmt Backup
Tiered storage
Auto-Tiered De-dup VT L
Storage Or tape
High perf.
Scan engine Automated movement between
tiered storage pools
These values have a broad applicability, across a multitude of industries and applications, and
are valuable to any industry and any environment needing easy-to-manage, high
performance, scalable NAS storage.
In the rest of this chapter, we examine a few of these usage cases in more detail:
Digital media and entertainment
Analytics for competitive advantage
Deep computing, high performance computing
Healthcare
Cloud Storage
Today in a media environment we are facing the convergence of two key trends:
The growing demand for rich media content
The exploding availability of affordable bandwidth.
This is causing enterprises and organizations of all types to evaluate how they can more
efficiently and strategically create, manage and distribute digital media.
The growing demand for compelling content, affordable bandwidth, and advancing network
technologies are prompting organizations to evaluate how they can efficiently manage the
digital media content throughout its entire lifecycle, as shown in the following figure. Efficient
digital media management can be used for the following purposes:
Increase the media impact on audiences
Enhance user experiences
Improve productivity
Tighten security
Reduce costs
Contents reutilization
Generate new revenue
The Digital Media environment basically can be understood as a three stage data lifecycle,
broken down into the Create, Manage, and Distribute sections.
In today’s digital media industry, this lifecycle is serviced by a large number of integrated
technologies provided by multiple technology vendors, including many technologies from IBM
and IBM partners. Consulting and integration skills in digital media are used to take
advantage of existing and evolving digital media business solutions.
This ecosystem can be diagrammed in detail as a digital media workflow (Figure 4-12):
Ingest &
Animation Indexing PC
Interactive Engine
Media Streaming
Media Server Wireless
Digital Rights Device
Video System
Content Edge
Infrastructure Content
Management Cache Wireless
Text Services Aggregation Gateway
System
Business
Image Rights System Kiosk
Distribution
E-sourcing Storage Scheduling WAN
Multi- Media
Audio repository
Gateway
Search
Middleware Managed Transaction
Business Server Playout
Hosting Server
Other Return
Data Network
The foregoing detailed digital media open “ecosystem” is commonly implemented today using
widely available common specifications for formats, protocols, and APIs - which facilitate
interoperable solutions based on multiple vendors, including IBM vendors and IBM business
partner components. This ecosystem framework provides the assurance of a solid
architectural and organization foundation for viable, repeatable, efficient workflows when
dealing with heterogeneous platforms and applications (which is almost always the case in
digital media).
From a competitive advantage standpoint, digital media production is all about managing the
workflow, making it efficient, and speeding the time to completion.
To bring value to this workflow, IBM SONAS is designed to provide the time-saving,
labor-saving efficiency of large central high performance shared data storage, providing fast
ingest, secure yet sharable data storage, and high performance delivery capability to this
ecosystem. SONAS as a centrally managed, centrally deployed, shared storage server allows
the digital media organization to implement a robust set of media-based accelerated
workflows that can quickly deliver projects, and move media through the workflow at ever
accelerating time to delivery. At the same time, SONAS provides standard NAS storage
interfaces to allow taking advantage of already installed application and technology
infrastructure.
The value of SONAS to today’s dynamic digital media workflows by providing centrally
managed, cost-effective, high performance storage, can be summarized as shown in
Figure 4-13.
Customers
(Playout,
Information & online, etc.)
Archives
Message &
Workgroups
Ca Control Services r
talo live
g De
Ac
c es
s
External
Central Deliver Customers
Storage
shes
& Ru
i ves
Arch Acc
Ingest ister ess
Workgroups Re g ro ug h
Acc / fi
ne c
og Pub ess u ts
dL
n l i sh
ss
Reg is
ea ts fini s
ni z ce Cu hed
Or
g a
Ac /Fine pr o
gr a Post
ter C
h m
ug Production
h Ro craft edit
onten
s
bli finishing
Production Pu
Workgroups
t
External
Production
Companies
Figure 4-13 Digital media - optimizing workflow using centralized, shared storage
We consider how this common theme of exploiting a centrally managed, high performance,
highly sharable storage subsystem is common across many digital media applications.
Broadcast environment
In a typical broadcasting environment, there are three distinct areas working with the high
quality digital streams (video and audio):
Ingest stations (incoming stream)
Non-linear editing (NLE) stations (handling stream)
Play out server (play the stream out)
In addition, for any high resolution stream, there will be a low resolution version created
(for catalog and archiving purposes). You will also find common IT tasks such as
backup/restore and database operations.
Analyzing an NLE system only, you typically have from two to five parallel data clips
referenced at a time, and most of the action is “reading,” which includes shuttling through the
material. When you decide to render the material, the read/write ratio is almost 50/50,
because an NLE system reads exactly one stream as input and writes one stream as an
output (with caching in between).
Looking in addition at an entire production day, you might consider that a lot of time is used by
seeking for appropriate material and this is read-only activity. In a typical NLE environment,
there are multiple files open at a time, and the result is one file being written when rendered.
Also, most of the edit work involves to shuttle through the content, which is heavy read
activity.
Editing
Direct & Near-line
Broadcast
Media Asset
Mgmt
Archive/HSM, DR
Copy
Compositing
Dubbing, graphics
Figure 4-14 Digital media - SONAS usage case for broadcast environment
Let us next see how SONAS might apply to a post production environment.
Post-production
Compared to broadcasting, the post-production (mastering and distribution) environments
require much higher bandwidth, because the resolution of the content is higher. Also, the
access behavior in post-production environments is more sequential, because this is
predefined work, which involves a “shuttle-in,” then working with the editing or color correction
stations. Hence, the material will be opened for read, updated on the workstation, and
written/rendered back to the storage.
The conclusion we can draw is that a centralized storage is the key to make the workflow
more efficient and to save costs. The storage in such an environment must meet the following
demands:
Storage must be scalable: Because islands are difficult to manage and/or are slowly
accessible (possibly needing their own content management), islands must be avoided.
Storage must perform: It must provide the speed and bandwidth for multiple read and write
streams that must be handled without interruption (concurrent access).
Storage must be highly available and must provide for data protection: Production must be
online at all times.
It is fair to recognize that storage is not the only component involved. However, the storage is
foundational key to analyze and improving the workflow capability of all infrastructure
components. Every component can have a big influence (impact) in the infrastructure’s
performance, scalability, or high availability.
In all cases, it comes back to having centrally manageable, high performance storage that
can be utilized across the entire workflow to operate in an optimum manner. This is the value
that centralized SONAS storage is intended to offer in the usage case for post-production, as
shown in Figure 4-15.
Film
Recorder
Ingest
Scanner, DataCine, File System (s)
Camera
SONAS
Media Asset
Mgmt
Archive/HSM, DR
Copy
Effects
Autodesk, DaVinci, CG, etc.
Figure 4-15 Digital media - SONAS usage case for post-production environment
Distribution
Upload
to Content
Ingest File System (s) Providers, CDNs
Upload Server, etc
SONAS
Media Asset
Mgmt
Archive/HSM, DR
Copy
Encoding
Figure 4-16 Digital media - SONAS usage case for content archiving and distribution
In summary, SONAS is designed to provide these values to the general NAS consolidation
and operations streamlining digital media usage case:
Provides a super-scalar, expandable NAS storage appliance: Providing ease of ordering,
deployment, and management
Provides centralized management of all NAS files, single namespace: Providing reduced
administrative costs and faster provisioning, along with high performance response time to
end users
File Placement policies including automation: Providing optimized storage costs, reduced
administrative costs
Provides all functions in one IBM SONAS Software license: Thus removing the need to
manage individual chargeable add-on software, leading to reduced Total Cost of
Ownership and simpler procurement process
Automated policy based hierarchical storage management: Providing global Hierarchical
Storage Management, integrated tiered storage, leading to reduced administrative costs
and optimized storage asset utilization
Independent scalability of storage and nodes: Providing simple but flexible configurations
that can be tailored to any customer specific workloads
Truly parallel file system with concurrent access to files from all nodes, from multiple
platforms, while maintaining high performance: Providing an optimum storage platform for
very high performance yet providing ability to reduce administrative and management
costs even as the file system grows to petabytes.
SONAS provides all of these values and more to the digital media environment.
Let us now turn our attention to other modern IT application and environments that require
this kind of storage capability.
In the face of exploding data volumes and shrinking batch-time windows, these businesses
are struggling to make real-time or near-real-time decisions, and thereby gain competitive
advantage.
If this data can be unleashed and utilized properly, then modern day businesses can make
better decisions to drive sales, improve processes and services, boost team productivity,
reduce the risks inherent in doing business, and streamline relationships with customers,
trading partners, and suppliers. These are the goals of today's modern day analytics
applications.
To achieve these goals, we fundamentally must supply a way to centralize and share data
with high performance on an enterprise level, efficiently and cost-effectively.
High performance, high concurrency, highly scalable, high capacity, easily manageable
central storage (such as IBM SONAS) provides a way to do exactly that, as shown in
Figure 4-17.
Spreadsheets
Data Model
central
storage
Cubing Services
Figure 4-17 SONAS provides ability for centralized, shareable data for enterprise-wide analytics
To understand and fully take advantage of the power of this usage case, we need to
appreciate the advanced characteristics of today’s modern analytics applications, and thus
understand the specific kinds of high performance and concurrent data access that they
demand. In the remainder of this chapter we see how the high performance central storage
(such as IBM SONAS) provides significant advantage in the usage case of supporting today’s
competitive advantage analytics applications.
Users
authorization
advantage
Security
Analytic Process Layer
applications Real-time computing and anal ysis, stream computing,
entity analytics, data mining, content management, text
analytics, etc.
Infrastructure layer
Virtualization, central end to end management, control,
data proximity, deployment on global virtual file server
with high performance scale out NAS storage
Cloud Storage
IBM SONAS IBM SONAS scale out based on IBM
technology capacity / performance SONAS
Figure 4-18 Analytics - location of shared, scalable storage in analytics application stack
In this diagram, you can see that SONAS provides a foundational of storage. But what kinds
of data access patterns are generated by today’s modern analytics applications? The answer
to that question will illustrate the reason that SONAS can be a significant differentiating
solution to the greatly expanded high performance, high scale types of data access required
by modern image-intensive, real-time compute intensive, stream computing applications.
WAS
We b
Google Gadge ts
Enterprise Applications
Quickr
Connections
Commerce
ERP
ECM
CRM
Legacy
Dashboards,
Mashups and Search
Figure 4-19 Analytics dashboards, mashups, and search are more data intensive than ever before
These kinds of application requirements are further expanded by the competitive need to
provide real-time updates on continuous basis, as depicted in Figure 4-20.
With stream computing, one can execute a process similar to a “continuous query,” however,
it is a query that continuously identifies and tracks the constantly changing data, identifying
peaks and trends of “analyze all traffic and energy usage data within 50 kilometers of city
center” - creating continuously generated and up-to-date real-time results. These results
might be broadcast to a broad variety of locations, in addition to mashup/dashboard/search
users. For example, these results can be sent as “smart city” controls and energy
reallocations to a widespread network of large and small and mobile devices such as GPSs
and mobile handsets - and be continually refreshed over time.
In the older traditional case of data queries, static questions are asked of static data; in the
stream computing case, continuously changing data is continuously evaluated by
continuously evolving questions. Modern day stream computing analytics in this way creates
an imaginative new extension of what is possible. A simple view of this new paradigm is
shown in Figure 4-21.
One can easily imagine how that this kind of computing starts to rapidly create major new
competitive advantage capabilities. As an example, consider how modern day analytics can
be applied and integrated into financial market workflow for trading decisions, as shown in
Figure 4-22.
A key aspect of this real-time stream computing technology trend is that extracting data from
a traditional OLTP database, and then loading it to a data warehouse, is becoming viewed as
strategically non-competitive. Traditional analytics had previously required data to be
extracted out of the transaction databases and loaded to the data warehouse, and this also
resulted in the creation of multiple silos of information. In addition to that problem, what is
real-time about a time-consuming extract of data from one source, and loading it to another.
The key requirement to enable real-time stream computing analytical processing has
therefore rapidly moved in the direction of converging the transaction data and the data
warehouse, and imbedding the analytics, onto a single data source.
This has been an elusive goal for businesses for a while now; with SONAS as a foundation,
we are closer to that goal.
IBM SONAS helps achieve this goal by providing the required parallel data access, very high
scalability, linear increase in performance, manageability, and high performance
characteristics required for such environments. SONAS can provide a foundation for both a
single data source for data and also real-time analytics - directly producing the ability to
implement real-time competitive advantage applications as a result of the reduced latency in
getting access to the data in a globally shareable manner.
With these capabilities now available through SONAS, now OLAP analytics can be performed
without moving the data. This no-copy analytics capability thus sets the stage for more even
more innovation in IT competitive advantage.
Data mining has always used advanced statistical techniques and mathematical algorithms to
analyze very large volumes of historical data. The objectives of data mining are to discover
and model unknown or poorly understood patterns and behaviors inherent in the data, thus
creating descriptive and/or predictive models to gain valuable insights and predict outcomes
with high business value.
One modern example is to use data mining to improve the human condition by highly tailored
health care. By studying vast amounts of health and disease characteristics spread across
entire populations, consider the possibilities of identifying medication formulas that have been
proven to be specifically effective with other patients that have a similar genetic and DNA
makeup as your own, as shown in Figure 4-23.
Data mining
Data mining
Process
Identify
‘clustering’
Figure 4-23 Analytics - data mining in real time on the Online Analytic Processing database
However, modern real-time analytics-based data mining, takes this to a new level. One of the
great advantages of real-time, in-database mining, compared to mining in a separate
analytical environment, is direct access to the data in its primary store rather than having to
move data back and forth between the database and the analytical environment. This
approach enables the mining functions to operate with lower latency, supporting real-time or
near-real-time mining as data arrives in the data warehouse, particularly with automated data
mining processes.
Modern data mining thus includes new descriptive and predictive techniques. Descriptive
mining methods (sometimes called “unsupervised learning”) include clustering
(segmentation), associations (link analysis), and sequences (temporal links), and much more.
Summary
When requirements include the need to build a modern analytic-centric IT competitive
advantage application environment, IBM SONAS can provide strategic benefits in delivering
the storage scalability, capacity, high performance, concurrent access, and central
manageability.
To achieve these goals, SONAS can provide a fundamental storage solution to centralize and
share data with high performance, for enterprise-wide analytics storage, with efficiency and
cost-effectiveness, as shown in Figure 4-24.
Spreadsheets
Data Model
IBM
SONAS
Storage
Cubing Services
Link: A copy of this booklet can be downloaded from the Internet at the following link:
ftp://ftp.software.ibm.com/common/ssi/pm/bk/n/imm14055usen/IMM14055USEN.PDF
From a usage case perspective, the goal of Deep Computing is to create solutions that help
clients develop deep insights that lead to innovation and competitive advantage. Obviously, a
large amount of storage is usually needed when building solutions that mean to exploit
today’s modern Deep Computing capabilities.
Today's deep computing can span many industries and many applications, as shown in
Figure 4-26:
Automotive/Aerospace
Engineering
Government & Scientific research,
Automotive, Aerospace
and Defense
Higher Educationweather/environmental
classified/defense,
These powerful technologies hold the possibility of not only of providing innovation and
competitive advantage, but also of enabling a smarter planet that improves the human
condition by stopping epidemics, inventing new medicines, understanding nature, and
protecting the environment. Modern deep computing capabilities can acquire, compute, and
create massive amounts of data, in pursuit of truly magnificent results (Figure 4-27).
All of these technologies require storage. Not surprisingly, the storage requirements for Deep
Computing environments faces the same challenges in high performance, high scalability,
manageability, concurrent parallel access, high availability, and cost-effectiveness as was
previously described in 4.2.1, “General NAS consolidation and streamlining of operations” on
page 110.
SONAS can provide high performance streaming storage capability for Deep Computing
environments where the throughput requirements for any specific application or specific user
can be handled by the high throughput capability of a single SONAS interface node (with an
appropriate amount of attached SONAS storage).
For Deep Computing, SONAS provides all of these capabilities of IBM GPFS, in an easy to
use, appliance form factor:
Provides a super-scalar, expandable high performance NAS storage appliance: Providing
high performance with ease of storage management, central deployment, and high
scalability
Provides centralized management of all NAS files, single namespace: Providing reduced
administrative costs and faster provisioning, along with high performance response time to
end users
File Placement policies including automation: Providing optimized storage costs, reduced
administrative costs for managing a Deep Computing environment
Automated policy based hierarchical storage management: Providing global Hierarchical
Storage Management, integrated tiered storage. This is especially useful when large
amounts of Deep Computing research data needs to be staged onto and off of the actual
disk storage, allowing storing of petabytes of data on terabytes of disk. Reduced
administrative costs and optimized storage asset utilization also can result
Independent scalability of storage and nodes: Providing very scalable, flexible
configurations that can be tailored to many types of Deep Computing workloads
Truly parallel file system with concurrent access to files from all nodes, from multiple
platforms, while maintaining high performance: Providing an optimum storage platform for
very high performance Deep Computing, supporting multiple processes running against
the same Deep Computing data, even as the file system grows to petabytes
Automatic load balancing and high availability, with integrated wide data striping:
Providing very high performance and fast access to data, in a 24 x 7 x 365 environment
High performance scan engine: Providing integrated ability to perform metadata scanning
at a very high rate of speed, thus allowing ability to locate, stage, and destage data in a
large data farm that can be petabytes in size
Integrated high performance backup / restore / HSM: Providing integration of the scan
engine and parallel data serving with IBM Tivoli Storage Manager, thus providing the
ability for high performance backups / restores / recalls of data that had previously been
migrated to external storage
Snapshots and asynchronous Replication: For data protection and disaster recovery
For more information on IBM General Parallel File System, see the URLs shown in
Figure 4-28.
http://www.ibm.com/systems/gpfs
http://www-03.ibm.com/systems/software/gpfs/resources.html
Figure 4-28 Deep computing - URLs for IBM General Parallel File System
Innovation is spanning multiple disciplines, and the creative solutions of the future are the
result of ever increasing collaboration and sharing of information between universities,
research institutes, and industry. In the end, it is not about the technology, it is what we can
do with it all that counts. Certainly, centrally manageable, securely sharable, high
performance storage clearly plays a key role in this collaboration and innovation.
With the availability of IBM SONAS, IBM now has an additional storage option to add to the
existing portfolio of IBM Deep Computing capabilities.
SONAS storage provides an additional capability, not a replacement for IBM General Parallel
File System. SONAS provides a means to gain access to powerful IBM GPFS parallel file
system deep computing storage capabilities, while simultaneously reducing the skill and
implementation time requirements that might have previously been needed to gain access to
these storage functions and performance.
Deep Computing and high performance computing installations are strongly influenced by
application needs. These applications can have a wide array of requirements, and the
specifics of the environments will determine when SONAS can provide the needed
capabilities, and when GPFS might be the preferred choice in others.
Healthcare
Ecosystem
PHARMACIES
RADIOLOGY LABS PATIENTS
Terminology
Mapping
Claims Message
Translation Normalization GOVERNMENT
PHYSICIANS’OFFICES
ACCESS CONTROL
PAYERS
THERAPEUTICS
1
Figure 4-29 Healthcare - ecosystem
There are many aspects to modern healthcare. We discuss one of the most important and
data intensive of these applications: modern medical imaging. Medical images are at the
center of today’s healthcare diagnostic procedures. Images have provided not only a
noninvasive means to view anatomical cross-sections of internal organs, tissues, bones, and
other human body features, but also as a means for physicians to evaluate the patient’s
diagnosis, monitor effects of treatment, and for scientific investigators to conduct research.
Medical images come from a broad spectrum of imaging sources such as computed axial
tomography (CT), magnetic resonance imaging (MRI), digital mammography and positron
emission tomography (PET), and they generate a staggering amount of image data and
important medical information. Modern medical healthcare therefore has large scale, high
performance, long term medical image data repository as a central, critical need. This image
data repository can contain comprehensive patient records, including information such as
demographics, medical history, clinical data and related diagnostic images, and
post-processed images.
With a large volume and complexity of the data, as well as diversified user access
requirements, the implementation of modern medical Picture Archiving and Communication
System (PACS) are sophisticated tasks which of course, require large amounts of centrally
manageable, cost-effective, highly available storage. Specific requirements for long term
secure archival and retrieval systems, that are scalable and upgradable with time, are also
part of the scope of these technologies. IT storage industry studies have indicated that by
2010, as much as 30% of the world’s storage can be medical images1 (Figure 4-30).
10X
18MB/2D
image
2004 2007
Regardless of the actual number, it is clear that storage requirements for Healthcare
environments supporting modern medical imaging and diagnostics, faces the same
challenges in high performance, high scalability, manageability, concurrent parallel access,
high availability, and cost-effectiveness as was previously described in 4.2.1, “General NAS
consolidation and streamlining of operations” on page 110.
We can envision IBM SONAS as providing capabilities that can help address these vital
needs for the future. In particular, previously described SONAS capabilities can be useful for
these requirements of a medical imaging system:
Medical image data is stored and managed based on integrated ILM rules:
– Rules dictate: number of replicas, location of replicas and tier of storage
– Data can be automatically destaged and restaged either on demand or according to
central policy engine
– Aligns value of data to storage costs. Can easily change criteria of data over time or
according to other needs
Medical image data is cached and indexed
Medical image data stored and retrieved using standard interface (such as CIFS/NFS):
– All other operations are transparent to users / applications
Optimized performance:
– Parallel loading of entire study / data set
– Intelligent caching speeds subsequent requests
Stream based transport:
– Low latency access
ILM Policy
# Copies
Storage tier
Radiology
Radiology Application
Applic ation
Storage location
1001010 1001010
0011011 0011011
1000101 1000101
0101010 0101010
Storage
Storage
Figure 4-31 Healthcare - medical image storage example using SONAS ILM
Upon retrieval, the SONAS provides the required capabilities to serve the image, from
wherever it happens to be in the tiered storage hierarchy, including restaging from external
storage, as shown in Figure 4-32.
Retrieval request
Radiology
Radiology Application
Application
CIFS Share
SONAS SONAS
1011 1011
1001010
0101 0101
0011011
1000101
1011 1011
0101010
SATA
0101 0101
Storage
St orage
1TB
SATA
Drives
Figure 4-32 Healthcare - medical image retrieval example using SONAS parallel reads
In the end, the goal of modern healthcare is to provide affordable, optimum patient care, as
well as healthcare organizational performance. SONAS can be a powerful component to
provide the storage foundation for healthcare systems for a smarter planet.
Telcos used to have switching centers staffed by many persons to connect calls and this
process has been replaced by switches that automate traffic to assure service and lower
cost.
Manufacturers used to rely heavily on human labor, but now manufacturers use robotics to
improve quality and lower cost.
Banks used to handle transactions manually and employ large numbers of persons to
perform calculations, now they use automated teller machines to improve service and
lower cost.
Breakthroughs like these are enabled by service management systems that enable whole
industries to become smarter and more efficient. IT needs to become smarter as well, and
cloud computing can bee seen as the natural evolution of IT to its next industrialized stage.
The drivers for cloud services can be summarized in the following points:
First and foremost, service consumers demand lower cost options, cost control is
paramount
As a consequence service consumers want to only pay for what they use
Service consumers want speedy service, they want to get it fast, when they need it.
Last but not least, to enable speed and cost control, they want it to be really easy to
manage.
Cases 4 and 5 differ in the management of access authorizations to the service, case 4 is in a
certain sense wholesale as the service consumer manages authorization and authentication
where case 5 is a retail case where individual users access the shared infrastructure and
authentication and authorization is managed by the service provider.
SONAS is a truly scalable solution, scalability is at the core of the SONAS architecture.
Physical storage capacity can be scaled to the tens of terabytes in a single file system and
front-end nodes can scale out horizontally to offer bandwidths up to tens of gigabytes per
second.
SONAS offers availability and resiliency functions such as snapshots, integrated backups and
replication that enable uninterruptible service delivery with minimum downtime for
housekeeping operations.
First, utilization of our storage cloud systems is greatly increased because all of the storage
capacity is managed as one unit. The data is spread evenly between all devices making
capacity management extremely simple. This makes it very easy for clients to order and then
consume what they really need, even in a private cloud system.
Our storage cloud offering also offers a simplified web user interface which IBM has
enhanced to be fully integrated into our public cloud offering as well for self-service
consumption of the service.
System management tasks such as provisioning, change and release management can all
executed by IBM greatly easing client management load.
Additionally, the storage cloud can also easily feed your chargeback and billing systems with
granular tracking of how assets are being used in the system, allowing you to build your own
internal shared storage clouds.
Current NAS solutions do not scale. File systems are often limited in practice to a few
terabytes and a few million objects, computing power in a single system is miniscule in an
Enterprise context, so you have to add box by box and manage them individually.
Difficult way to apply policies across this independent data islands. Workflow, regulatory
compliance hampered. “Find the file” becomes “find the server with the file, then find the file”.
Massive scalability in multiple dimensions, ability to scale Interface and Storage nodes
separately to suit workload, and greatly simplified administration/management.)
Deep integration of Information Lifecycle Management (ILM) functions into the file system
becomes more important as the amount of TB's explode – scalability and speed critical to
actual execution of an ILM or Tiered Storage strategy. Global namespace, deeply integrated
policy engine, integrated backup and replication of single logical device. Backup windows are
a big issue and get worse while the amount of data continuously increase. Internal distribution
of data often causes “hotspots” and very low total utilization (as low as ~15%)
We see commercial requirements on data access rates and response times to individual files
that have been unique to HPC environments in the past. Logical storage pools of many
physical devices increases utilization, eliminates hotspots and increases performance by
wide striping of data across all devices. Making file servers clustered is easy in theory but
hard to build and maintain, there is much complexity.
Migration, integration or removal of storage for file services is typically difficult. Very proven,
mature IBM core technology coupled to simplified interfaces allowing non-disruptive online
building block addition or removal of individual components, easing growth or technology
upgrades.
IBM has announced the client-premise hosted service called Smart Business Storage Cloud
that can be implemented behind client firewall in managed or un-managed configurations and
as an IBM hosted offering. IBM has also announced the Smart Business Storage on the IBM
Cloud service that offers access by subscription to IBM hosted storage cloud services on IBM
infrastructure based on SONAS. Figure 4-36 illustrates these two offerings.
We have also launched a true public cloud for storage as well, called Smart Business Storage
on the IBM Cloud. Delivered in a public cloud model with a full pay-per-usage pricing model
and all the high degrees of automation, this cloud will also be focused on the needs of our
enterprise clients. We are raising the bar for storage cloud offerings by providing higher
performance, more reliability, and an all-inclusive pricing model that the competition cannot
match.
It is important to understand these services in context of one another as IBM sees the future
of cloud technology for most installations to be in the use of what we call ‘hybrid’ cloud
implementation models. In this model the clients workloads can be even further expanded
through the use of both cloud types in tandem. Inherent technology in the cloud software will
allow us to enable seamless, policy-based movement of data to and from the private and
public cloud spaces.
IBM Smart Business Storage Cloud offers design and implementation services as well as
ongoing basic, premium and fully managed support services, providing you with a single point
of contact for asset, configuration, performance and problem call management. Through
around-the-clock monitoring and reporting of the production cycle, we help you optimize your
data storage environment. Additionally, you also have the option of allowing us to manage
your storage cloud, so you can reduce the need for high skills in network file system and
storage technologies.
A Service consumer customer accessing file services might have their own LDAP or AD
server that contains the profiles of the service consumer customer’s end users. Alternatively,
the service consumer can rely on an AD/LDAP service provided by the service provider.
Currently a SONAS cluster can be connected to a single AD or LDAP server, and so we have
two possible use cases:
In the first case the service consumer or customer uses a central AD/LDAP service
provided by the service provider, the service provider manages and assigns users for
each customer. In this case multiple service consumers can access the same SONAS
cluster.
In the second case the service consumer or user uses and maintains his own AD/LDAP
server, in this case the SONAS cluster can be used by a single service consumer or
customer.
We then saw that large scale, fast growing, modern competitive advantage IT applications,
scale out NAS can often be a highly desirable approach to solving requirements for:
Tremendous growth of “unstructured data” with the same (or less) administrative staff
Need for NAS solutions tot scale in both performance and capacity
Provide parallel access to data to support modern analytics and compute intensive
competitive advantage application
Provide ability to centrally manage hundreds of terabytes to petabytes of data, while
driving high utilization of storage with high performance
Resolve situation where backup windows are too small for 100’s of terabytes
Provide an easy way to centrally apply and enforce policies across all the NAS storage
Provide method to transparently and automatically migrate data as needed
Provide an automated method to globally implement centrally managed, centrally
deployed Automatic Tiered Storage (i.e. Information Lifecycle Management)
IBM SONAS is designed as a scale out NAS appliance, to provide a solution to all of these
issues.
IBM SONAS is designed to provide a cost-effective, highly scalable, centrally managed NAS
storage solution for all of these usage cases. In the remainder of this book, we review the
high-level concepts of IBM Scale Out NAS (SONAS).
After reading this chapter, you will know where to go next to investigate and evaluate IBM
SONAS further for possible use in your IT organization.
This chapter directs you to information about SONAS in other Redbooks publications, to
websites and white papers, and provides information about IBM Services related to SONAS.
This book goes into more detail on IBM SONAS, and discusses topics in detail, such as:
SONAS Hardware architecture
SONAS Software architecture
SONAS networking considerations
SONAS backup and restore, availability, and resiliency
SONAS configuration and sizing
SONAS installation and configuration
SONAS administration, including:
– Security
– Information Lifecycle Management and automated tiered storage
– Asynchronous replication
– Tivoli Storage Manager integration
SONAS migration overview
Getting started with SONAS
You can download this Redbooks publication for more detailed SONAS information.
IBM has created a whole SONAS Eco-System in order to validate the SONAS solution into
existing and relevant ISV environments. SONAS provides File Services and Resource
You might already be committed to an ISV solution in wide domains such as Storage
Management, Data Protection, Virtualization or Data Base environment. And obviously it is a
hard requirement to have business continuity when your SONAS solution is integrated into
your existing infrastructure. You then have to ensure that the SONAS Storage solution is
compatible with your ISV solutions and it will scale to meet your expectations.
The validation of the SONAS solution to be interoperable with middleware such as databases
or security protections, as well as with your business applications, is a key criteria, but
SONAS has also to provide you acceptable performance and throughputs.
This validation process done by IBM with ISVs is aimed to help you to reduce the deployment
effort and risks by delivering you already tested applications. For additional information about
the SONAS Storage solution and compatibility with your ISV solutions, see the website:
https://www-03.ibm.com/systems/storage/solutions/isv/index.html#symantec
Cluster manager
SONAS
HSM and ILM Monitoring Agents
Parallel Software
File System
Backup & Restore GUI/CLI mgmt functions
And Interfaces
Enterprise Linux
IBM Servers
There is a broad variety of existing information available on the Internet about each of these
functions and components, and we list a few of those here.
To learn more about IBM GPFS concepts and capabilities, watch this 21-minute self-playing
website video:
http://www-03.ibm.com/systems/data/flash/clusters/software/gpfs/training/intro/she
ll_popup.html
There are a number of Redbooks publications that have been written on IBM GPFS. You can
search for these books on the Redbooks publications website:
http://www.redbooks.ibm.com/cgi-bin/searchsite.cgi?query=GPFS
An existing Tivoli Storage Manager customer can use their existing Tivoli Storage Manager
Server to perform backups and restores on SONAS, as well as to perform HSM. The SONAS
Software (5639-SN1) includes imbedded Tivoli Storage Manager client code inside the
SONAS (you do not need a separate license for the internal Tivoli Storage Manager client
code inside the SONAS). The following links provide additional information about Tivoli
Storage Manager:
General information about Tivoli Storage Manager:
http://www-01.ibm.com/software/tivoli/products/storage-mgr/
Tivoli Storage Manager manuals online:
http://publib.boulder.ibm.com/infocenter/tsminfo/v6/index.jsp
Tivoli Storage Manager publications on the Redbooks website:
http://www.redbooks.ibm.com/cgi-bin/searchsite.cgi?query=Tivoli+AND+storage+AND
+manager
5.1.5 Education
For general educational information about clustered file systems, a set of web articles on this
topic has been written by the independent consultant “Clustermonkey”. You can view these
articles at:
http://www.clustermonkey.net/
To learn more about IBM Implementation Services for Network Attached Storage systems –
IBM Scale Out Network Attached Storage, contact your IBM representative or IBM Business
Partner, or visit the following website:
http://www-03.ibm.com/systems/storage/network/sonas/
This assessment covers information about the existing storage hardware infrastructure that
can either being direct attached storage (DAS), SAN and NAS storage from major vendors.
The tool allows you to asses Information about installed storage and fabric components,
planned and unplanned outages, personnel costs and maintenance costs.
All the above information, together with the expected growth rates for storage, number of
users and servers is used to generate a SONAS design proposal. This proposal together with
the existing infrastructure data is then used to calculate and compare the cost breakdown in
the following categories: hardware, software, networking, services, facilities, ongoing
personnel, and downtime.
The results, presented both in tables or graphs, together with all input data can be used to
generate an extensive report about the current storage and proposed SONAS environments.
The publications listed in this section are considered particularly suitable for a more detailed
discussion of the topics covered in this book.
Online resources
These websites are also relevant as further information sources:
SONAS Support Site:
http://www.ibm.com/storage/support/
From this website, select:
Product family: Network Attached Storage (NAS)
Product: Scale Out Network Attached Storage
Click Go.
Support for IBM System Storage, TotalStorage and Tivoli Storage products:
http://www.ibm.com/storage/support/
Additional GPFS documentation sources:
http://www.ibm.com/systems/gpfs
http://www-03.ibm.com/systems/software/gpfs/resources.html
NFS V4 ACL information:
http://www.nfsv4.org/
Index 153
154 IBM Scale Out Network Attached Storage Concepts
IBM Scale Out Network Attached
Storage Concepts
IBM Scale Out Network Attached
Storage Concepts
IBM Scale Out Network Attached Storage Concepts
IBM Scale Out Network Attached Storage Concepts
(0.2”spine)
0.17”<->0.473”
90<->249 pages
IBM Scale Out Network Attached
Storage Concepts
IBM Scale Out Network Attached
Storage Concepts
Back cover ®
Introduces IBM Scale IBM Scale Out Network Attached Storage (IBM SONAS) is a Scale Out
NAS offering designed to manage vast repositories of information in INTERNATIONAL
Out NAS concepts
enterprise environments requiring very large capacities, high levels of TECHNICAL
performance, and high availability. SUPPORT
Details hardware and
ORGANIZATION
software architecture SONAS provides a range of reliable, scalable storage solutions for a
variety of storage requirements. These capabilities are achieved by
Provides using network access protocols such as NFS, CIFS, HTTP, and FTP.
configuration
considerations Utilizing built-in RAID technologies, all data is well protected with BUILDING TECHNICAL
options to add additional protection through mirroring, replication, INFORMATION BASED ON
snapshots, and backup. These storage systems are also characterized PRACTICAL EXPERIENCE
by simple management interfaces that make installation,
administration, and troubleshooting uncomplicated and IBM Redbooks are developed
straightforward. by the IBM International
Technical Support
In this IBM Redbooks publication, we discuss the marketplace Organization. Experts from
requirements that led to the SONAS stand-alone storage system. We IBM, Customers and Partners
introduce storage architectures and file systems. We then introduce from around the world create
the reader to SONAS and describe the hardware and software that timely technical information
make up the SONAS appliance.
based on realistic scenarios.
Specific recommendations
are provided to help you
implement IT solutions more
effectively in your
environment.