Isilon Training
Isilon Training
Copyright © 2017 Dell Inc. or its subsidiaries. All Rights Reserved. Dell, EMC, and other trademarks are trademarks of Dell Inc. or
its subsidiaries. Other trademarks may be the property of their respective owners. Published in the USA.
THE INFORMATION IN THIS PUBLICATION IS PROVIDED “AS IS.” DELL EMC MAKES NO REPRESENTATIONS OR
WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION, AND SPECIFICALLY
DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.
Use, copying, and distribution of any DELL EMC software described in this publication requires an applicable software license. The
trademarks, logos, and service marks (collectively "Trademarks") appearing in this publication are the property of DELL EMC
Corporation and other parties. Nothing contained in this publication should be construed as granting any license or right to use any
Trademark without the prior written permission of the party that owns the Trademark.
AccessAnywhere Access Logix, AdvantEdge, AlphaStor, AppSync ApplicationXtender, ArchiveXtender, Atmos, Authentica,
Authentic Problems, Automated Resource Manager, AutoStart, AutoSwap, AVALONidm, Avamar, Aveksa, Bus-Tech, Captiva,
Catalog Solution, C-Clip, Celerra, Celerra Replicator, Centera, CenterStage, CentraStar, EMC CertTracker, CIO Connect,
ClaimPack, ClaimsEditor, Claralert,CLARiiON, ClientPak, CloudArray, Codebook Correlation Technology, Common Information
Model, Compuset, Compute Anywhere, Configuration Intelligence, Configuresoft, Connectrix, Constellation Computing, CoprHD,
EMC ControlCenter, CopyCross, CopyPoint, CX, DataBridge, Data Protection Suite, Data Protection Advisor, DBClassify, DD
Boost, Dantz, DatabaseXtender, Data Domain, Direct Matrix Architecture, DiskXtender, DiskXtender 2000, DLS ECO, Document
Sciences, Documentum, DR Anywhere, DSSD, ECS, elnput, E-Lab, Elastic Cloud Storage, EmailXaminer, EmailXtender, EMC
Centera, EMC ControlCenter, EMC LifeLine, EMCTV, Enginuity, EPFM, eRoom, Event Explorer, FAST, FarPoint, FirstPass,
FLARE, FormWare, Geosynchrony, Global File Virtualization, Graphic Visualization, Greenplum, HighRoad, HomeBase,
Illuminator, InfoArchive, InfoMover, Infoscape, Infra, InputAccel, InputAccel Express, Invista, Ionix, Isilon, ISIS,Kazeon, EMC
LifeLine, Mainframe Appliance for Storage, Mainframe Data Library, Max Retriever, MCx, MediaStor, Metro, MetroPoint,
MirrorView, Mozy, Multi-Band Deduplication, Navisphere, Netstorage, NetWitness, NetWorker, EMC OnCourse, OnRack,
OpenScale, Petrocloud, PixTools, Powerlink, PowerPath, PowerSnap, ProSphere, ProtectEverywhere, ProtectPoint, EMC Proven,
EMC Proven Professional, QuickScan, RAPIDPath, EMC RecoverPoint, Rainfinity, RepliCare, RepliStor, ResourcePak, Retrospect,
RSA, the RSA logo, SafeLine, SAN Advisor, SAN Copy, SAN Manager, ScaleIO Smarts, Silver Trail, EMC Snap, SnapImage,
SnapSure, SnapView, SourceOne, SRDF, EMC Storage Administrator, StorageScope, SupportMate, SymmAPI, SymmEnabler,
Symmetrix, Symmetrix DMX, Symmetrix VMAX, TimeFinder, TwinStrata, UltraFlex, UltraPoint, UltraScale, Unisphere, Universal
Data Consistency, Vblock, VCE. Velocity, Viewlets, ViPR, Virtual Matrix, Virtual Matrix Architecture, Virtual Provisioning, Virtualize
Everything, Compromise Nothing, Virtuent, VMAX, VMAXe, VNX, VNXe, Voyence, VPLEX, VSAM-Assist, VSAM I/O PLUS, VSET,
VSPEX, Watch4net, WebXtender, xPression, xPresso, Xtrem, XtremCache, XtremSF, XtremSW, XtremIO, YottaYotta, Zero-
Friction Enterprise Storage.
In the 1960s, magnetic tape eclipsed punch cards as the way to store corporate computer data. During the
mid-1960s magnetic tape gave way to the hard disk drive. The first hard drive was the size of two
refrigerators, required 50 disks to store less than four megabytes of data, and was called the IBM 3330.
During the 1980s the personal computer revolution introduced miniaturization. This brought a wide array of
storage form factors. Less than 30 years after two refrigerator-sized units stored less than four megabytes,
the average consumer could store about one-third of that amount on a three-and-a-half inch plastic disk.
The evolution of data storage was one of constant progress, improvement, and improved efficiency paired
with increased storage density.
Structured data often resides within a fixed field inside a larger record or file. A large fie often requires a
data model that defines the type of data, such as a numeric or alphanumeric model. In addition to defining
the type of data, access to the data and the processing of the data are important. Therefore, in today’s
data storage world, structured data is most often expressed in a relational database. The rigid table
structure makes structured data easy to query. Spreadsheets, library catalogs, inventory sheets, phone
directories, and customer contact information are all examples of structured data that fit neatly into the
rows and columns of a database.
Unstructured data does not fit into neat rows and columns because it has little or no classification data.
Image files, photographs, graphics files, video, and audio files are all examples of unstructured data.
Imagine you have a spreadsheet with information about your pet dog. The spreadsheet might have the
dog’s name, birthdate, breed, color, weight, parent’s names and information, breeder information, location,
etc. This structured data would be very easy to plug into the predefined field of a database, as the
information deals with classifying individual traits. Now imagine what would happen if you tried to fit a
photograph of your dog into those same fields: it wouldn’t fit. There is no way to classify an image in the
same way that we list out dog’s name, birthdate, breed, or color. According to industry analysts, the
creation rate of unstructured data outpaces structured data, with unstructured data comprising 80 to 90%
of all digital data.
In 1986, the entire world had the technical capacity to store merely 2.6 Exabyte's. By 2020, the world will
need to store more than 40,000 Exabyte's. Much of this growth occurs because a person formerly had to
sit in an office to use a computer, but today, billions of individuals generate data, all day, everywhere they
go, from mobile devices.
Thus, studies document that the world’s data storage needs are not merely growing; they are
mushrooming. Fortunately Isilon specializes in the storage of unstructured data, but we will come back to
that after we look at a few more ways to classify data storage.
A block of data is a sequence of bits or bytes in a fixed length; the length is determined by the file system.
Saving a single piece of data requires the operating system, or OS, to break the file into blocks and each
block is written to a particular sector, or area, of the drive. A single file may require compiling many blocks
together. Block-based data is especially useful when working with small bits of information that need to be
accessed or written frequently. An excellent example of block-based data would be a large database of
postal codes. Someone querying the database probably wants only a small number of the postal codes,
but rarely wants all of them. block-based data makes it easy to gather information in partial sets and is
particularly adept at handling high volumes of small transactions, such as stock trading data. Data
associated with stock trading can generate one billion relatively small files in only a few hours. Block
format is the go-to format when you need flexibility and intensive speed of input and output operations.
File-based data, on the other hand, is created depending upon the application and protocol being used.
Some applications store data as a whole file, which is broken up and sent across the network as packets.
All of the data packets are required to reassemble the file. Unlike block-based data, where you can grab
only one type of postal code, in file-based data storage you need the whole file in order for it to be useful.
For example, a PDF file is generally not readable unless you have all of it accessible; having only part of
the file will generate an error and not allow the file to be opened.
File-based data is organized in chunks too large to work well in a database or in an application that deals
with a high volume of transactions. In IT applications, block-based data usually relates to structured data
while file-based data usually relates to unstructured data. Isilon specializes in handling file-based data.
Can Isilon do block-based storage? Technically, yes, but if you are looking for a block-based solution there
are other Dell EMC products that specialize in block and would best handle that type of workflow.
As applications proliferated, soon there were many servers, each with its own DAS. This worked fine, with some
drawbacks. If one server’s DAS was full while another server’s DAS was half empty, the empty DAS couldn’t share its
space with the full DAS. People thought, “What if we took all these individual storage stacks and put them in one big
stack, then used the network to let all the servers access that one big pool of storage? Then our servers could share
capacity!”
Accomplishing that approach required a traffic cop to keep track of what data went with what application. Thus, the
volume manager was invented. Adding a volume manager to the storage system created the Storage Attached Network
(SAN).
SAN was optimized for block-based data. It worked fine until employers began giving their employees computers.
Employees then needed to get to the stored data, but they couldn’t: SAN was set up for servers, not personal computers
(PCs). PCs worked differently from the storage file server and network communications only communicate from one file
system to another file system. The answer arrived when corporations put employee computers on the network, and
added to the storage a file system to communicate with users. And from this, Network Attached Storage (NAS) was
born.
NAS works pretty well. But it could be improved. For example, now the server is spending as much time servicing
employee requests as it is doing the application work it was meant for. The file system doesn’t know where data is
supposed to go, because that’s the volume manager’s job. The volume manager doesn’t know how the data is
protected; that’s RAID’s job. If high-value data needs more protection than other data, you need to migrate the data to a
different volume that has the protection level that data needs. So there is opportunity to improve NAS.
To alleviate these issues, Isilon combined the file system, the volume manager, and the data protection into one
seamless, self-aware Operating System (OS): OneFS.
Some advantages of this approach include the simplicity of having all data in a single file system and a single volume.
When you have storage capacity without hard architectural limitations, your system is easier to manage and grow.
Isilon was designed to work in a mixed environment. Even if the clients attached to the server are a mix of Windows,
UNIX, and Mac OS X operating systems, Isilon offers a single unified platform for all.
Scale-up NAS came first, represented here with a green line. In this architecture, a pair of controllers or
filers manages a stack of disk trays. You can readily add capacity – if you need more storage, you simply
add more drives. But the architecture doesn’t let you pile on more filers. As disk space grows, computing
resources do not.
In contrast, scale-out NAS, represented here with a blue line, uses nodes. Each node contains drives, but
it also contains more processing and more memory. By adding nodes, performance and capacity scale out
in proportion. The green line shows that over time, the filers must work harder and harder to manage the
growing capacity. Result: performance slows.
The blue line shows that as you add nodes, performance improves, because every node can exploit all the
resources of every other node.
DAS, SAN, and scale-up NAS have their places, but they were invented before the Big Data era. Scale-
out NAS systems were designed for Big Data. Thus, in many regards, the Isilon approach using scale-out
NAS architecture makes managing Big Data less challenging.
There are two types of Isilon clusters currently in the field. There are original Generation 5 nodes which
require a minimum of three nodes to form a cluster. Additionally you will find the Generation 6 clusters. 6th
generation Isilon clusters require a minimum of four nodes, which are contained in one chassis. Let’s
move on and review the different hardware types in a cluster and their positioning.
The Generation 6 clusters start with a minimum of four nodes. Generation 5 and 6 nodes can also exist
within the same cluster. Having both types of nodes in one cluster is the typical path for a hardware
refresh as our customers move toward the Next Generation Isilon brand. Currently, a cluster can have up
to 144 nodes, regardless to which node types are present.
Architecturally, every Isilon node is equal to every other Isilon node of the same type in a cluster. No one
specific node has to be the controller or filer. Instead, OneFS unites the entire cluster in a globally
coherent pool of memory, CPU, and capacity. OneFS writes files in stripes across the nodes for built-in
high availability. So when a file request is received, any available node will communicate the request to
the nodes assembling the file. Then that node will deliver the file requested. Therefore requests are not
processed through one controller node, but rather the node which is most accessible based on availability.
With the sixth generation Isilon architecture, a single cluster can theoretically scale to 1000+ petabytes.
The “availability” possibilities become even greater as the next generation Isilon nodes become a larger
percentage of a cluster.
Starting at the Client/Application layer, the Isilon NAS architecture supports mixed modes. Windows,
UNIX, and OSX operating systems can all connect to an Isilon cluster and access the same files.
At the networking level, the Isilon OneFS operating system supports key industry-standard protocols over
Ethernet, including network file shares, Server Message Block (SMB), HTTP, FTP, Hadoop Distributed
File System (HDFS) for data analytics, SWIFT, and REST for object and cloud computing requirements.
As a file-based storage system, Isilon does not support protocols associated with block-based data.
The next logical level would be the nodes. Nodes are combined into one volume by the OneFS operating
system. All information is shared among nodes, thus allowing a client to connect to any node in the cluster
and access any directory in the file system. And finally on the back end, all the nodes are connected with
either an InfiniBand fabric switch (for Generation 5 and Generation 6 clusters) or with an 40Gb Ethernet
network (Generation 6 clusters only) for low-latency internal communication with one another.
And there you have a slightly deeper look at the levels, or layers, of the scale-out NAS architecture. Next,
let’s take a look at external networking.
The 10 GbE and 40 GbE interfaces support link aggregation. Link aggregation creates a logical interface
that clients connect to. In the event of a Network Interface Controller (NIC) or connection failure, clients do
not lose their connection to the cluster. For stateful protocols, such as SMB and NFSv4, this prevents
client-side timeouts and unintended reconnection to another node in the cluster. Instead, clients maintain
their connection to the logical interface and continue operating normally. Support for Continuous
Availability (CA) for stateful protocols like SMB and NFSv4 is available with OneFS 8.0.
The enhanced functionality includes continuous availability for SMBv3. This feature allows SMB, Network
File System version three (NFSv3), and NFSv4 clients to dynamically move to another node in the event
the node they are connected to goes down.
This feature applies to Microsoft Windows 8, Windows 10 and Windows Server 2012 R2 clients. This
feature is part of Isilon's non-disruptive operation initiative to give customers more options for continuous
work and less down time. The CA option allows seamless movement from one node to another and no
manual intervention on the client side. This enables a continuous workflow from the client side with no
appearance or disruption to their working time. CA supports home directory workflows as well.
In the N+M data model, N represents the number of nodes in the cluster, and M represents the number of
simultaneous hardware failures that the cluster can withstand without incurring data loss. “Hardware
failures” refers to drive loss, node loss, or a combination of the simultaneous loss of drives and/or nodes.
As the system writes the data, it also protects the data with parity bits called Forward Error Correction
(FEC). The OneFS operating system spreads the data across numerous drives in multiple nodes so that if
part of the data goes missing, the missing data can be recalculated and restored. This involves complex
mathematics, but to illustrate the concept, we’ll use a basic example using three colored boxes with some
contents. Let’s assume “5” and “3” be the contents of the blue and yellow boxes, respectively, on the first
row. The sum of these two numbers is “8”, which we are calling a parity bit, and is the contents of the
green box. For here on, once we know the contents of these three colored boxes we can successfully tell
what value inside of a missing box on any of the subsequent rows of boxes, can’t we? That is, on the third
row, knowing the contents of the blue and green boxes as “5” and “8” we can conclude the value of the
missing yellow box to be “3”. In general, the basic idea of the parity bit is a value that OneFS sets to show
what total should result when the binary data is added together.
In another words, with the parity bit the OneFS can easily rebuild the missing data. With the aid of the
parity bit, any one value could vanish, and OneFS could readily recalculate and restore it. Now that we
understand how the N+M data protection model works, let’s look at Isilon’s approach using FEC.
With N + 1 protection, data is 100% available even if a single drive or node fails. With N + 2 protection,
two components can fail, but the data will still be 100% available. OneFS supports up to N+4 – users can
organize their cluster so that as many as four drives, or four entire nodes, can fail without loss of data or of
access to the data.
RAID is disk-based, so when you choose a level of protection – that is, how many parity bits – you’ve
chosen for the entire RAID volume. With Isilon’s FEC approach, you can set different levels of protection
for different nodes, directories, or even different files. Also you can change protection levels on the fly,
non-disruptively. Unlike RAID where you have the same protection level across all the disks and this
cannot be changed without reformatting the disks.
When a client connects though a single node and saves data, the write operation occurs across multiple
nodes in the cluster. This is also true for read operations. When a client connects to a node and requests
a file from the cluster, the node to which the client has connected uses the backend InfiniBand or 40GbE
network to coordinate with other nodes to retrieve, rebuild, and deliver the file back to the client.
The process of striping spreads all write operations from a client across the nodes of a cluster. The
example in this animation demonstrates how a file is broken down into chunks, after which it is striped
across disks in the cluster along with parity, also known as FEC.
Though a client connects to only one node, when that client saves data to the cluster, the write operation
occurs in multiple nodes in the cluster. Each node contains between 12 and 59 hard disk drives, or a
combination of SSDs and disk drives. As the system lays out data across the cluster, it distributes stripe
units across nodes, and each node in turn may utilize numerous drives, if the file is large enough. This
method minimizes the role of any specific drive or node. If one piece of hardware stops working, the data it
contains can be reconstructed.
The node with the requesting client then uses the backend InfiniBand or 40GbE to communicate with other
nodes to retrieve, rebuild, and deliver the file.
The density of the drives in an Generation 6 chassis are notably heavier than the previous Generation 5.
Therefore extreme care and every safety precaution should be taken when rack mounting or working with
the nodes. Having reviewed the Generation 6 hardware design, let’s take a look at the Generation 5
hardware design.
The Generation 5 portfolio consists of four storage node series and two non-storage nodes. The storage
nodes series are the S-Series, X-Series, NL-Series, and the HD-Series. Each of these are pictured here.
Not pictured are the non-storage: the A100 accelerator and the backup accelerator. The S-Series is for
high-performance primary storage and is designed for high-transactional and IO-intensive tier one
workflows. The X-Series strikes a balance between large capacity and high-performance storage. X-
Series nodes are best for high-throughput and high-concurrency tier two workflows and also for larger files
with fewer users. The NL-Series is designed to provide a cost-effective solution for tier three workflows,
such as nearline storage and data archiving. NL nodes are ideal for nearline archiving and for disk-based
backups. The HD-Series is a high-density, deep archival platform. This platform is used for archival level
data that must be retained for long, if not indefinite, periods of time but is rarely accessed. Moving forward,
the Generation 5 nodes will be joined by the Generation 6 nodes. Let’s review those now.
The video linked here was shot by an enthused Isilon customer who posted it on YouTube with the
comment that adding a node was “insanely fast”. It is not especially exciting to watch, until you realize that
accomplishing the same tasks with another NAS solution takes 26 steps and multiple hours.
If you look at the free space available in the pie chart, this Isilon customer took his system from 280
terabytes to 403 terabytes, adding 120 terabytes of storage in a minute.
The first is Data At Rest Encryption (DARE). The term “data at rest” refers to any data sitting on your
drives. DARE is used for confidential or sensitive information.
Isilon’s security protocols are top notch, and though that is not the topic of discussion today for the sake of
this example let’s assume the following scenario. Somehow a hostile party infiltrated your network, they
were able to access your Isilon cluster, and somehow acquired the various levels of permissions/access to
see the data striped on the clusters. With DARE these hypothetical super hackers still could not read the
data, because it is encrypted. So even assuming this almost-impossible, worst case scenario, no
information would be compromised. DARE is not just an excellent security protocol. DARE is, in fact,
required by many vertical markets. For example around the globe federal governments, financial services,
and HIPAA-compliant health care providers all must encrypt stored data. This is where a feature, such as
DARE, is obviously valuable.
A less obvious benefit of DARE occurs when it’s time to upgrade your hardware. If you run a corporation in
one of the previously mentioned regulated industries and you need to dispose of an old drive, you may
have a problem. That is, anytime data is left on a bad, or good, disk drive there is the potential that the
data can be retrieved. No matter how many erase programs are used, nor how many times it is run, such
programs do not literally delete the data. They only mark the sector of the drive that holds the data as
available for overwrite. Thus, the data still remains on the drive and it can potentially be recovered with
hacker tools. For that reason, a whole industry has sprung up around physically destroying retired
hardware. DARE provides an easier solution. DARE ensures all data at rest has been encrypted. If the
data has been encrypted with a 256-bit key for all its life, you can recycle the drive as is, without fear that
anyone can read it. Isilon implements DARE by offering optional Self-Encrypting Drives, or SEDs.
As the system writes the data, it also protects the data. Previously we reviewed N + M protection and
parity bits. The technical way to describe that kind of fault tolerance protection level is to say that OneFS
uses Reed Solomon FEC algorithms. OneFS also uses Flexprotect. FlexProtect creates an n-way,
redundant fabric that scales as nodes are added to the cluster, providing 100% data availability even with
four simultaneous node failures. Flexprotect utilizes advanced technology to provide redundancy and
availability capabilities far beyond those of traditional RAID. Unlike scale-up NAS, in scale-out NAS there
is not a singular master node or device that controls the cluster. Each node is a peer that shares the
management workload and acts independently as a point of access for incoming data requests. That way,
bottlenecking does not occur when there are a large influx of simultaneous requests. Thus, there is a copy
of OneFS on every node in the cluster. This approach prevents downtime since every node can take over
activities for another node if it happens to go offline. Within the cluster, OneFS coordinates all the nodes
on the back end, across either the InfiniBand or 40GbE networks. Now that we know how to store data
using OneFS, let’s look at how OneFS benefits this storage approach through optimization.
To enable multiprotocol file access, Isilon translates Windows Security Identifiers (SIDS) and UNIX User
Identities (UIDs) into a common identity format. OneFS stores these identities on the cluster, tracking all
the user IDs from the various authentication sources. OneFS also stores the appropriate permissions for
each identity or group. We call this common identity format stored on the cluster the “on-disk
representation” of users and groups.
For instance, the SMB protocol exclusively uses SIDs for authorization data. If a user needs to retrieve a
file for a Windows client, as OneFS starts to retrieve the file, it converts the on-disk identity into a
Windows-friendly SID and checks the permissions. Or, if the user is saving a file, OneFS would do the
same kind of translation, from the on-disk representation to SIDs, before saving the file. This works the
same way on the UNIX side using UIDs and Globally Unique Identifiers (GUIDs) instead of SIDs. And
that’s how all users can access OneFS files in a mixed-platform client environment.
OneFS supports four methods for authenticating users: AD, Lightweight Directory Access Protocol
(LDAP), Network Information Service (NIS), and local/file provider accounts on the cluster.
You are likely already familiar with AD. While Active Directory can serve many functions, the primary
reason for joining the cluster to an AD domain is to let the AD domain controller perform user and group
authentication. Each node in the cluster shares the same AD computer account, making it easy to
administer and manage.
You probably know LDAP, too. A primary advantage of LDAP is the open nature of its directory services
and the ability to use LDAP across many platforms. OneFS can use LDAP to authenticate user and group
access to the cluster.
NIS is Sun Microsystem’s directory access protocol. To avoid confusion, please note NIS differs from
NIS+, which the Isilon cluster does not support.
Isilon supports local user and group authentication using the web administration interface which fall under
the category of local or file provider authentication.
Enabling multiple authentication methods on a cluster is possible with OneFS. However it is important to
test how the different methods interact. A word of caution: work methodically, it is quite easy to get
confused about which authentication is authorizing who.
Policies drive every process in OneFS. That includes the way data is distributed across the cluster and on
each node. It also includes how client connections get distributed among the nodes, when and how
maintenance tasks are executed. Policies are very important to the functionality of OneFS because they
enable so many automated activities.
First, you can manually upload log files to the Dell EMC support FTP site. The log files provide detailed
information about the cluster activities when a client needs technical support. Dell EMC’s support
personnel request these files at the beginning of a support call.
The second option is EMC Secure Remote Services (ESRS) which monitors your Isilon cluster, and with
your permission, allows remote access to Dell EMC Technical Support personnel to gather cluster data
and troubleshoot issues. ESRS is a secure, IP-based customer service support system. ESRS features
include 24x7 remote monitoring and secure authentication with AES 256-bit encryption and RSA digital
certificates. When configured, ESRS monitors your Isilon cluster on a node-by-node basis and sends
alerts regarding the health of your devices. Dell EMC Technical Support personnel can establish remote
sessions through SSH or the web administration interface. During remote sessions, support personnel can
run scripts that gather diagnostic data about cluster settings and operations, which is sent to a secure FTP
site, and troubleshoot open support cases on your cluster.
Several policies determine how the data distribution occurs. The default policy is for data to write
anywhere in the cluster. Data distributes among the different node pools based on highest percentage of
available space. So the data target can be pool or a tier anywhere on the cluster. If the distribution of the
data needs to be more specific, the customer should license SmartPools. If SmartPools is licensed, an
individual node pool can be the target for the data distribution.
The Pre-fetch, or read ahead, is an optimization algorithm that attempts to predict what data will be
needed next, before the request is made. When clients open larger files, especially streaming formats like
video and audio, the cluster assumes that you will generally watch minute four of that video after minute
three. So it proactively loads minutes four, five, and maybe even six into memory ahead of when it is
requested. Because of this function, delivering those minutes will be faster than if the cluster had to go to
the hard drive repeatedly for each request. With OneFS you can configure the pre-fetch cache
characteristics to work best with the selected access pattern.
The system is enabled, by default, for virtual hot spares (VHS). VHS enables you to allocate disk space
that will hold the data as it's rebuilt when a disk drive fails.
We mentioned FEC, earlier in this course, and how OneFS uses FEC for data protection. Well, this is
different than RAID. RAID works at the disk level – once you’ve chosen a RAID type, that whole RAID
volume can only be that type of RAID; and if you want to change the RAID type, you’d have to move all the
data off the RAID disks before you can reformat.
Because OneFS uses FEC for protection, you can set the protection level differently based on tier, node
pool, directory, and even by the individual file. Extra protection creates extra overhead because OneFS
writes more FEC stripes. You decide how to trade off extra capacity (meaning, less protection) with
greater redundancy (meaning, less capacity). Based on the value of the data, you can set the protection
level differently.
For example, a Research and Development department has a node pool dedicated to testing. Because the
test data is not that valuable because it's not production data, they’ve set minimal N+1 protection. Their
customer database, however, is the company’s most valuable asset. Customer data is written to a
different node pool set and to a higher level of protection ‒ N+4. Protection is flexible with OneFS so they
could even set up to eight mirrors of each file and while this is not space efficient, for very frequently read
files it can really speed things up.
The standard functionality is available in the unlicensed version of SmartPools, which is sometimes
referred to as SmartPools Basic. If you have the license for SmartPools, you get enhanced capabilities.
Quotas let you implement thin provisioning. For example, let us say you advise group A, “you may use up
to one terabyte of storage”. Group A will not instantly utilize the full terabyte of storage you have allocated
to them. They may never use it all . But with quota-based thin provisioning, you can keep showing the
group an available terabyte of storage, even if you don’t have a full terabyte actually available on the
cluster currently.
OneFS has three primary types of quotas: accounting or advisory quotas; plus two levels of enforcement
quotas, soft limit and hard limit. Advisory quotas are informational only. If a user exceeds their advisory
storage quota, OneFS lets them; but the cluster provides a comparison between the quota allocation and
actual usage. In contrast, if a user exceeds a soft limit quota, the system notifies the user with a warning
email. If the user exceeds a hard limit quota, the system will from then on deny the user the ability to write.
It also notifies the user that the quota has been violated.
You can customize the quota notifications in OneFS so that they meet your requirements.
Quotas are enhanced functionality that requires licensing. To get the feature, you must purchase a
SmartQuotas license for each node in the cluster.
You can run deduplication jobs against a specific directory path or on the entire directory structure.
OneFS also provides a dry run deduplication assessment tool without a need for a license. This allows you
to test drive deduplication to see how much capacity you would save if you ran the actual deduplication
process. To enable the deduplication full functionality, you need a SmartDedupe license for each node in
the cluster.
InsightIQ monitors many aspects of system performance, such as CPU utilization and interface
throughput. The tool also reports on the file system analytics including quota usage, files per user, files per
directory, average file size, and more.
An external VMware system or standalone Linux server is required to use InsightIQ. The separate server
runs external to the cluster and collects data from the cluster in scheduled intervals. To enable these
capabilities, you get a free InsightIQ license for each cluster.
Each stripe is protected separately with FEC blocks, or parity. Stripes are spread across the nodes and
not contained in a single node. Only one or two data or protection stripe units are contained on a single
node for any given data stripe. Protecting at this granular level allows you to vary your protection levels
and set them separately for node pools, directories, or even individual files.
What’s the point of all this? Well, in most popular SANs and in typical scale-up NAS, you have a pair of
heads so that each one can back up the other, and that’s what provides high availability. With OneFS, you
could say that high availability is baked right in to every data transaction, because the data is spread onto
many drives and multiple nodes, all of them ready to pitch in and help reassemble the data if a component
fails. This approach creates an amazingly resilient platform.
OneFS uses the copy-on-write or redirect-on-write snapshot methodology depending on the size of the
change. This is an automated action that is determined by OneFS. The COW snapshot approach keeps
the live version of data intact while storing differences in a snapshot, while ROW will write changes into
available space and then update pointers to look at the new changes. Because the system is only writing
changes, the writes are very fast.
Snapshot policies are used to determine the snapshot schedule, the path to the snapshot location, and
snapshot retention periods. Snapshot deletions happen as part of a scheduled job, or you can also delete
them manually. And although you can delete snapshots out of chronological order it is not recommended
and not a best practice.
Some OneFS system processes use snapshots internally. No license is required for system-based
snapshot usage. However, to use snapshots for data resiliency requires a SnapshotIQ license for each
node in the cluster.
A backup application external to the cluster manages the backup process. You can set this up in one of
two ways:
• For Generation 5 and Generation 6 platforms, send cluster data over your LAN to the backup
device
• For Generation 5 platform only, send data directly from the cluster to the backup device using
Isilon backup accelerators
Depending upon the amount of data and the interfaces selected on the external network, backing up
across a network might not be as efficient as using the backup accelerator. The backup accelerator
provides access to the data across the fast InfiniBand internal network and delivers it to the backup device
over Fibre Channel ports. NDMP support comes standard with OneFS.
With copy, any new files on the source are copied over to the target, while files that have been deleted on
the source remain unchanged on the target.
With synchronization, both the source and target clusters maintain identical file sets, except that files on
the target are read-only.
You can set replication policies to run synchronization jobs at specified times or you can replicate
automatically if the source data changes. The policies can be set up per directory or for specific data
types, and you can also configure exceptions to include or exclude specific files.
OneFS also empowers you to limit the bandwidth used for replication, in order to optimize the traffic for
more important workflows.
Enterprise is more flexible than Compliance, and meets most companies’ retention requirements. It can
allow privileged deletes by an administrator.
Compliance level of retention is even more secure, designed to meet SEC regulatory requirements. In
Compliance mode, once data is committed to disk, no one can change or delete the data until the retention
clock expires. A common hacker ploy for beating retention safeguards is to temporarily change the system
clock to some date way in the future, thus releasing all files. Compliance mode defeats this approach by
relying upon a specialized clock that prohibits clock changes.
You can still use SyncIQ to replicate the files that have retention policies applied.
The storage industry says that Big Data is digital data having too much volume, velocity, or variety to be
stored traditionally. To make sure the three V’s of Big Data are perfectly clear, let’s consider some
examples.
What’s an example of velocity? Machine-generated workflows produce massive volumes of data. For
example, the longest stage of designing a computer chip is physical verification, where the chip design is
tested in every way to see not only if it works, but also if it works fast enough. Each time researchers fire
up a test on a graphics chip prototype, sensors generate many terabytes of data per second. Storing
terabytes of data in seconds is an example of Big Data velocity.
Perhaps the best example of variety is the world’s migration to social media. On a platform such as
Facebook, people post all kinds of file formats: text, photos, video, polls, and more. According to a CNET
article from June 2012, Facebook was taking in more than 500 terabytes of data per day, including 2.7
billion Likes and 300 million photos. Every day. That many kinds of data at that scale represents Big Data
variety.
The “Three Vs” – volume, velocity, and variety – often arrive together. When they combine, administrators
truly feel the need for high performance, higher capacity storage. The three Vs generate the challenges of
managing Big Data.
Growing data has also forced an evolution in storage architecture over the years due to the amount of data
that needs to be maintained- sometimes for years or beyond. Isilon is a Big Data solution because it can
handle the volume, velocity, and variety that defines the fundamentals of Big Data. These topics will be
addressed as the course continues.
Challenge: Complex data architecture. SAN and scale-up NAS data storage architectures encounter a
logical limit at 16 terabytes, meaning, no matter what volume of data arrives, a storage administrator has
to subdivide it into partitions smaller than 16 terabytes. This is part of why customers wind up with silos of
data. To simplify this challenge, scale-out NAS such as an Isilon cluster holds everything in one single
volume with one LUN. Isilon is like one gigantic bucket for your data, and really can scale seamlessly
without architectural hard stops forcing subdivisions on the data.
Challenge: Low utilization of raw capacity. SAN and scale-up NAS architectures must reserve much of
the raw capacity of the system for management and administrative overhead, such as RAID parity disks,
metadata for all those LUNs and mega-LUNs, duplicate copies of the file system, and so on. As a result,
conventional SAN and NAS architectures often use only half of the raw capacity available, because you
have to leave headroom on each separate stack of storage. Suppose you have seven different silos of
data. As soon as you put them in one big volume, you immediately get back the headroom from six of the
seven stacks. In that way, Isilon offers high utilization. Isilon customers routinely use 80% or more of raw
disk capacity.
Challenge: Non-flexible data protection. When you have Big Data volumes of information to store, it
had better be there, dependably. If an organization relies on RAID to protect against data loss or
corruption, the failure of a single disk drive causes disproportionate inconvenience. The most popular
RAID implementation scheme allows the failure of only two drives before data loss. (A sizable Big Data
installation will easily have more than 1000 individual hard drives, so odds are at least one drive is down at
any given time). The simpler answer is to protect data using a different scheme. Shortly you’ll learn about
Isilon’s clustered architecture based on nodes that do not use RAID. Nodes full of hard drives rely less on
any single drive and can recover a failed drive as a non-emergency.
Challenge: Difficult to scale performance. Some data storage architectures use two controllers, sometimes
referred to as servers or filers, to run a stack of many hard drives. You can scale capacity by adding more hard
drives, but it’s difficult to scale performance. In a given storage stack, the hard drives offer nothing but capacity - all
the intelligence of the system, including computer processing and RAM, must come from the two filers. If the
horsepower of the two filers becomes insufficient, the architecture does not allow you to pile on more filers. You have
to start over with another stack and two more filers. In contrast, every node in an Isilon cluster contains capacity
PLUS compute power PLUS memory. The nodes can work in parallel, so each node you add scales out linearly – in
other words, all aspects of the cluster scale up, including capacity and performance.
Challenge: Silos of data. Due to the architectural restrictions we just discussed, SAN and scale-up NAS end up with
several isolated stacks of storage. Many customer sites have a different storage stack for each application or
department. If the R&D stack performs product testing that generates results at Big Data velocity, the company may
establish an HPC stack, which could reach capacity rapidly. Other departments or workflows may have independent
storage stacks that have lots of capacity left, but there’s no automated way for R&D to offload their HPC overflow to,
for example, a backup storage stack. Instead, an administrator has to manually arrange a data migration. In contrast,
an Isilon cluster distributes data across all its nodes to keep them all at equal capacity. You don’t have one node
taking a pounding while other nodes sit idle. There are no hot spots, and thus, no manual data migrations. Automated
balancing makes much more sense if the goal is to keep pace with Big Data velocity.
Challenge: Concurrency. In conventional storage, a file is typically confined to a RAID stripe. That means that the
maximum throughput of reading that file is limited to how fast those drives can deliver the file. But in modern
workflows, you may have a hundred engineers or a thousand digital artists all needing access to a file, and those
RAID drives can’t keep up. Perhaps the two filers on that stack can’t process that many requests efficiently. Isilon’s
answer is that every node has at least a dozen drives, plus more RAM and more computer processing, for more
caching and better concurrent access. When there is heavy demand for a file, several nodes can deliver it.
Challenge: Many manual processes. Besides manual data migrations, conventional storage has many more
manual processes. An administrator over a SAN or a scale-up NAS product spends a significant amount of time
creating and managing LUNs, partitioning storage, establishing mounts, launching jobs, and so on. In contrast, Isilon
is policy-driven. Once you define your policies, the cluster does the rest automatically.
Isilon was purpose-built to ease the challenges of processing, storing, managing, and delivering data at
scale. Isilon’s positioning is to provide simple yet powerful answers for Big Data storage administrators.
CloudPools lets your customers address rapid data growth and optimize data center storage resources by
using the cloud as a highly economical storage tier with massive storage capacity for cold or frozen data
that is rarely used or accessed. This enables more valuable on premise storage resources to be used for
more active data and applications. To secure data that is archived in the cloud, CloudPools encrypts data
that is transmitted from the Isilon cluster at the core data center to the cloud storage service. This data
remains encrypted in the cloud until it is retrieved and returned to the Isilon cluster at the data center.
With built-in multi-protocol capabilities, Isilon can support a wide range of traditional and next-generation
applications on a single platform, including powerful Big Data analytics that provide you with better insight
and use of your stored information.
Data at edge locations, such as remote or branch offices, are growing. These edge locations are often
inefficient islands of storage, running with limited IT resources and inconsistent data protection practices.
Data at the edge generally lives outside of the business Data Lake, making it difficult to incorporate into
data analytics projects. The edge-to-core-to-cloud approach can extend your Isilon Data Lake to edge
locations and out into the cloud, thus enabling consolidation, protection, management, and backups of
remote edge location data.
PARTICIPANT GUIDE
PARTICIPANT GUIDE
Internal Use - Confidential
PowerScale Concepts-SSP
Course Objectives...................................................................................................... 8
Course Objectives................................................................................................................ 9
PowerScale Concepts-SSP
PowerScale Concepts-SSP
PowerScale Concepts-SSP
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 1
Prerequisite Skills
Prerequisite Skills
PowerScale Concepts-SSP
Prerequisite Skills
To understand the content and successfully complete this course, a student must
have a suitable knowledge base or skill set. The student must have an
understanding of:
• Networking fundamentals such as TCP/IP, DNS and routing
• An introduction to storage such as NAS and SAN differences and basic storage
principles and features
• Installation process of an PowerScale cluster
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 3
Rebranding - Isilon is now PowerScale
PowerScale Concepts-SSP
PowerScale Concepts-SSP
PowerScale Concepts-SSP
The graphic shows the PowerScale Solutions Expert certification track. You can
leverage the Dell Technologies Proven Professional program to realize your full
potential. A combination of technology-focused and role-based training and exams
to cover concepts and principles as well as the full range of Dell Technologies'
hardware, software, and solutions. You can accelerate your career and your
organization’s capabilities.
PowerScale Solutions
(C) - Classroom
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 7
Course Objectives
Course Objectives
PowerScale Concepts-SSP
Course Objectives
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 9
Data Storage Overview
PowerScale Concepts-SSP
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 11
Data Storage Overview
Module Objectives
PowerScale Concepts-SSP
Storage Evolution
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 13
Data Storage Overview
During the data storage evolution, two types of data developed: structured data and
unstructured data. PowerScale specializes in storing unstructured data.
• Resides in fixed field of records or files • Does not reside in fixed model
PowerScale Concepts-SSP
Block-based data
• Sequence of bytes at fixed length
• Single piece of file or whole file
File-based data
• Discrete unit of information defined by application or created by user
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 15
Data Storage Overview
Digital Transformation
IDC projects, that through 2022, 75% of successful digital strategies will be built by
a transformed IT organization, with modernized and rationalized infrastructure,
applications, and data architectures.
[...] within the next four years, the global economy will finally reach digital
supremacy, with more than half of Gross domestic product (GDP)Gross domestic
product (GDP) is a monetary measure of the market value of all the final goods and
services produced in a specific time period. - IDC FutureScape1
At the same time, many organizations still struggle to tactically apply DX learnings
to their own business.
PowerScale Concepts-SSP
With unstructured data being the majority of data storage growth, a solution was
needed. An International Data Corporation (IDC) study published in 2018 showed
that the amount of digital data created, captured, and replicated worldwide grew
exponentially. This finding was based on the proliferation of then-new technologies
such as Voice over IP, RFID, smartphones, and consumer use of GPS. Also, the
continuance of data generators such as digital cameras, HD TV broadcasts, digital
games, ATMs, email, videoconferencing, medical imaging, and so on.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 17
Data Storage Overview
PowerScale clusters are a NAS solution. There are two types of NAS architectures;
scale-up and scale-out.
Scale-Up
3The two controllers can run active/active or active-passive. For more capacity,
add another disk array. Each of these components is added individually. As more
systems are added, NAS sprawl becomes an issue.
PowerScale Concepts-SSP
Controller with
disk shelves
Independent systems on
network - separate points
of management
Client
s
Structured or
Unstructured storage
Scale-Out
• With a clustered NAS solutions, or scale-out architecture, all the NAS boxes, or
PowerScale nodes, belong to a unified cluster with a single point of
management.
• In a scale-out solution4, the computational throughput, disks, disk protection,
and management are combined and exist for a single cluster.
4Not all clustered NAS solutions are the same. Some vendors overlay a
management interface across multiple independent NAS boxes. This gives a
unified management interface, but does not unify the file system. While this
approach does ease the management overhead of traditional NAS, it still does not
scale well.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 19
Data Storage Overview
Unstructured storage
1000+ PBS
Clients
PowerScale Concepts-SSP
Scale-Out NAS
Scale-out NAS5 is now a mainstay in most data center environments. The next
wave of scale-out NAS innovation has enterprises embracing the value6 of NAS
and adopting it as the core of their infrastructure.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 21
Data Storage Overview
PowerScale Concepts-SSP
DAS
SAN
SAN
As applications proliferated, soon there were many servers, each with its own DAS.
This worked fine, with some drawbacks. If one server’s DAS was full while another
server’s DAS was half empty, the empty DAS couldn’t share its space with the full
DAS. Due to this limitation with DAS, SAN was introduced which effectively utilized
volume manager and RAID.
NAS
NAS
SAN was set up for servers, not personal computers (PCs). PCs worked differently
from the storage file server and the network communications in PCs, only
communicate from one file system to another file system. The breakthrough came
when corporations put employee computers on the network, and added to the
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 23
Data Storage Overview
storage a file system to communicate with users. From this, Network Attached
Storage (NAS) was born.
NAS works pretty well, but there is room for improvement. For example, the server
is spending as much time servicing employee requests as it is doing the application
work it was meant for. The file system doesn’t know where data is supposed to go,
because that’s the volume manager’s job. The volume manager doesn’t know how
the data is protected; that’s RAID’s job. If high-value data needs more protection
than other data, you need to migrate the data to a different volume that has the
protection level that data needs. So there is opportunity to improve NAS.
PowerScale Concepts-SSP
With traditional NAS systems the file system7, volume manager8, and the
implementation of RAID9 are all separate entities.
OneFS is the operating system and the underlying file system that drives and
stores data. OneFS is a single file system that performs the duties of the volume
manager and applies protection.
• Creates a single file system for the cluster.10
• Volume manager and protection.11
7The file system is responsible for the higher-level functions of authentication and
authorization.
10As nodes are added, the file system grows dynamically and content is
redistributed.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 25
Data Storage Overview
12Because all information is shared among nodes, the entire file system is
accessible by clients connecting to any node in the cluster.
13Each PowerScale storage node contains globally coherent RAM, meaning that,
as a cluster becomes larger, it also becomes faster. When a node is added, the
performance scales linearly.
PowerScale Concepts-SSP
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 27
PowerScale Physical Architecture
PowerScale Concepts-SSP
Module Objectives
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 29
PowerScale Physical Architecture
14The Gen 6 platform reduces the data center rack footprints with support for four
nodes in a single 4U chassis. It enables enterprise to take on new and more
demanding unstructured data applications. The Gen 6 can store, manage, and
protect massively large datasets with ease. With the Gen 6, enterprises can gain
new levels of efficiency and achieve faster business outcomes.
15 The ideal use cases for Gen 6.5 (F200 and F600) is remote office/back office,
factory floors, IoT, and retail. Gen 6.5 also targets smaller companies in the core
verticals, and partner solutions, including OEM. The key advantages are low entry
price points and the flexibility to add nodes individually, as opposed to a chassis/2
node minimum for Gen 6.
PowerScale Concepts-SSP
Network: There are two types of networks that are associated with a cluster:
internal and external.
Ethernet
Clients connect to the cluster using Ethernet connections17 that are available on all
nodes.
16In general, keeping the network configuration simple provides the best results
with the lowest amount of administrative overhead. OneFS offers network
provisioning rules to automate the configuration of additional nodes as clusters
grow.
17Because each node provides its own Ethernet ports, the amount of network
bandwidth available to the cluster scales linearly.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 31
PowerScale Physical Architecture
OneFS supports a single cluster18 on the internal network. This back-end network,
which is configured with redundant switches for high availability, acts as the
backplane for the cluster.19
19 This enables each node to act as a contributor in the cluster and isolating node-
to-node communication to a private, high-speed, low-latency network. This back-
end network utilizes Internet Protocol (IP) for node-to-node communication.
PowerScale Concepts-SSP
The external network provides connectivity for clients over standard file-based
protocols. It supports link aggregation, and network scalability is provided through
software in OneFS. A Gen 6 node has to 2 front-end ports - 10 GigE, 25 GigE, or
40 GigE, and one 1 GigE port for management. Gen 6.5 nodes have 2 front-end
ports - 10 GigE, 25 GigE, or 100 GigE. In the event of a Network Interface
Controller (NIC) or connection failure, clients do not lose their connection to the
cluster. For stateful protocols, such as SMB and NFSv4, this prevents client-side
timeouts and unintended reconnection to another node in the cluster. Instead,
clients maintain their connection to the logical interface and continue operating
normally. Support for Continuous Availability (CA) for stateful protocols like SMB
and NFSv4 is available with OneFS 8.0.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 33
PowerScale Physical Architecture
Back-end Network
InfiniBand
Ethernet
PowerScale Concepts-SSP
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 35
PowerScale Physical Architecture
The clients can access the cluster using DNS, and the enhanced functionality20
provides connection distribution policies as shown in the graphic. Also, they provide
continuous availability21 (CA) capabilities.
1 4
2
3
PowerScale Concepts-SSP
2: Determines the average CPU utilization on each available network interface and
selects the network interface with lightest processor usage.
3: Selects the next available network interface on a rotating basis. This selection is
the default method. Without a SmartConnect license for advanced settings, this is
the only method available for load balancing.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 37
PowerScale Physical Architecture
N + M Data Protection
OneFS sets parity bits, also called FEC protection. In the example below, using the
parity bit (green), OneFS determines the missing pieces.
Here, if blue + yellow = green, the missing pieces are identified using the parity
bits.
Parity bit
Then
PowerScale Concepts-SSP
FEC enables the customer to choose the number of bits of parity to implement.
One bit of parity for many disks is known as N+1; two parity points for many disks
are known as N+2, and so on.
With the N+1 protection, data is 100% available even if a drive or a node fails.
Failure
With N+2, N+3, and N+4 protection, data is 100% available if multiple drives or
nodes fail.
Failure Failure
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 39
PowerScale Physical Architecture
During the write operation, with OneFS, the file from the client is striped across the
nodes. The system breaks the file-based data into smaller logical sections called
stripe units. The smallest element in a stripe unit is 8 kilobytes and each stripe unit
is 128 kilobytes, or sixteen 8 kilobytes blocks. If the datafile is larger than 128
kilobytes, the next part of the file is written to a second node. If the file is larger than
256 kilobytes, the third part is written to a third node, and so on. The graphic
illustrates a 384-kilobytes file with 3 stripe units and 1 FEC unit.
File
Stripe Unit
FEC
Node 3 Node 4
Node 1 Node 2
Leaf
Leaf
Spine
PowerScale Concepts-SSP
PowerScale Nodes
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 41
PowerScale Nodes
PowerScale Nodes
PowerScale Concepts-SSP
Module Objectives
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 43
PowerScale Nodes
The design goal for the PowerScale nodes are to keep the simple ideology of NAS,
provide the agility of the cloud, and the cost of commodity.
The Gen 6x family has different offerings that are based on the need for
performance and capacity. As Gen 6 is a modular architecture, you can scale out
compute and capacity separately. All the nodes are powered by OneFS.
PowerScale Concepts-SSP
PowerScale Family
Click each tab to learn about the different offerings that Gen 6 family provides.
F-Series
• F80022
• F81023
• F60024
• F20025
22 The F800 is suitable for workflows that require extreme performance and
efficiency.
23 The F810 is suitable for workflows that require extreme performance and
efficiency. The F810 also provides high-speed inline data deduplication and in-line
data compression. It delivers up to 3:1 efficiency, depending on your specific
dataset and workload.
24 Ideal for small, remote clusters with exceptional system performance for small
office/remote office technical workloads.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 45
PowerScale Nodes
H-Series
After F-series nodes, next in terms of computing power are the H-series nodes.
These are hybrid storage platforms that are highly flexible and strike a balance
between large capacity and high-performance storage to provide support for a
broad range of enterprise file workloads.
• H40026
• H50027
• H560028
• H60029
25Ideal for low cost all-flash node pool for existing Gen6 clusters. Ideal for small,
remote clusters.
27The H500 is a versatile hybrid platform that delivers up to 5 GB/s bandwidth per
chassis with a capacity ranging from 120 TB to 720 TB per chassis. It is an ideal
choice for organizations looking to consolidate and support a broad range of file
workloads on a single platform.
28The H5600 combines massive scalability – 960 TB per chassis and up to 8 GB/s
bandwidth in an efficient, highly dense, deep 4U chassis. The H5600 delivers inline
data compression and deduplication. It is designed to support a wide range of
demanding, large-scale file applications and workloads.
PowerScale Concepts-SSP
A-Series
The A-series nodes namely have lesser compute power compared to other nodes
and are designed for data archival purposes. The archive platforms can be
combined with new or existing all-flash and hybrid storage systems into a single
cluster that provides an efficient tiered storage solution.
• A20030
• A200031
30The A200 is an ideal active archive storage solution that combines near-primary
accessibility, value and ease of use.
31The A2000 is an ideal solution for high density, deep archive storage that
safeguards data efficiently for long-term retention.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 47
PowerScale Nodes
Gen 6 requires a minimum of four nodes to form a cluster. You must add nodes to
the cluster in pairs.
The chassis holds four compute nodes and 20 drive sled slots.
Both compute modules in a node pair power-on immediately when one of the
nodes is connected to a power source.
Gen 6 chassis
1 10 9
2 8
4
6
3
5 7
1: The compute module bay of the two nodes make up one node pair. Scaling out a
cluster with Gen 6 nodes is done by adding more node pairs.
2: Each Gen 6 node provides two ports for front-end connectivity. The connectivity
options for clients and applications are 10 GbE and 40 GbE.
3: Each node can have 1 or 2 SSDs that are used as L3 cache, global namespace
acceleration (GNA), or other SSD strategies.
4: Each Gen 6 nodes provides two ports for back-end connectivity. A Gen 6 node
supports 10 GbE, 40 GbE, and InfiniBand.
5: Power supply unit - Peer node redundancy: When a compute module power
supply failure takes place, the power supply from the peer compute module in the
node pair will temporarily provide power to both nodes.
PowerScale Concepts-SSP
6: Each Node has five drive sleds. Depending on the length of the chassis and type
of the drive, each node can handle up to 30 drives or as few as 15.
8: The sled can be either a short sled or a long sled. The types are:
9: The chassis comes in two different depths, the normal depth is about 37 inches
and the deep chassis is about 40 inches.
10: Large journals offer flexibility in determining when data should be moved to the
disk. Each node has a dedicated M.2 vault drive for the journal. A node mirrors
their journal to its peer node. The node writes the journal contents to the vault when
a power loss occurs. A backup battery helps maintain power while data is stored in
the vault.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 49
PowerScale Nodes
Gen 6.5 requires a minimum of three nodes to form a cluster. You can add single
nodes to the cluster. The F600 and F200 are a 1U form factor and based on the
R640 architecture.
8 5
1: Scaling out an F200 or an F600 node pool only requires adding one node.
3: Each F200 node has four SAS SSDs. Each F600 node has 8 NVMe SSDs.
4: Each Gen F200 and F600 node provides two ports for backend connectivity. The
PCIe slot 1 is used.
5: Redundant power supply units - When a power supply fails, the secondary
power supply in the node provides power. Power is supplied to the system equally
from both PSUs when the Hot Spare feature is disabled. Hot Spare is configured
using the iDRAC settings.
7: The nodes come in two different 1U models. The graphic shows the F200.
8: The F200 frontend connectivity uses the rack network daughter card (rNDC).
PowerScale Concepts-SSP
Important: The F600 nodes have a 4-port 1 GB NIC in the rNDC slot.
The NIC is not allocated to any OneFS function.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 51
PowerScale Nodes
Node Interconnectivity
1: Backend ports int-a and int-b. The int-b port is the upper port. Gen 6 backend
ports are identical for InfiniBand and Ethernet, and cannot be identified by looking
at the node. If Gen 6 nodes are integrated in a Gen 5 or earlier cluster, the backend
will use InfiniBand. Note that there is a procedure to convert an InfiniBand backend
to Ethernet if the cluster no longer has pre-Gen 6 nodes.
2: PowerScale nodes with different backend speeds can connect to the same
backend switch and not see any performance issues. For example, an environment
has a mixed cluster where A200 nodes have 10 GbE backend ports and H600
nodes have 40 GbE backend ports. Both node types can connect to a 40 GbE
switch without effecting the performance of other nodes on the switch. The 40 GbE
switch provides 40 GbE to the H600 nodes and 10 GbE to the A200 nodes.
4: There are two speeds for the backend Ethernet switches, 10 GbE and 40 GbE.
Some nodes, such as archival nodes, might not need to use all of a 10 GbE port
bandwidth while other workflows might need the full utilization of the 40 GbE port
bandwidth. The Ethernet performance is comparable to InfiniBand so there should
be no performance bottlenecks with mixed performance nodes in a single cluster.
Administrators should not see any performance differences if moving from
InfiniBand to Ethernet.
Gen 6 nodes can use either an InfiniBand or Ethernet switch on the backend.
InfiniBand was designed as a high-speed interconnect for high-performance
PowerScale Concepts-SSP
computing, and Ethernet provides the flexibility and high speeds that sufficiently
support the PowerScale internal communications.
Gen 6.5 only supports Ethernet. All new, PowerScale clusters support Ethernet
only.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 53
PowerScale Nodes
Quick Scalability
Ready to write
PowerScale Concepts-SSP
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 55
PowerScale OneFS Operating System
PowerScale Concepts-SSP
PowerScale Concepts-SSP
Module Objectives
PowerScale Concepts-SSP
33When nodes are added, OneFS redistributes the content to use the resources of
the entire cluster.
PowerScale Concepts-SSP
• FlexProtect35.
• Runs on all nodes
• Each node is a peer36.
• Prevents bottlenecking37.
• A copy of OneFS is on every cluster node.
• 10 GBE, 40 GbE (Gen 6): 10 GBE, 25 GBE, 100 GBE(Gen 6.5) and infiniBand
handle all intracluster communications.
35 Creates an n-way, redundant fabric that scales as nodes are added to the
cluster, providing 100% data availability even with four simultaneous node failures.
36 Each node shares the management workload and acts independently as a point
of access for incoming data request.
PowerScale Concepts-SSP
Benefits of OneFS
38When a node is added to the cluster, it adds computing power, storage, caching,
and networking resources.
PowerScale Concepts-SSP
OneFS supports access to the same file using different protocols and
authentication methods simultaneously. SMB clients that authenticate using Active
Directory (AD), and NFS clients that authenticate using LDAP, can access the
same file with their appropriate permissions applied.
• OneFS translates Windows Security Identifiers (SIDS) and UNIX User Identities
(UIDs) into a common identity format.
• Different authentication sources.
• Permissions activities are transparent to client.
• Authenticate against correct source.
• File access behavior as protocol expects.
• Correct permissions applied - stores the appropriate permissions for each
identity or group.
PowerScale Concepts-SSP
Authentication
1
5 2
1: Active Directory (AD): The primary reason for joining the cluster to an AD
domain is to let the AD domain controller perform user and group authentication.
4: Local or File Provider: OneFS supports local user and group authentication
using the web administration interface.
PowerScale Concepts-SSP
Policy-Based Automation
• Includes the way data is distributed across the cluster and on each node.
• Includes how client connections get distributed among the nodes, when and
how maintenance tasks are run.
PowerScale Concepts-SSP
Management Interfaces
• Serial Console39
• Web Administration Interface (WebUI)40
• Command Line Interface (CLI)41
39The serial console is used for initial cluster configurations by establishing serial
access to the node designated as node 1.
PowerScale Concepts-SSP
42The PAPI is divided into two functional areas: one area enables cluster
configuration, management, and monitoring functionality, and the other area
enables operations on files and directories on the cluster.
43The Front Panel Display is located on the physical node or chassis. It is used to
perform basic administrative tasks onsite.
PowerScale Concepts-SSP
So who is allowed to access and make configuration changes using the cluster
management tools? In addition to the integrated root and admin users, OneFS
provides role-based access control (RBAC). With RBAC, you can define privileges
to customize access to administration features in the OneFS WebUI, CLI, and for
PAPI management.
• Grant or deny access to management features.
Configured user with restricted privileges
• RBAC
• Set of global admin privileges
• Five preconfigured admin roles
• Zone RBAC (ZRBAC)
• Set of admin privileges specific to an access zone
• Two preconfigured admin roles
• Can create custom roles.
• Assign users to one or more roles.
PowerScale Concepts-SSP
Dell Technologies
Support Location
Cluster
Location
If there is an issue with your cluster, there are two types of support available. You
can manually upload logfiles to the Dell Technologies support FTP site, or use
Secure Remote Services.
• Manually FTP upload logfiles
• As needed.
• Support requests logfiles.
• Secure Remote Support
• Broader product support.
• Manual logfile uploads.
• 24x7 remote monitoring - node-by-node basis and sends alerts regarding
the health of devices.
• Allows remote cluster access - requires permission.
• Secure authentication with AES 256-bit encryption and RSA digital
certificates.
• Log files provide detailed information about the cluster activities.
• Remote session that is established through SSH or the WebUI - support
personnel can run scripts that gather diagnostic data about cluster settings and
operations. Data is sent to a secure FTP site where service professionals can
open support cases and troubleshoot on the cluster.
PowerScale Concepts-SSP
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 69
Data Management and Security
PowerScale Concepts-SSP
Module Objectives
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 71
Data Management and Security
Node Pool
Data distribution is how OneFS spreads data across the cluster. Various models of
PowerScale nodes, or node types can be present in a cluster. Nodes are assigned
to node pools based on the model type, number of drives, and the size of the
drives. The cluster can have multiple node pools, and groups of node pools can be
combined to form tiers of storage. Data distributes among the different node pools
that are based on the highest percentage of available space. This means that the
data target can be a pool or a tier anywhere on the cluster.
PowerScale Concepts-SSP
Data IO Optimization
Manage directories or
Manage cluster-wide by files
default
random
concurrent
sequential
You can optimize data input and output to match the workflows for your business.
By default, optimization is managed cluster-wide, but you can manage individual
directories or individual files. The data access pattern can be optimized for random
access, sequential access, or concurrent access. For example, sequential
optimization has aggressive prefetching. The prefetch, or read ahead, is an
optimization algorithm that attempts to predict what data is needed next, before the
request is made. When clients open larger files, especially streaming formats like
video and audio, OneFS assumes that you will watch minute four of the video after
minute three. Prefetch proactively loads minutes four, five, and sometimes even six
into memory before it is requested. Prefetch delivers those minutes faster than
returning to the hard drive for each request. With OneFS, you can configure the
prefetch cache characteristics to work best with the selected access pattern.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 73
Data Management and Security
Performance optimization is the first thing a customer notice about their cluster in
day to day operations. But what does the average administrator notice second?
They notice when a cluster has issues after they notice how great it works. They
want it fast, and they want it to work. That is a reason why data protection is
essential.
Data protection level refers to how many components in a cluster can malfunction
without loss of data.
• Flexible and configurable.
• Virtual hot spare - allocate disk space to hold data as it is rebuilt when a disk
drive fails.
• Select FEC protection by node pool, directory, or file.
• Extra protection creates more FEC stripes, increasing overhead.
• Standard functionality is available in the unlicensed version of SmartPools.
PowerScale Concepts-SSP
You can subdivide capacity usage by assigning storage quotas to users, groups,
and directories.
• Policy-based quota management.
• Nesting - place a quota on a department, and then a smaller quota on each
department user, and a different quota on the department file share.
• Thin provisioning - shows available storage even if capacity is not available.
• Quota types
• Accounting - informational only, can exceed quota.
• Enforcement soft limit - notification sent when exceeded
• Enforcement hard limit - deny writes.
• Customizable quota notifications.
• Requires SmartQuotas license.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 75
Data Management and Security
PowerScale Concepts-SSP
InsightIQ is a powerful tool that monitors one or more clusters and then presents
data in a robust graphical interface with reports you can export. You can examine
the information and break out specific information you want, and even take
advantage of usage growth and prediction features. InsightIQ offers:
• Monitor system usage - performance and file system analytics.
• Requires a server or VMware system external to cluster.
• Free InsightIQ license.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 77
Data Management and Security
Each stripe is protected separately with forward error correction (FEC) protection
blocks, or parity. Shown is a 1-megabyte file that is divided into two stripe units with
N+2 protection.
• Protected at data stripe - one or two data or protection stripe units are contained
on a single node for any given data stripe.
• Striped across nodes.
• Variable protection levels - set separately for node pools, directories, or even
individual files.
• Set at node pool, directory, or file.
• High availability is integrated - data is spread onto many drives and multiple
nodes, all ready to help reassemble the data when a component fails.
PowerScale Concepts-SSP
Data resiliency is the ability to recover past versions of a file that has changed over
time. Sooner or later, every storage admin gets asked to roll back to a previous
“known good” version of a file. OneFS provides this capability using snapshots.
• File change rollback technology - called snapshots.
• Copy-on-write (CoW) - writes the original blocks to the snapshot version first,
and then writes the data to the file system, incurs a double write penalty but less
fragmentation.
• Redirect-on-write (RoW) - writes changes into available file system space and
then update pointers to look at the new changes, there is no double write
penalty but more fragmentation.
• Policy-based
• Scheduled snapshots
• Policies determine the snapshot schedule, path to the snapshot location,
and snapshot retention periods.
• Deletions happen as part of a scheduled job, or are deleted manually.
• Out of order deletion allowed, but not recommended.
• Some system processes use with no license required.
• Full capability requires SnapshotIQ license.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 79
Data Management and Security
PowerScale Concepts-SSP
Source Target
Replication keeps a copy of data from one cluster on another cluster. OneFS
replicates during normal operations, from one PowerScale cluster to another.
Replication may be from one to one, or from one to many PowerScale clusters.
Cluster-to-cluster synchronization
Cluster-to-cluster synchronization
• Copy - new files on the source are copied to the target, while files deleted on
the source remain unchanged on the target.
• Synchronization - only works in one direction and both the source and target
clusters maintain identical file sets, except that files on the target are read-only.
Per directory or for specific types of data and can set exceptions to include or
exclude specific files.
• Manual start
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 81
Data Management and Security
• On schedule
• When changes made
Bandwidth throttling
Bandwidth throttling - used on replication jobs to optimize resources for high priority
workflows.
PowerScale Concepts-SSP
Data Retention
Data retention is the ability to prevent data from being deleted or modified before
some future date. In OneFS, you can configure data retention at the directory level,
so that different directories can have different retention policies. You can also use
policies to automatically commit certain types of files for retention.
• Two modes of retention
• Enterprise (more flexible) - enable privileged deletes by an administrator.
• Compliance (more secure) - designed to meet SEC regulatory requirements.
Once data is committed to disk, individuals cannot change or delete the data
until the retention clock expires - OneFS prohibits clock changes.
• Compatible with SyncIQ replication.
• Requires SmartLock license.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 83
PowerScale and Big Data
PowerScale Concepts-SSP
PowerScale Concepts-SSP
Module Objectives
PowerScale Concepts-SSP
PowerScale Concepts-SSP
The “Three v's”– volume, velocity, and variety – often arrive together. When they
combine, administrators truly feel the need for high performance, higher capacity
storage. The three V's generate the challenges of managing Big Data.
Growing data has also forced an evolution in storage architecture over the years
due to the amount of maintained data. PowerScale is a Big Data solution because
it can handle the volume, velocity, and variety that defines the fundamentals of Big
Data.
1: Challenge: Nonflexible data protection. When you have Big Data volumes of
information to store, it had better be there, dependably. If an organization relies on
RAID to protect against data loss or corruption, the failure of a single disk drive
causes a disproportionate inconvenience. The most popular RAID implementation
scheme allows the failure of only two drives before data loss. (A sizable Big Data
installation easily has more than 1000 individual hard drives, so odds are at least
one drive is down at any time.) The simpler answer is to protect data using a
different scheme.
What is meant by volume? Consider any global website that works at scale. One
example of Big Data volume is the YouTube press page that says YouTube ingests
100 hours of video every minute.
PowerScale Concepts-SSP
graphics chip prototype, sensors generate many terabytes of data per second.
Storing terabytes of data in seconds is an example of Big Data velocity.
3: Perhaps the best example of variety is the migration of the world to social media.
On a platform such as Facebook, people post all kinds of file formats: text, photos,
video, polls, and more. Many kinds of data at that scale represent Big Data variety.
PowerScale Concepts-SSP
45 Challenge: SAN and scale-up NAS architectures must reserve much of the raw
capacity of the system for management and administrative overhead. Overhead
includes RAID parity disks, metadata for all the LUNs and mega LUNs, duplicate
copies of the file system, and so on. As a result, conventional SAN and NAS
architectures often use half of the raw capacity available, because of the headroom
for each separate stack of storage. Suppose that you have seven different silos of
data. When you put them in one large volume, you immediately get back the
headroom from six of the seven stacks. In that way, PowerScale offers high
utilization. PowerScale customers routinely use 80% or more of raw disk capacity.
PowerScale Concepts-SSP
46 Some data storage architectures use two controllers, sometimes called servers
or filers, to run a stack of many hard drives. You can scale capacity by adding more
hard drives, but it is difficult to scale performance. In a given storage stack, the
hard drives offer nothing but capacity. All the intelligence of the system, including
computer processing and RAM, must come from the two filers. If the horsepower of
the two filers becomes insufficient, the architecture does not enable you to pile on
more filers. You start over with another stack and two more filers. In contrast, every
node in an PowerScale cluster contains capacity plus computing power plus
memory. The nodes can work in parallel, so each node you add scales out linearly.
In other words, all aspects of the cluster scale up, including capacity and
performance.
47 Due to the architectural restrictions, SAN and scale-up NAS end up with several
isolated stacks of storage. Many sites have a different storage stack for each
application or department. A backup storage stack is an example. Instead, an
administrator has to manually arrange a data migration. If the R&D stack performs
product testing that generates results at Big Data velocity, the company may
establish an HPC stack, which could reach capacity rapidly. Other departments or
workflows may have independent storage stacks with lot of capacity remaining, but
there is no automated way for R&D to offload their HPC overflow. In contrast, an
PowerScale cluster distributes data across all its nodes to keep them all at equal
capacity. You do not have one node that is overworked while other nodes sit idle.
PowerScale Concepts-SSP
There are no hot spots, and thus, no manual data migrations. If the goal is to keep
pace with Big Data velocity, automated balancing makes more sense.
48In conventional storage, a file is typically confined to a RAID stripe. That means
that the maximum throughput of reading that file is limited to how fast those drives
can deliver the file. In modern workflows where a hundred engineers or a thousand
digital artists access a file, the RAID drives cannot keep up. Perhaps the two filers
on that stack cannot process that many requests efficiently. With PowerScale,
every node has at least a dozen drives, plus more RAM and more computer
processing, for more caching and better concurrent access. When there is heavy
demand for a file, several nodes can deliver it.
49 Besides manual data migrations, conventional storage has many more manual
processes. A SAN or a scale-up NAS administrator spends a significant amount of
time creating and managing LUNs, partitioning storage, establishing mounts,
launching jobs, and so on. In contrast, PowerScale is policy-driven. Once you
define your policies, the cluster does the rest automatically.
PowerScale Concepts-SSP
A scale-out Data Lake is a large storage solution where vast amounts of data from
other solutions or locations are combined into a single store. Elements of a data
lake are:
• Digital repository to store massive data.
• Variety of formats.
• Can do computations and analytics on original data.
• Helps address the variety issue with Big Data.
• Data can be secured, analyzed, and actions taken based on insights.
• Enterprises can eliminate the cost of having silos of information.
• Provides scaling capabilities in terms of capacity, performance, security, and
protection.
PowerScale Concepts-SSP
PowerScale Concepts-SSP
Unmatched
Easy Growth
Efficiency
Cloud Tiering
Ready Hadoop Enabled
A Data Lake is a central data repository that stores data from various sources, such
as file shares, web apps, and the cloud. It enables businesses to access the same
data for various uses and enables the manipulation of data using various clients,
analyzers, and applications. The data is real-time production data with no need to
copy or move it from an external source, like another Hadoop cluster, into the Data
Lake. The Data Lake provides tiers that are based on data usage, and the ability to
instantly increase the storage capacity when needed. This slide identifies the key
characteristics of a scale-out Data Lake.
PowerScale Concepts-SSP
PowerScale CloudPools
The PowerScale CloudPools software enables you to select from various public
cloud services or use a private cloud. CloudPools offers the flexibility of another tier
of storage that is off-premise and off-cluster. Essentially what CloudPools do is
provide a lower TCO50 for archival-type data.
• Treat cloud storage as another cluster-connected tier.
• Policy-based automated tiering
• Address rapid data growth and optimize data center storage resources - use
valuable on-site storage resources for active data.
• Send rarely used or accessed data to cloud.
• Seamless integration with data – retrieve at any time.
PowerScale Concepts-SSP
PowerScale Concepts-SSP
Edge locations are often inefficient islands of storage, running with limited IT
resources, and inconsistent data protection practices. Data at the edge generally
lives outside of the Data Lake, making it difficult to incorporate into data analytics
projects. The edge-to-core-to-cloud approach extends the Data Lake to edge
locations and out into the cloud. It enables consolidation, protection, management,
and backups of remote edge location data.
PowerScale Concepts-SSP
Course Summary
PowerScale Concepts-SSP
Course Summary
PowerScale Concepts-SSP
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 101
Appendix
PowerScale Nodes
Individual PowerScale nodes provide the data storage capacity and processing
power of the PowerScale scale-out NAS platform. All of the nodes are peers to
each other and so there is no single 'master' node and no single 'administrative
node'.
• No single master
• No single point of administration
Administration can be done from any node in the cluster as each node provides
network connectivity, storage, memory, non-volatile RAM (NVDIMM) and
processing power found in the Central Processing Units (CPUs). There are also
different node configurations, compute, and capacity. These varied configurations
can be mixed and matched to meet specific business needs.
Each contains.
• Disks
• Processor
• Cache
PowerScale Concepts-SSP
Tip: Gen 5 and Gen 6 nodes can exist within the same cluster. Every
PowerScale node is equal to every other PowerScale node of the
same type in a cluster. No one specific node is a controller or filer.
PowerScale Concepts-SSP
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 103
PowerScale Concepts-SSP
COURSE GUIDE
(V3)
PARTICIPANT GUIDE
Internal Use - Confidential
PowerScale Administration-SSP1
PowerScale Administration-SSP1
PowerScale Administration-SSP1
PowerScale Administration-SSP1
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 5
NAS, PowerScale, and OneFS
Scenario
Storage Technologies
DAS
In the early days of system data, corporations1 stored data on hard drives in a
server. To minimize risk, corporations mirrored the data on a RAID. This technique
is called Direct Attached Storage (DAS).
1The intellectual property of the company depended entirely upon that hard drive's
continued functionality.
PowerScale Administration-SSP1
RAID
DAS
SAN
As applications proliferated, soon there were many servers, each with its own DAS.
This worked fine, with some drawbacks2. Due to this limitation with DAS, SAN was
introduced which effectively utilized volume manager and RAID.
SAN
NAS
2If one server’s DAS was full while another server’s DAS was half empty, the
empty DAS could not share its space with the full DAS.
3PCs worked differently from the storage file server and the network
communications in PCs, only communicate from one file system to another file
system.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 7
NAS, PowerScale, and OneFS
NAS
CAS
PowerScale Administration-SSP1
Cloud
Cloud storage stores data over the Internet to a cloud provider. The cloud provider
manages and protects the data. Typically, cloud storage is delivered on demand
with just-in-time capacity and costs.
NAS Overview
NAS provides the advantages of server consolidation by eliminating the need for
multiple file servers.
• Consolidates the storage that is used by the clients onto a single system,
making it easier to manage the storage.
• Uses network and file-sharing protocols to provide access to the file data5.
• Uses its own operating system6 and integrated hardware and software
components to meet specific file-service needs.
PowerScale clusters are a NAS solution. There are two types of NAS architectures;
scale-up and scale-out.
5NAS enables both UNIX and Microsoft Windows users to share the same data
seamlessly.
6 Its operating system is optimized for file I/O and, therefore, performs file I/O better
than a general-purpose server. As a result, a NAS device can serve more clients
than general-purpose servers and provide the benefit of server consolidation.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 9
NAS, PowerScale, and OneFS
Scale-Up
Controller with
disk shelves
Independent systems on
network - separate
Clients points of management
Structured or Unstructured
storage
Scale-Out
• With a clustered NAS solutions, or scale-out architecture, all the NAS boxes, or
PowerScale nodes, belong to a unified cluster with a single point of
management.
• In a scale-out solution, the computational throughput, disks, disk protection, and
management are combined and exist for a single cluster.
PowerScale Administration-SSP1
Unstructured storage
PowerScale cluster
1000+ PBS
Scale-Out NAS
Scale-out NAS7 is now a mainstay in most data center environments. The next
wave of scale-out NAS innovation has enterprises embracing the value8 of NAS
and adopting it as the core of their infrastructure.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 11
NAS, PowerScale, and OneFS
PowerScale Administration-SSP1
With traditional NAS systems the file system9, volume manager10, and the
implementation of RAID11 are all separate entities.
OneFS is the operating system and the underlying file system that drives and
stores data.
OneFS is a single file system that performs the duties of the volume manager and
applies protection.
OneFS is built on FreeBSD.
• Creates a single file system for the cluster.12
• Volume manager and protection.13
• Data shared across cluster.14
• Scale resources.15
9The file system is responsible for the higher-level functions of authentication and
authorization.
12As nodes are added, the file system grows dynamically and content is
redistributed.
13 OneFS performs the duties of the volume manager and applies protection to the
cluster as a whole. There is no partitioning, and no need for volume creation. All
data is striped across all nodes.
14Because all information is shared among nodes, the entire file system is
accessible by clients connecting to any node in the cluster.
15Each PowerScale storage node contains globally coherent RAM, meaning that,
as a cluster becomes larger, it also becomes faster. When adding a node, the
performance scales linearly.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 13
NAS, PowerScale, and OneFS
Challenge
IT Manager:
Open participation question:
Question: What is the difference between scale-up and scale-out
architecture?
PowerScale Administration-SSP1
PowerScale
Scenario
Gen 6 highlights.
Gen 6.5 highlights.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 15
NAS, PowerScale, and OneFS
Gen 6 requires a minimum of four nodes to form a cluster. You must add nodes to
the cluster in pairs.
The chassis holds four compute nodes and 20 drive sled slots.
Both compute modules in a node pair power-on immediately when one of the
nodes is connected to a power source.
Gen 6 chassis
1 10 9
2 8
4
6
3
5 7
1: The compute module bay of the two nodes make up one node pair. Scaling out a
cluster with Gen 6 nodes is done by adding more node pairs.
2: Each Gen 6 node provides two ports for front-end connectivity. The connectivity
options for clients and applications are 10 GbE, 25 GbE, and 40 GbE.
3: Each node can have 1 or 2 SSDs that are used as L3 cache, global namespace
acceleration (GNA), or other SSD strategies.
4: Each Gen 6 node provides two ports for back-end connectivity. A Gen 6 node
supports 10 GbE, 40 GbE, and InfiniBand.
5: Power supply unit - Peer node redundancy: When a compute module power
supply failure takes place, the power supply from the peer node temporarily
provides power to both nodes.
6: Each node has five drive sleds. Depending on the length of the chassis and type
of the drive, each node can handle up to 30 drives or as few as 15.
PowerScale Administration-SSP1
8: The sled can be either a short sled or a long sled. The types are:
9: The chassis comes in two different depths, the normal depth is about 37 inches
and the deep chassis is about 40 inches.
10: Large journals offer flexibility in determining when data should be moved to the
disk. Each node has a dedicated M.2 vault drive for the journal. A node mirrors
their journal to its peer node. The node writes the journal contents to the vault when
a power loss occurs. A backup battery helps maintain power while data is stored in
the vault.
Gen 6.5 requires a minimum of three nodes to form a cluster. You can add single
nodes to the cluster. The F600 and F200 are a 1U form factor and based on the
R640 architecture.
1
5
8 2
7 4
1: Scaling out an F200 or an F600 node pool only requires adding one node.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 17
NAS, PowerScale, and OneFS
3: Each Gen F200 and F600 node provides two ports for backend connectivity. The
PCIe slot 1 is used.
4: Redundant power supply units - When a power supply fails, the secondary
power supply in the node provides power. Power is supplied to the system equally
from both PSUs when the Hot Spare feature is disabled. Hot Spare is configured
using the iDRAC settings.
5: Disks in a node are all the same type. Each F200 node has four SAS SSDs.
6: The nodes come in two different 1U models, the F200 and F600. You need
nodes of the same type to form a cluster.
7: The F200 front-end connectivity uses the rack network daughter card (rNDC).
PowerScale offers nodes for different workloads of performance and capacity. The
table below shows some of the node specifications. To get the latest and a
complete list of specification and compare between the node offerings, browse the
product page.
PowerScale Administration-SSP1
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 19
NAS, PowerScale, and OneFS
PowerScale Administration-SSP1
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 21
NAS, PowerScale, and OneFS
PowerScale Administration-SSP1
PowerScale Features
The design goal for the PowerScale nodes are to keep the simple ideology of NAS,
provide the agility of the cloud, and the cost of commodity. Click each tab to learn
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 23
NAS, PowerScale, and OneFS
more on the features provided by PowerScale. See the student guide for more
information.
16A Media and Entertainment production house needs high single stream
performance at PB scale that is cost optimized. The organization requires cloud
archive in a single namespace, archive optimized density with a low Total Cost of
Ownership (TCO) solution. This environment typically has large capacities and
employs new performance technologies at will.
PowerScale Administration-SSP1
Data Protection
Sizing
17Financial sectors rely heavily on data protection and availability to operate. Data
loss such as customer transactions or system downtime can negatively affect the
business.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 25
NAS, PowerScale, and OneFS
The Gen 6x platforms addresses the challenges of agility and lower TCO by:
• Dedicated cache drives
• Modular architecture
• Non-disruptive upgrades
PowerScale has no dependency on the flash boot drive. Gen 6 nodes boot from
boot partitions on the data drives. These drives are protected using erasure coding
to remove the dependency on dedicated boot drives. Next, PowerScale uses SSD
drives for the journal to remove the NVRAM dependency present on Gen 5 nodes.
There are now multiple distributed copies of the journal.
18A simplicity and agility use case is a small start-up company growing at rapid
pace, who needs to start with limited capacity and then grow on demand for scale
and new workloads.
PowerScale Administration-SSP1
By creating smaller failure domains with significantly fewer drives in each node pool
and neighborhood, increases the reliability of the system by reducing the spindle-
to-CPU ratio. The increased reliability enables the cluster to use larger capacity
drives, without the risk of overburdening the system in the event of a drive failure.
PowerScale enables predictable failure handling at Petabyte (PB) densities.
Gen 6 platforms have dedicated cache drives for dedicated cache. The caching
options offered are 1 or 2 SSD configurations in various capacities to maximize
front end performance. Gen 6 hardware is focused on support and serviceability,
based on a modular architecture with full redundancy. It is possible to increase
performance with data in place, increase cache without disruption, and upgrade
speeds and feeds non-disruptively.
PowerScale Family
The Gen 6x family has different offerings that are based on the need for
performance and capacity. You can scale out compute and capacity separately.
OneFS runs on all nodes. Click each tab to learn more about the different offerings.
F-Series
The F-series nodes sit at the top of both performance and capacity, with the all-
flash arrays. The all-flash platforms can accomplish 250-300k protocol operations
per chassis, and get 15 GB/s aggregate read throughput from the chassis. Even
when the cluster scales, the latency remains predictable.
• F800
• F810
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 27
NAS, PowerScale, and OneFS
• F600
• F200
H-Series
After F-series nodes, next in terms of computing power are the H-series nodes.
These are hybrid storage platforms that are highly flexible and strike a balance
between large capacity and high-performance storage to provide support for a
broad range of enterprise file workloads.
• H400
• H500
• H5600
• H600
A-Series
The A-series nodes namely have lesser compute power compared to other nodes
and are designed for data archival purposes. The archive platforms can be
combined with new or existing all-flash and hybrid storage systems into a single
cluster that provides an efficient tiered storage solution.
• A200
• A2000
PowerScale Administration-SSP1
Node Interconnectivity
1: Backend ports int-a and int-b. The int-b port is the upper port. Gen 6 backend
ports are identical for InfiniBand and Ethernet, and cannot be identified by looking
at the node. If Gen 6 nodes are integrated in a Gen 5 or earlier cluster, the backend
will use InfiniBand. Note that there is a procedure to convert an InfiniBand backend
to Ethernet if the cluster no longer has pre-Gen 6 nodes.
2: PowerScale nodes with different backend speeds can connect to the same
backend switch and not see any performance issues. For example, an environment
has a mixed cluster where A200 nodes have 10 GbE backend ports and H600
nodes have 40 GbE backend ports. Both node types can connect to a 40 GbE
switch without effecting the performance of other nodes on the switch. The 40 GbE
switch provides 40 GbE to the H600 nodes and 10 GbE to the A200 nodes.
4: There are two speeds for the backend Ethernet switches, 10 GbE and 40 GbE.
Some nodes, such as archival nodes, might not need to use all of a 10 GbE port
bandwidth while other workflows might need the full utilization of the 40 GbE port
bandwidth. The Ethernet performance is comparable to InfiniBand so there should
be no performance bottlenecks with mixed performance nodes in a single cluster.
Administrators should not see any performance differences if moving from
InfiniBand to Ethernet.
Gen 6 nodes can use either an InfiniBand or Ethernet switch on the backend.
InfiniBand was designed as a high-speed interconnect for high-performance
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 29
NAS, PowerScale, and OneFS
computing, and Ethernet provides the flexibility and high speeds that sufficiently
support the PowerScale internal communications.
Gen 6.5 only supports Ethernet. All new, PowerScale clusters support Ethernet
only.
Network: There are two types of networks that are associated with a cluster:
internal and external.
19In general, keeping the network configuration simple provides the best results
with the lowest amount of administrative overhead. OneFS offers network
provisioning rules to automate the configuration of additional nodes as clusters
grow.
PowerScale Administration-SSP1
Ethernet
Clients connect to the cluster using Ethernet connections20 that are available on all
nodes.
20Because each node provides its own Ethernet ports, the amount of network
bandwidth available to the cluster scales linearly.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 31
NAS, PowerScale, and OneFS
OneFS supports a single cluster21 on the internal network. This back-end network,
which is configured with redundant switches for high availability, acts as the
backplane for the cluster.22
The Gen 6x back-end topology in OneFS 8.2 and later supports scaling a
PowerScale cluster to 252 nodes. See the participant guide for more details.
27 uplinks per
spine switch
Leaf-Spine is a two level hierarchy where nodes connect to leaf switches, and leaf
switches connects to spine switches. Leaf switches do not connect to one another,
and spine switches do not connect to one another. Each leaf switch connects with
22 This enables each node to act as a contributor in the cluster and isolating node-
to-node communication to a private, high-speed, low-latency network. This back-
end network utilizes Internet Protocol (IP) for node-to-node communication.
PowerScale Administration-SSP1
each spine switch and all leaf switches have the same number of uplinks to the
spine switches.
The new topology uses the maximum internal bandwidth and 32-port count of Dell
Z9100 switches. When planning for growth, F800 and H600 nodes should connect
over 40 GbE ports whereas A200 nodes may connect using 4x1 breakout cables.
Scale planning enables for nondisruptive upgrades, meaning as nodes are added,
no recabling of the backend network is required. Ideally, plan for three years of
growth. The table shows the switch requirements as the cluster scales. In the table,
Max Nodes indicate that each node is connected to a leaf switch using a 40 GbE
port.
Challenge
IT Manager:
Open participation question:
Question: What are the differences between Gen 6 nodes and
Gen 6.5 nodes?
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 33
NAS, PowerScale, and OneFS
Resources
PowerScale Administration-SSP1
Scenario
• Serial Console
• Web Administration Interface (WebUI)
• Command Line Interface (CLI)
• Platform Application Programming Interface (PAPI)
• Front Panel Display
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 35
NAS, PowerScale, and OneFS
This video provides an overview on the serial console. See the student guide for a
transcript of the video.
https://edutube.emc.com/Player.aspx?vno=jHnaLyBuvlzyrARCLAU/jw==&autoplay
=true
Four options are available for managing the cluster. The web administration
interface (WebUI), the command-line interface (CLI), the serial console, or the
platform application programming interface (PAPI), also called the OneFS API. The
first management interface that you may use is a serial console to node 1. A serial
connection using a terminal emulator, such as PuTTY, is used to initially configure
the cluster. The serial console gives you serial access when you cannot or do not
want to use the network. Other reasons for accessing using a serial connection
may be for troubleshooting, site rules, a network outage, and so on. Shown are the
terminal emulator settings.
PowerScale Administration-SSP1
configuration Wizard, running the isi config command enables you to change
the configuration settings.
isi config
Common commands -
shutdown, status, name
Change
s
prompt
to >>>
Other "isi" commands not available in
configuration console
The isi config command, pronounced "izzy config," opens the configuration
console. The console contains configured settings from the time the Wizard started
running.
Use the console to change initial configuration settings. When in the isi config
console, other configuration commands are unavailable. The exit command is
used to go back to the default CLI.
OneFS
version
User must have logon privileges
Connect to
any node in
cluster over
HTTPS on
port 8080
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 37
NAS, PowerScale, and OneFS
The WebUI requires at least one IP address configured23 on one of the external
Ethernet ports present in one of the nodes.
• Out-of-band24
• In-band25
Both methods are done using any SSH client such as OpenSSH or PuTTY. Access
to the interface changes based on the assigned privileges.
OneFS commands are code that is built on top of the UNIX environment and are
specific to OneFS management. You can use commands together in compound
24Accessed using a serial cable connected to the serial port on the back of each
node. As many laptops no longer have a serial port, a USB-serial port adapter may
be needed.
PowerScale Administration-SSP1
command structures combining UNIX commands with customer facing and internal
commands.
4
1
5
3 6
5: The CLI command use includes the capability to customize the base command
with the use of options, also known as switches and flags. A single command with
multiple options result in many different permutations, and each combination
results in different actions performed.
6: The CLI is a scriptable interface. The UNIX shell enables scripting and execution
of many UNIX and OneFS commands.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 39
NAS, PowerScale, and OneFS
CLI Usage
Option
explanation
The man isi or isi --help command is an important command for a new
administrator. These commands provide an explanation of the available isi
commands and command options. You can also view a basic description of any
command and its available options by typing the -h option after the command.
26A chief benefit of PAPI is its scripting simplicity, enabling customers to automate
their storage administration.
PowerScale Administration-SSP1
3: Some commands are not PAPI aware, meaning that RBAC roles do not apply.
These commands are internal, low-level commands that are available to
administrators through the CLI. Commands not PAPI aware: isi config, isi
get, isi set, and isi services
4: The number indicates the PAPI version. If an upgrade introduces a new version
of PAPI, some backward compatibility ensures that there is a grace period for old
scripts to be rewritten.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 41
NAS, PowerScale, and OneFS
The Gen 6 front panel display is an LCD screen with five buttons used for basic
administration tasks27.
The Gen 6.5 front panel has limited functionality28 compared to the Gen 6.
Challenge
Lab Assignment: Launch the lab image and connect to the cluster
using the WebUI and the CLI.
27Some of them include: adding the node to a cluster, checking node or drive
status, events, cluster details, capacity, IP and MAC addresses.
28You can join a node to a cluster and the panel display node name after the node
has joined the cluster.
PowerScale Administration-SSP1
Scenario
Your Challenge: The new IT manager has given you a task to describe
the OneFS licensing and add the new nodes to the PowerScale cluster.
Licensing
Evaluation licensing
No individual per-
enabled from cluster
feature keys
Upgrades translate
keys into file
WebUI Cluster management > Licensing > Open Activation File Wizard or use the "isi license"
command.
In OneFS 8.1 and later a single license file contains all the licensed feature
information in a single location.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 43
NAS, PowerScale, and OneFS
Device ID cannot be
changed
Changing the LNN 3 to LNN 5 to maintain the sequential numbering of the nodes.
You should have an understanding of the two different numbers that identify a
node. The numbers are the device ID and logical node number or LNN.
The status advanced command from the isi config sub menu shows the
LNNs and device ID.
When a node joins a cluster, it is assigned a unique node ID number. If you remove
and rejoin a node from the cluster, the node is assigned a new device ID.
You can change an LNN in the configuration console. To change the LNN to
maintain the sequential numbering of the nodes use lnnset <OldNode#>
<NewNode#>.
PowerScale Administration-SSP1
When adding new nodes to a cluster, the cluster gains more CPU, memory, and
disk space. The methods for adding a node are:
• Front panel
• Configuration Wizard
• WebUI
• CLI
Join the nodes in the order that the nodes should be numbered.
Adding a node not connected to the external network (NANON) increases the
storage and compute capacity of the cluster.
Nodes are automatically assigned node numbers and IP addresses on the internal
and external networks. A node joining the cluster with a newer or older OneFS
version is automatically reimaged to match the OneFS version of the cluster. A
reimage may take up to 5 minutes.
Compatibility
Hardware compatibility is a concern when combining dissimilar Gen 6.5 nodes. For
example, when adding a single F200 node with 48 GB RAM to an F200 node pool
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 45
NAS, PowerScale, and OneFS
that has nodes with 96 GB of RAM. Without compatibility, a minimum of three F200
nodes with 48 GB RAM is required, which creates a separate node pool.
Node series compatibility depends upon the amount of RAM, the SSD size, number
of HDDs, and the OneFS version.
Cluster Shutdown
Can shutdown
entire cluster
PowerScale Administration-SSP1
Administrators can restart or shutdown the cluster using the WebUI29 or the CLI30.
Challenge
Lab Assignment: Launch the lab and add a node using the
Configuration Wizard and add a node using the WebUI.
29The WebUI Hardware page has a tab for Nodes to shut down a specific node, or
the Cluster tab to shut down the cluster.
30
Native UNIX commands do not elegantly interact with OneFS, because the
OneFS file system is built as a separate layer on top of UNIX.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 47
NAS, PowerScale, and OneFS
Scenario
IT Manager: Good, looks like you know what the different PowerScale
management tools are. Now I want you to focus on the directory
structure that OneFS uses. This is important as it sets up the directory
structure we will use moving foreward.
OneFS root
directory
At the core of OneFS, is the single file system across the cluster (/ifs). The single
file system in practice is a common directory structure.
PowerScale Administration-SSP1
Using or intervening with the built-in directory paths is not recommended unless
explicitly instructed to do so.
• Using a single file system starting with a newly created directory under /ifs is
recommended.
• For example, in the simplest form, you can create /ifs/engineering where
the engineering department data is the top-level directory for the engineering
organization.
OneFS root
Cluster root
Authentication
and segregation
root Location to situate data and create
exports and shares as per
requirement
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 49
NAS, PowerScale, and OneFS
Use case:
• A company that is named X-Attire plans to implement a single cluster for their
engineering team.
• After conversations with the customer, you identify that the customer does not
plan to have another cluster for remote disaster recovery.
• The company name or authentication domain name is used as the access zone
name (x-attire).
PowerScale Administration-SSP1
Use case:
• X-Attire plans to implement a disaster recovery solution.
• X-Attire wants to replicate the Boston/homedirs directory to the Seattle data
center.
• from Seattle, they plan to replicate the /groupdirs directory to Boston.
• Having the directory structure design up front makes the implementation easier.
On the /ifs directory, do not set inherited ACLs and do not propagate ACL
values.
Permissions on levels 1 through 5 are customer-specific and you should define the
appropriate permissions and inherited permissions starting at the appropriate level.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 51
NAS, PowerScale, and OneFS
Challenge
Lab Assignment: Go to the lab and build the base directories. The
base directories are used throughout your implementation of the
PowerScale cluster.
PowerScale Administration-SSP1
PowerScale Administration-SSP1
Authentication Providers
Scenario
IT Manager: Now, the next thing to do is get the cluster pointed to the
Active Directory and LDAP servers. Before our clients can access files
that are stored on the cluster, they must be authenticated. Make sure
that you have a good understanding of the authentication providers that
the cluster supports.
1 2 3
4 5 6
PowerScale Administration-SSP1
reason for joining the cluster to an Active Directory domain is to perform user and
group authentication.
5: The local provider provides authentication, and lookup facilities for user accounts
added by an administrator.
PowerScale Administration-SSP1
Authentication
Authentication provider
source / directory
Access control architectural components that show two configured access zones.
lsassd is between the access protocols and the lower-level services providers.
The lsassd daemon mediates between the authentication protocols that clients
use and the authentication providers in the third row.
The authentication providers check their data repositories, which are shown on the
bottom row. The process determines user identity and subsequent access to files.
Function
Active Directory can serve many functions, but the primary reason for joining the cluster to an AD domain is to enable domain
users to access cluster data.
To join the cluster to AD, specify the fully qualified domain name, which can be
resolved to an IPv4 or an IPv6 address, and a username with join permission.
Areas to consider:
• Creates a single AD machine account
• Establishes trust relationship
PowerScale Administration-SSP1
PowerScale Administration-SSP1
Link:
https://edutube.emc.com/Player.aspx?vno=Xu/3IyDNSxbuNMOcLHrqBg==&autopl
ay=true
Select the Join a domain button. This demonstration shows the barest configuration
to join a domain. Start by entering the provider name. The NetBIOS requires that
computer names be 15 characters or less. Two to four characters are appended to
the cluster name you specify to generate a unique name for each node. If the
cluster name is more than 11 characters, you can specify a shorter name in the
Machine Name field. Enter the user name of the account that has the right to add
computer accounts to the domain, and then enter the account password. The
Enable Secure NFS checkbox enables users to log in using LDAP credentials, but
to do this, Services for NFS must be configured in the AD environment.
Shown is the CLI equivalent command used to join Active Directory. To display a
list of command options, run the isi auth ads create -h command at the
CLI. Now, before connecting to an LDAP server you should decide which optional
PowerScale Administration-SSP1
customizable parameters you want to use. Refer the Isilon Web Administration
Guide for details on each of the settings.
Click the Join button. While joining the domain, the browser window displays the
status of the process and confirms when the cluster has successfully joined the AD
domain. The join creates a single computer account is for the entire cluster.
And that is the most basic configuration. Note that AD and LDAP both use TCP
port 389. Even though both services can be installed on one Microsoft server, the
cluster can only communicate with one of services if they are both installed on the
same server. This concludes the demonstration.
PowerScale Administration-SSP1
31 The easiest method is to synchronize the cluster and the authentication servers
all to the same NTP source.
32The cluster time property sets the date and time settings, either manually or by
synchronizing with an NTP server. After an NTP server is established, setting the
date or time manually is not allowed.
33After a cluster is joined to an AD domain, adding an NTP server can cause time
synchronization issues. The NTP server takes precedence over the SMB time
synchronization with AD and overrides the domain time settings on the cluster.
35 Nodes use NTP between themselves to maintain cluster time. When the cluster
is joined to an AD domain, the cluster must stay synchronized with the time on the
domain controller. If the time differential is more than five minutes, authentication
may fail.
PowerScale Administration-SSP1
NTP Configuration
Chimers nodes can contact the external Non-chimer nodes use chimers as NTP
NTP servers servers
WebUI > General settings > NTP page to configure NTP and chimer settings.
You can configure specific chimer nodes by excluding other nodes using the
isi_ntp_config {add | exclude} <node#> command. The list excludes
nodes using their node numbers that are separated by a space.
LDAP Overview
Function
OneFS can authenticate users and groups against an LDAP repository in order to grant them access to the cluster. OneFS
supports Kerberos authentication for an LDAP provider.
PowerScale Administration-SSP1
37Each attribute has a name and one or more values that are associated with it
that is similar to the directory structure in AD.
PowerScale Administration-SSP1
Link:
https://edutube.emc.com/Player.aspx?vno=JKBFLVJaUoqGz8DJmH4zqg==&autop
lay=true
In this demonstration, we’ll go through the steps needed to configure LDAP for the
PowerScale cluster. Let us navigate to Access and then to Authentication providers
page. Next, select the LDAP tab. Now click the Add an LDAP provider button.
For this demonstration, I am only showing the barest configuration. Let us give our
LDAP a provider name. Next, I will enter the URI to the LDAP server. You must
configure a base distinguished name. Often issues involve either misconfigured
base DNs or connecting to the LDAP server. The top-level names almost always
mimic DNS names; for example, the top-level Isilon domain would be dc=isilon,
dc=com for Isilon.com. Our environment is DEES and lab.
Shown is the CLI equivalent command used to configure LDAP. To display a list of
these commands, run the isi auth ldap create -h command at the CLI.
And that is the most basic configuration.
PowerScale Administration-SSP1
Now, before connecting to an LDAP server you should decide which optional
customizable parameters you want to use. If there are any issues while configuring
or running the LDAP service, there are a few commands that can be used to help
troubleshoot. The ldapsearch command runs queries against an LDAP server to
verify whether the configured base DN is correct. The tcpdump command verifies
that the cluster is communicating with the assigned LDAP server.
You have the option to enter a netgroup. A netgroup, is a set of systems that reside
in a variety of different locations, that are grouped together and used for permission
checking. For example, a UNIX computer on the 5th floor, six UNIX computers on
the 9th floor, and 12 UNIX computers in the building next door, all combined into
one netgroup.
Select the Add LDAP Provider button. After the LDAP provider is successfully
added, the LDAP providers page displays a green status. This means that the
cluster can communicate with the LDAP server. Note that AD and LDAP both use
TCP port 389. Even though both services can be installed on one Microsoft server,
the cluster can only communicate with one of services if they are both installed on
the same server. This concludes the demonstration.
Challenge
Lab Assignment:
• Join the cluster to Active Directory
• Configure the cluster for LDAP
PowerScale Administration-SSP1
Access Zones
Scenario
IT Manager: Now that you have configured the cluster for Active
Directory and LDAP, it is time to take the next step in implementation.
You are configuring access zone for two organizations, finance and
engineering. Finance is a Microsoft Windows environment and
engineering is a Linux environment. Before you configure the cluster, I
want to ensure you understand access zones and what they do.
This video provides an overview for access zones. See the student guide for a
transcript of the video.
PowerScale Administration-SSP1
Link: https://edutube.emc.com/Player.aspx?vno=w/pzpXjL6ZCFlcdx0riu5A
Although the default view of a cluster is that of one physical machine, you can
partition a cluster into multiple virtual containers called access zones. Access
zones enable you to isolate data and control who can access data in each zone.
Access zones support configuration settings for authentication and identity
management services on a cluster. Configure authentication providers and
provision protocol directories, such as SMB shares and NFS exports, on a zone-by-
zone basis. Creating an access zone, automatically creates a local provider, which
enables you to configure each access zone with a list of local users and groups.
You can also authenticate through a different authentication provider in each
access zone.
The OneFS identity management maps users and groups from separate directory
services to provide a single combined identity. It also provides uniform access
control to files and directories, regardless of the incoming protocol.
PowerScale Administration-SSP1
External Protocols
Clients use the external access protocols to connect to the PowerScale cluster.
The supported protocols are SMB, NFS, S3, HTTP, FTP, HDFS, and SWIFT.
lsassd Daemon
The lsassd (L-sass-d) daemon mediates between the external protocols and the
authentication providers, with the daemon contacting the external providers for user
lookups.
PowerScale Administration-SSP1
External Providers
Internal Providers
Internal providers sit within the cluster operating system and are the Local, or File
Providers.
• File provider - authoritative third-party source of user and group information.
• Local provider - provides authentication and lookup facilities for user accounts
added by an administrator.
• Local provider automatically created in access zone.
PowerScale Administration-SSP1
1 4
2
4: The /ifs/eng base directory partitions data from the /ifs/dvt directory.
5: The base directory of the default System access zone is /ifs and cannot be
modified. Avoid using the OneFS built-in directories as base directories.
A base or root directory defines the tree structure of the access zone.
The access zone cannot grant access to any files outside of the base directory,
essentially creating a unique namespace.
PowerScale Administration-SSP1
This demonstration provides a look at access zone configuration. See the student
guide for a transcript of the video.
Link: https://edutube.emc.com/Player.aspx?vno=08ieHpVlyvyD+A8mTzHopA
In this demonstration, we will go through the steps to create access zones using
the WebUI and the CLI. First, let’s use the WebUI.
Navigate to Access and then to the Access zones page. Note that the System
access zone is shown in the table. The System zone is created by OneFS. Select
the Create an access zone button. In the window, enter the zone name for the new
access zone. Next enter the zone base directory. This should be unique, and you
should avoid using the OneFS built-in directories such as /ifs/data. Our base
directory is /ifs/sales.
Since we have not created this directory before creating the access zone, select
the checkbox to create the base directory automatically. Notice that we already
configured the authentication providers. This access zone is dedicated for the
Active Directory users. Add the AD provider and then select Create zone.
PowerScale Administration-SSP1
Next, we will create another access zone using the CLI. We are logged in via SSH
to node 1 and using the isi zone command. The name of this access zone is
engineering. The unique base directory is /ifs/engineering. Since the
/ifs/engineering directory does not exist, use the option to create it. And
finally, we will add the LDAP authentication provider to the zone.
Next verify that the zones are created. Use the list option. Moving back to the
WebUI, check the access zone page to verify the zones display. Instead of waiting
for the refresh, click on another page and then back.
This demonstration showed configuring access zones using the WebUI and the
CLI. This concludes the demonstration.
Listed are areas to consider when configuring and discussing access zones.
• The number of access zones should not exceed 50.
• As a good practice, configure an access zone for a specific protocol if multi-
protocol access is not needed. For example, an implementation with both NFS
and SMB access should have an access zone for the NFS access and another
access zone for the SMB access.
• Access zones and authentication providers must be in only one groupnet.
• Authentication sources are joined to the cluster and "seen" by access zones -
multiple instances of the same provider in different access zones is not
recommended.
• Authentication providers are not restricted to one specific zone.
• Only join AD providers not in same forest (untrusted forest).
• Shared UIDs in same zone can potentially cause UID/GID conflicts.
• You can overlap data between access zones for cases where workflows require
shared data - however, overlapping adds complexity that may lead to issues
with client access.
You can avoid configuration problems on the cluster when creating access zones
by following best practices guidelines.
PowerScale Administration-SSP1
System zone is for global admin Employ ZRBAC for zone administration.
access only.
Create zones to isolate data for Do not isolate if workflow requires shared
different clients. data.
Challenge
PowerScale Administration-SSP1
Groupnets
Scenario
Configure SmartConnect
IP address, VLAN, and
MTU on the subnet
Groupnets reside at the top tier of the networking hierarchy and are the
configuration level for managing multiple tenants on your external network.
PowerScale Administration-SSP1
A subnet can also be called the SmartConnet zone and contain one or more pools.
Pools enable more granular network configuration.
Multi-Tenancy Overview
SmartConnect: isilon.xattire.com
192.168.0.0/24
192.168.2.0/24
SmartConnect: isilon.gearitup.com
Groupnets are the configuration level for managing multiple tenants39 on the
external network of the cluster.
In the X-Attire scenario, the solution must treat each business unit as a separate
and unique tenant with access to the same cluster. The graphic shows how each
organization has its own groupnet and access zone.
PowerScale Administration-SSP1
Multi-tenancy Considerations
Groupnets are an option for those clusters that will host multiple companies,
departments, or clients that require their own DNS settings. Some areas to
consider are:
• DNS settings are per groupnet
• Create another groupnet only if separate DNS settings required.
• Follow proper build order:
1. Create groupnet
2. Configure authentication provider
3. Create access zone, and add authentication provider
4. Configure subnet with SmartConnect
5. Create pool, and add access zone
• In a multiple tenant solution, a share can span access zones. Combining
namespaces and overlapping shares is an administrative decision.
This video provides an overview of the groupnet and access zone relationship. See
the student guide for a transcript of the video.
PowerScale Administration-SSP1
Link:
https://edutube.emc.com/Player.aspx?vno=b4A2l5FzF2na/Txqk2AUTA==&autopla
y=true
Because groupnets are the top networking configuration object, they have a close
relationship with access zones and the authentication providers. Having multiple
groupnets on the cluster means that you are configuring access to separate and
different networks, which are shown as org1 and org2. Different groupnets enable
portions of the cluster to have different networking properties for name resolution.
Configure another groupnet if separate DNS settings are required. If necessary, but
not required, you can have a different groupnet for every access zone. The
limitation of 50 access zones enables the creation of up to 50 groupnets.
When the cluster joins an Active Directory server, the cluster must know which
network to use for external communication to the external AD domain. Because of
this, if you have a groupnet, both the access zone and authentication provider must
exist within same groupnet. Access zones and authentication providers must exist
within only one groupnet. Active Directory provider org2 must exist in within the
same groupnet as access zone org2.
PowerScale Administration-SSP1
The graphic shows the Cluster management > Network configuration > external network > Add
a groupnet window.
When creating a groupnet with access zones and providers in the same zone, you
need to create them in the proper order:
1. Create the groupnet.
2. Create the access zone and assign to the groupnet.
3. Create the subnet and pool.
4. Add the authentication provider and associate them with the groupnet
5. Associate the authentication providers with the access zone.
When creating a groupnet with access zones and providers in the same zone, you
should create them in the proper order.
PowerScale Administration-SSP1
PowerScale Administration-SSP1
Challenge
IT Manager:
Because you configure the network components together, you will not
go to the lab until the other topics are discussed. Open participation
question:
Question: When would you create a groupnet?
PowerScale Administration-SSP1
Scenario
This video provides an overview of SmartConnect. See the student guide for a
transcript of the video.
PowerScale Administration-SSP1
Link: https://edutube.emc.com/Player.aspx?vno=L7mXSvTcNQl8+LLKzNEzkw
SmartConnect provides name resolution for the cluster. The cluster appears as a
single network element to a client system. Both cluster and client performance can
be enhanced when connections are more evenly distributed.
In Isilon OneFS 8.2, SmartConnect supports connection service for 252 nodes.
SmartConnect Architecture
SmartConnect: isilon.xattire.com
SIP: 192.168.0.100 - 192.168.0.104
192.168.0.0/24
192.168.2.0/24
SmartConnect: isilon.gearitup.com
SIP: 192.168.2.100 - 192.168.2.104
The example shows two unique groups using the same cluster, X-Attire and GearItUp.
You can configure SmartConnect into multiple zones to provide different levels of
service for different groups of clients.
PowerScale Administration-SSP1
For example, SmartConnect directs X-Attire users to F800 nodes for their needed
performance. GearItUp users access the H500 nodes for general-purpose file
sharing. The zones are transparent to the users.
The SmartConnect Service IPs40 (SSIP or SIP) are addresses that are part of the
subnet.
SmartConnect Licensing
The table shows the differences between the SmartConnect basic and
SmartConnect Advanced.
40Do not put the SIPs in an address pool. The SIPs are a virtual IP within the
PowerScale configuration, it is not bound to any of the external interfaces.
PowerScale Administration-SSP1
The SIPs, SmartConnect zone, and the DNS entries are the configuration
components for SmartConnect.
PowerScale Administration-SSP1
This demonstration shows the initial network configuration for the cluster. See the
student guide for a transcript of the video.
Link: https://edutube.emc.com/Player.aspx?vno=4hL0i4iBe2BLqJzlT4dN/Q
In this demonstration, we’ll go through the step for an initial configuration of the
cluster external network. The demonstration shows configuring SmartConnect and
a dedicated pool for an access zone.
First, login to the WebUI and navigate to the Cluster management, Network
configuration page. The External network tab is selected by default. Note that
groupnet0 and subnet0 is automatically created by OneFS. On the subnet0 line,
PowerScale Administration-SSP1
select View / Edit. There are no values for SmartConnect. Select Edit. Go to the
SmartConnect service IPs and enter the range of SmartConnect IP addresses.
OneFS versions prior to OneFS 8.2 do not allow you to enter a range of IP
addresses. For this demonstration we will be using a SmartConnect service name.
Select Save changes. The CLI equivalent to add the SmartConnect service
address is the isi network subnet modify command. Now that
SmartConnect is configured, we will configure the IP address pool for the access
zone. On the subnet0 line, click on the More dropdown and select Add pool.
Enter the pool name and then select the access zone. For this implementation the
authentication providers and the access zones are already created.
Next enter the range of IP address for this pool. Select the external node interfaces
that will carry the client traffic. The SmartConnect basic fully qualified zone name is
sales.dees.lab. We have the SmartConnect advanced license activated. Here is
where we can configure the advanced functions. For the demonstration, we will
keep the default settings. Select Add pool. The CLI equivalent to create a pool is
the isi network pools create command.
This demonstration showed the initial configuration of the network. This concludes
the demonstration.
SmartConnect Considerations
PowerScale Administration-SSP1
• Static pools are best used for stateful clients, and dynamic pools are best for
stateless clients.
• Time-to-live value41.
Challenge
IT Manager:
Because you configure the network components together, you will not
go to the lab until the other topics are discussed. Open participation
question:
Question: What are the SmartConnect Advanced benefits?
PowerScale Administration-SSP1
IP Address Pools
Scenario
IP Address Pools
OneFS configures
groupnet0, subnet0,
pool0
Control connectivity to
access zones
More subnets are configured as either IPv4 or IPv6 subnets. Other IP address
pools are created within subnets and associated with a node, a group of nodes,
NIC ports or aggregated ports.
The pools of IP address ranges in a subnet enable you to customize42 how users
connect to your cluster.
PowerScale Administration-SSP1
Use case: Say that X-Attire adds 4 F800 nodes for a video media group. X-Attire
wants the video media team to connect directly to the F800 nodes to use various
high I/O applications. The administrators can separate the X-Attire connections.
Access to the home directories connect to the front end of the H500 nodes while
the video media group accesses the F800 nodes. This segmentation keeps the
home directory users from using bandwidth on the F800 nodes.
Link Aggregation
Physical NIC
Single
Logical NIC
Physical NIC
Aggregation combining the two physical interfaces into a single, logical interface.
43The link aggregation mode determines how traffic is balanced and routed among
aggregated network interfaces.
PowerScale Administration-SSP1
Click each tab to learn more about the link aggregation modes.
LACP
Configure LACP at the switch level and on the node. Enables the node to negotiate
interface aggregation with the switch.
PowerScale Node
Logical NIC
Physical NIC
Switch
Physical NIC
Round Robin
Round robin is a static aggregation mode that rotates connections through the
nodes in a first-in, first-out sequence, handling all processes without priority.
Round robin balances outbound traffic across all active ports in the aggregated link
and accepts inbound traffic on any port.
Client requests are served one after the other based on their arrival.
PowerScale Administration-SSP1
PowerScale Node
Rotates connections in a first-in, first-
out sequence
Logical NIC
1
3
Physical NIC 1 Incoming client requests
5 6 7 8
Physical NIC 2
4
2
The graphic shows, client request 2, client request 3 and so on follow client request 1.
Note : Round Robin is not recommended if the cluster is using TCP/IP workloads.
Failover
Active/Passive failover is a static aggregation mode that switches to the next active
interface when the primary interface becomes unavailable. The primary interface
handles traffic until there is an interruption in communication. At that point, one of
the secondary interfaces takes over the work of the primary.
1 2 3 5 6
Physical NIC 1
4
Physical NIC 2
In the graphic, the nodes serve the incoming client requests. If any of the nodes become
unavailable or interrupted due to an issue, the next active node takes over and serves the upcoming
client request.
FEC
Typically used with older Cisco switches - LACP preferred in new generation
PowerScale nodes.
PowerScale Administration-SSP1
FEC accepts all incoming traffic and balances outgoing traffic over aggregated
interfaces that is based on hashed protocol header information that includes source
and destination addresses.
Outgoing traffic
Logical NIC
Incoming client requests 3 1
2
Physical NIC 1
6 5 4 3 2 1
6 5 4
Physical NIC 2
The graphic shows, the node accepts and serves all the incoming client requests. The node
balances outgoing traffic.
PowerScale Administration-SSP1
Allocation Method
PowerScale Administration-SSP1
Static
If there are more IP addresses than nodes, new nodes that are added to the pool
get the additional IP addresses.
Once allocating an IP address, the node keeps the address indefinitely unless
deleting the member interface from the pool, or removing the node from the cluster.
PowerScale Administration-SSP1
Dynamic
Dynamic pools are best used for stateless protocols such as NFSv3. Also configure
for NFSv4 with continuous availability (CA).
PowerScale Administration-SSP1
The graphic shows a two SmartConnect zones, each with different IP allocation methods.
Static pools are best used for SMB clients because of the stateful nature of the
SMB protocol.
Dynamic pools are best used for stateless protocols such as NFSv3. You can
identify a Dynamic range by the way the IP addresses present in the interface as
.110 -.112 or .113 -.115 instead of a single IP address like 0.10.
Challenge
PowerScale Administration-SSP1
PowerScale Administration-SSP1
Scenario
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 97
Configuring Identity Management and Authorization
Overview
A user who is assigned to more than one role has the combined privileges of those
roles.
The root and admin users can assign others to built-in or custom roles that have
login and administrative privileges to perform specific administrative tasks.
The example shows that user Jane is assigned the Backup Administrator role.
Many of the privileges that user Root has are not visible to user Jane.
Role-based access enables you to separate out some administrative privileges and
assign only the privileges that a user needs. Granting privileges makes access to
the configuration of the cluster less restrictive.
Roles
OneFS includes built-in administrator roles with predefined sets of privileges that
you cannot modify. You can also create custom roles and assign privileges. Click
the tabs to learn more about each role.
PowerScale Administration-SSP1
Built-in Roles
Built-in roles44 are included in OneFS and have been configured with the most
likely privileges necessary to perform common administrative functions.
44You cannot modify the list of privileges that are assigned to each built-in role.
However, you can assign users and groups to built-in roles.
47The AuditAdmin built-in role enables you to view all system configuration
settings.
48 The BackupAdmin built-in role enables backup and restore of files from /ifs.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 99
Configuring Identity Management and Authorization
Custom roles
You can create custom roles50 and assign privileges mapped to administrative
areas in your PowerScale cluster environment.
The following list describes what you can and cannot do through roles:
• You can assign privileges to a role but not directly to users or groups.
• You can create custom roles and assign privileges to those roles.
• You can copy an existing role.
• You can add any user or group of users, to one or more roles as long as the
users can authenticate to the cluster.
PowerScale Administration-SSP1
The video provides an overview of role creation. See the student guide for a
transcript of the video.
Link: https://edutube.emc.com/Player.aspx?vno=tQkWrNubtdORFBHxoRlMAg
Login as admin, a user that can assign privileges. Navigate to Access, Membership
and roles. On the Membership and roles page, note that the access zone selected
is System. Go to the Roles tab. Before moving on to the configuration, note that
OneFS has a number of built-in roles that cover most access needs. There may be
a need to define a custom role. In these instances, you can select the Create a
Role button. I will demonstrate this in a moment. A great place to learn more about
the different privileges is the Isilon OneFS Web Administration Guide.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 101
Configuring Identity Management and Authorization
The next example is to add a Windows administrator, Sai, to the sales access
zone. Adding Sai to a role specific to the access zone prevents him from
accidentally configuring Windows shares in other zones. In fact, Sai will have no
visibility into other zones. On the Roles tab, select the sales access zone. Note the
two built-in roles really do not provide the level of access for Sai. Create a role. The
role name is WinAdmin and add a short description. Shown is the CLI command to
create a zone role. Remember OneFS version 8.2 introduces zone-aware roles.
Previous version CLI commands do not have the --zone option. boston-2# isi
auth roles create --zone sales WinAdmin. Just as in the previous
example, add a member to this role. Select the provider and then the domain. Next
Search and select Sai. Now add privileges to the role. First, add the ability to log in
to the WebUI. Next, add the privilege to configure SMB. Give Read/write access to
this privilege. Now save the role. boston-2# isi auth roles modify
WinAdmin --zone sales --add-priv ISI_PRIV_LOGIN_PAPI --add-
priv ISI_PRIV_SMB –-add-user dees\\sai. Now verify the privileges of
the users.
Logout and then log in as Hayden, the AuditAdmin. The first indication is the
Access menu. Notice the options are missing. Navigating to Protocols, Windows
sharing, notice Hayden cannot create a share, only view. Also, since added to a
System zone role, Hayden can audit information in other zones. System zone
administrators are global.
Log out of the WebUI and login as Sai. You must login at an IP address or netBios
associated with the sales access zone. Viewing the Access options, Sai does not
have the privileges. Navigating to Protocols, Windows sharing, notice Sai cannot
switch to another access zone, but can configure SMB shares. This demonstration
stepped through configuring RBAC and ZRBAC. This concludes the demonstration.
PowerScale Administration-SSP1
Role Management
You can view, add, or remove members of any role. Except for built-in roles, whose
privileges you cannot modify, you can add or remove OneFS privileges on a role-
by-role basis.
View Roles
Command Description
isi auth roles list A basic list of all roles on the cluster
isi auth roles view <role> Detailed information about a single role,
where <role> is the name of the role
View Privileges
User Privileges are performed through the CLI. The table shows the commands
that can view a list of your privileges or of another user.
Command Description
isi auth mapping token <user> List of privileges for another user,
where <user> is a placeholder for
another user by name:
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 103
Configuring Identity Management and Authorization
You can create an empty custom role and then add users and privileges to the role.
Deleting a role does not affect the privileges or users that are assigned to it. Built-in
roles cannot be deleted.
The table shows the commands used to create, modify and delete a custom role.
Command Description
isi auth roles create <name> [-- To create a role, where <name> is
description <string>] the name that you want to assign
to the role and <string> specifies
an optional description
isi auth roles modify <role> [-- To add a user to the role, where
add-user <string>] <role> is the name of the role and
<string> is the name of the user
isi auth roles modify <role> [-- To add a privilege with read/write
add-priv <string>] access to the role, where <role> is
the name of the role and <string>
is the name of the privilege
isi auth roles modify <role> [-- To add a privilege with read-only
add-priv-ro <string>] access to the role, where <role> is
the name of the role and <string>
is the name of the privilege
Privileges
PowerScale Administration-SSP1
List privileges
The graphic shows built-in roles that have a predefined set of privileges. Red outlines are the only
privileges available for ZRBAC.
Note: The WebUI privileges names differ from the names that are
seen in the CLI.
ZRBAC - ISI_PRIV_AUTH_Privilege
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 105
Configuring Identity Management and Authorization
1: If zone is created by the system zone admins, only the system zone admins can
modify and delete. Local zone admin can only view and add access zones.
If zone is created by a nonsystem zone admin, both the zone admin and
nonsystem zone admin can view, modify, and delete.
4: The IP address in the IP address pool associated with the access zone.
Challenge
Lab Assignment: Go to the lab and create user accounts for RBAC and
ZRBAC.
PowerScale Administration-SSP1
Scenario
Your Challenge: The IT manager has tasked you to determine the on-
disk identity to configure on the cluster. Before configuring, you should
have an understanding of how identity management works. The
manager expects you to describe identity management, user tokens,
and on-disk identity.
Layers of Access
Identity Assignment
Based on authentication or
mediated in cluster
Cluster connectivity has four layers of interaction. The third layer is identity
assignment. The layer is straightforward and based on the results of the
authentication layer.
There are some cases that need identity mediation within the cluster, or where
roles are assigned within the cluster that are based on user identity.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 107
Configuring Identity Management and Authorization
Identity Management
The OneFS identity management maps the users and groups from separate
services. The mapping provides a single unified identity on a cluster and uniform
access control to files and directories, regardless of the incoming protocol. Click on
the "i" icons for a high-level information about the process.
PowerScale Administration-SSP1
2: The authentication providers uses OneFS to first verify a user identity after which
users are authorized to access cluster resources. The top layers are access
protocols – NFS for UNIX clients, SMB for Windows clients, and FTP and HTTP for
all.
3: Between the protocols and the lower-level services providers and their
associated data repositories, is the OneFS lsassd daemon. lsassd mediates
between the authentication protocols that clients and the authentication providers,
who check their data repositories for user identity and file access, use.
The video describes the access token generation. See the student guide for a
transcript of the video.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 109
Configuring Identity Management and Authorization
URL:
https://edutube.emc.com/Player.aspx?vno=MmSHIH1OvcP5nHsi0hd51g==&autopl
ay=true
When the cluster receives an authentication request, the lsassd searches the
configured authentication sources for matches to the incoming identity. If the
identity is verified OneFS generates an Access Token. Access Token form basis of
who you are when performing actions on the cluster. Shown is the output of the
users mapping token. The token supplies the primary owner and group identities to
use during file creation. For most protocols the access token is generated from the
PowerScale Administration-SSP1
user name or from the authorization data that is received during authentication.
Access tokens are also compared against permissions on an object during
authorization checks. The access token includes all identity information for the
session OneFS exclusively uses the information in the token when determining if a
user has access to a particular resource.
Access tokens form the basis of who you are when performing actions on the
cluster. The tokens supply the primary owner and group identities to use during file
creation. When the cluster builds an access token, it must begin by looking up
users in external directory services. By default, the cluster matches users with the
same name in different authentication providers and treats them as the same user.
The ID-mapping service populates the access token with the appropriate identifiers.
Finally, the on-disk identity is determined.
Primary Identities
OneFS supports three primary identity types, UIDs, GIDs, and SIDs.
UIDs and GIDs from Local, NIS, LDAP providers range from 1 to 65k.
OneFS automatically allocates UIDs and GIDs from the range 1,000,000-
2,000,000.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 111
Configuring Identity Management and Authorization
1
2
1: The user identifier, or UID, is a 32-bit string that uniquely identifies users on the
cluster. UNIX-based systems use UIDs for identity management.
2: The security identifier, or SID, is a unique identifier that begins with the domain
identifier and ends with a 32-bit Relative Identifier (RID). Most SIDs take the form
S-1-5-21-<A>-<B>-<C>-<RID>, where <A>, <B>, and <C> are specific to a domain
or system, and <RID> denotes the object inside the domain. SID is the primary
identifier for users and groups in Active Directory.
3: The group identifier, or GID, for UNIX serves the same purpose for groups that
UID does for users.
Secondary Identities
PowerScale Administration-SSP1
1: Windows provides a single namespace for all objects that is not case-sensitive,
but specifies a prefix that targets the dees Active Directory domain. UNIX assumes
unique case-sensitive namespaces for users and groups. For example, Sera and
sera can represent different objects.
2: Kerberos and NFSv4 define principals that require all names to have a format
similar to an email address. For example, given username sera and the domain
dees.lab, dees\sera and [email protected] are valid names for a single object in
Active Directory. With OneFS, whenever providing a name as an identifier, the
correct primary identifier of UID, GID, or SID is requested.
Multiple Identities
The graphic shows a user that has both a Windows and Linux account. Multiple
identity, or multiprotocol access, could include configuring mapping to ensure user
IDs correctly map to one another.
OneFS is RFC 2307 compliant. Enable RFC 2307 to simplify user mapping.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 113
Configuring Identity Management and Authorization
See the participant guide for information about mapping challenges and
considerations.
ID Mapper Database
1 3
1: The user mapper feature can apply rules to modify the user identity OneFS
uses, add supplemental user identities, and modify the group membership of a
user. The user mapping service combines user identities from different directory
services into a single access token. The mapping service then modifies it according
to the rules that you create.
PowerScale Administration-SSP1
On-Disk Identity
Identifies preferred
identity to store on
disk
Determines identity
stored in ACLs - SID
or UID/GIDs
The graphic shows the token of Windows user Sera with a UID as the on-disk identity.
OneFS uses an on-disk identity store for a single identity for users and groups.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 115
Configuring Identity Management and Authorization
The available on-disk identity types are Native, UNIX, and SID. The on-disk identity
is a global setting. Because most protocols require some level of mapping to
operate correctly, choose the preferred identity to store on-disk.
The use case for the default Native setting is an environment that has NFS and
SMB client and application access. With the Native on-disk identity set, lsassd
attempts to locate the correct identity to store on disk by running through each ID-
mapping method. The preferred object to store is a real UNIX identifier. OneFS
uses a real UNIX identifier when found. If a user or group does not have a real
UNIX identifier (UID or GID), OneFS stores the real SID. Click on the highlighted
icon to learn more.
Troubleshooting Resources
Challenge
PowerScale Administration-SSP1
Authorization
Scenario
Permissions Overview
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 117
Configuring Identity Management and Authorization
3
4
6
1
5
1: OneFS supports NFS and SMB protocols. It accesses the same directories and
files with different clients.
6: OneFS supports two types of authorization data on a file, access control lists, or
ACLs, and UNIX permissions, or POSIX mode bits.
PowerScale Administration-SSP1
The internal representation, which can contain information from either the POSIX
mode bits or the ACLs, is based on RFC 3530.
POSIX Overview
53A file can only be in one of the states at a time. That state is authoritative. The
actual permissions on the file are the same, regardless of the state.
55 OneFS must store an authoritative version of the original file permissions for the
file sharing protocol and map the authoritative permissions for the other protocol.
OneFS must do so while maintaining the security settings for the file and meeting
user expectations for access. The result of the transformation preserves the
intended security settings on the files. The result also ensures that users and
applications can continue to access the files with the same behavior.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 119
Configuring Identity Management and Authorization
1 2 3
4 5
2: Group permissions
4: Configure permission flags to grant read (r), write (w), and execute (x)
permissions to users, groups, and others in the form of permission triplets. The
classes are not cumulative. OneFS uses the first class that matches. Typically,
grant permissions in decreasing order, giving the highest permissions to the file
owner and the lowest to users who are not the owner or the owning group.
5: These permissions are saved in 16 bits, which are called mode bits.
6: The information in the upper 7 bits can also encode what the file can do,
although it has no bearing on file ownership. An example of such a setting would
be the “sticky bit.”
PowerScale Administration-SSP1
Triplets
9 mode bits
Triplet classes
Modify UNIX permissions in the WebUI on the File system > File system explorer page. Click
image to enlarge.
The graphic shows root user who is logged in and the /ifs/boston/hr
directory. Only root user can view and edit the owner and group of the object.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 121
Configuring Identity Management and Authorization
chmod Command
Changing the permissions on a directory so that group members and all others can only read the
directory.
OneFS supports the standard UNIX tools for changing permissions: chmod and
chown. The change mode command, chmod, can change permissions of files and
directories. The man page for chmod documents all options.
Changes that are made using chmod can affect Windows ACLs.
chown Command
The output shows that penni is an LDAP user who is responsible for the content of the
/ifs/boston/hr directory.
The chown command is used to change ownership of a file. Changing the owner of
a file requires root user access. The basic syntax for chown is chown [-R]
PowerScale Administration-SSP1
newowner filenames. Using the -R option changes the ownership on the sub
directories.
The chgrp command changes the group. View the man pages for command
definitions.
Access control
elements
No permissions = no access
List of advanced
List of basic permissions
permissions
On Windows host: Properties > Security tab > Advanced > Edit window
While you can apply permissions for individual users, Windows administrators
usually use groups to organize users, and then assign permissions to groups
instead of individual users.
Windows includes many rights that you can assign individually or you can assign
rights that are bundled together as permissions. For example, the Read permission
includes the rights to read and execute a file while the Full Control permission
assigns all user rights. Full Control includes the right to change ownership and
change the assigned permissions of a file or folder.
When working with Windows, note the important rules that dictate the behavior of
Windows permissions. First, if a user has no permission that is assigned in an ACL,
then the user has no access to that file or folder. Second, permissions can be
explicitly assigned to a file or folder and they can be inherited from the parent
folder. By default, when creating a file or folder, it inherits the permissions of the
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 123
Configuring Identity Management and Authorization
OneFS has configurable ACL policies that manage permissions. You can change
the default ACL settings globally or individually, to best support the environment.
The global permissions policies change the behavior of permissions on the system.
For example, selecting UNIX only changes the individual ACL policies to
correspond with the global setting. The permissions settings of the cluster are
handled uniformly across the entire cluster, rather than by each access zone.
The WebUI > Access > ACL policy settings page and how the policy settings
translate in the CLI command output. You can also use the "isi auth settings acls
modify" command to configure the ACL settings.
1
2
3
4
2: Use case: Permissions operate with UNIX semantics - prevents ACL creation.
3: Use case: Permissions operate with Windows semantics - errors for UNIX
chmod.
PowerScale Administration-SSP1
1 4
1: The ls -le command shows actual permissions stored on disk and ACL from
security descriptor.
2: The ls -len command shows numerical (n) owner and group SID or UID/GID.
4: The long format includes file mode, number of links, owner, group, MAC label,
number of bytes, abbreviated month, day file last modified, hour file last modified,
minute file last modified, and the path name.
OneFS takes advantage of standard UNIX commands and has enhanced some
commands for specific use with OneFS.
The list directory contents, ls, command provides file and directory permissions
information, when using an SSH session to the cluster. PowerScale has added
specific options to enable reporting on ACLs and POSIX mode bits.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 125
Configuring Identity Management and Authorization
Tip: The ls command options are all designed for long notation
format, which is displayed when the -l option is used. The -l
option also displays the actual permissions that are stored on disk.
Running the ls -le command shows the synthetic ACLs for files and directories (the -d flag lists
directory entries).
A Windows client processes only ACLs, it does not process UNIX permissions.
When viewing the permission of a file from a Windows client, OneFS must translate
the UNIX permissions into an ACL.
If a file has Windows-based ACLs (and not only UNIX permissions), OneFS
considers it to have advanced, or real ACLs56.
56Advanced ACLs display a plus (+) sign when listed using an ls –l, or as shown,
the ls -led command. POSIX mode bits are present when a file has a real ACL,
however these bits are for protocol compatibility and are not used for access
checks.
PowerScale Administration-SSP1
The video discusses authentication and authorization. See the student guide for a
transcript of the video.
Link:
https://edutube.emc.com/html5/videoPlayer.htm?vno=EN8uMS3WuRwjY4Q0mIUa
Zw
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 127
Configuring Identity Management and Authorization
corporate building, thus the user has permission to enter. Share level permissions
work similarly in that users get access to the share before they can gain access to
any of the share directories. A user that has access to a directory (office) can then
access the files within the directory, providing permission to the file is given.
Two options are available when creating a share, Do not change existing
permissions and Apply Windows default ACLs. Understand the Apply
Windows default ACLs settings. This setting can destroy or at a minimum alter
explicitly defined directory permissions that are created on the share. For example,
carefully migrated permissions can change, creating more work and the potential of
causing data unavailability. Files and directories can be either POSIX authoritative
or ACLs authoritative.
A synthetic ACL does not exist on the file system and is not stored anywhere.
Instead, OneFS generates a synthetic ACL as needed, and then discards it. OneFS
creates the synthetic ACL in memory when a client that only understands ACLs,
such as Windows clients, queries the permissions on a file that only has POSIX
permissions.
With synthetic ACLs, POSIX mode bits are authoritative. POSIX mode bits handle
permissions in UNIX environments and govern the synthetic ACLs. Permissions
are applied to users, groups, and everyone, and allow or deny file and directory
access as needed. The read, write, and execute bits form the permissions triplets
for users, groups, and everyone. The mode bits can be modified using the WebUI
or the CLI standard UNIX tools such as chmod and chown. Since POSIX governs
the synthetic ACLs, changes made using chmod change the synthetic ACLs. For
example, running chmod 775 on the /ifs/dvt directory changes the mode bits to
read-write-execute for group, changing the synthetic ACL for the group. The same
behavior happens when making the access more restrictive, for example, running
chmod 755, changes the synthetic ACL to its corresponding permission. The
chmod behavior is different when ACLs are authoritative.
PowerScale Administration-SSP1
In the example, the directory /ifs/dvt/win has a real ACL. The POSIX mode bits are
775. Running chmod 755 does not change to the POSIX mode bits since merging
775 with 755 gives the combined value of 775. Shown is an excerpt from the Isilon
cluster WebUI page that shows the different behaviors.
The first example shows that the share permission is everyone read-only although
the POSIX indicates read-write-execute. Windows users can write to the share
based on the synthetic ACLs. The second example shows POSIX at 755. Although
the ACL is set to a user with full control, the user cannot write to the share—POSIX
is authoritative.
The “+” indicates a real or native ACL that comes directly from Windows and is
applied to the file. Access control entries make up Windows ACLs. An administrator
can remove the real ACL permission using the chmod -b command. ACLs are
more complex than mode bits and can express a richer set of access rules.
However, not all POSIX mode bits can represent Windows ACLs any more than
Windows ACLs can represent POSIX mode bits.
Once a file is given an ACL, its previous POSIX mode bits are no longer
enforced—the ACL is authoritative. The first example shows a real ACL used,
POSIX set for 777, and the share permissions for the user set to read-only.
Although the POSIX show read-write-execute for everyone, the user cannot write
because of the ACL. In contrast, the second example shows the case where the
user can write.
Troubleshooting Resources
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 129
Configuring Identity Management and Authorization
Challenge
Lab Assignment:
Log in to the cluster and verify the ACL policy setting.
• Permissions and ownership using the WebUI
• Permissions and ownership using the CLI
• ACL authoritative
• ACL policy setting
PowerScale Administration-SSP1
PowerScale Administration-SSP1
OneFS Caching
Scenario
IT Manager: The next thing that I would like to know more about is how
the PowerScale caches data.
4
2
1
3
2: Accelerate access. The immediacy determines how the cache is refreshed, how
long the data is available, and how the data is emptied or flushed from cache.
3: Different cache levels to account for differing data immediacy. The cache levels
provide guidance to the immediacy of information from a client-side transaction
perspective.
PowerScale Administration-SSP1
4: Cache is temporary. Because cache is a copy of the metadata and user data,
any data that is contained in cache is temporary and can be discarded when no
longer needed.
Caching maintains a copy of the metadata57 and/or the user data blocks in a
location other than primary storage.
Cache in OneFS is divided into levels. Each level serves a specific purpose in read
and write transactions.
Cache Levels
OneFS caching consists of the client-side level 1, or L1, cache and write coalescer,
and level 2, or L2 storage and node-side cache.
Both L1 cache and L2 cache are managed and maintained in RAM. However,
OneFS is also capable of using SSDs as level 3, or L3 cache.
57
The copy is used to accelerate access to the data by placing the copy on a
medium with faster access than the drives.
PowerScale Administration-SSP1
Each cache has its own specialized purpose and works together to provide
performance improvements across the entire cluster.
L1 Cache
Client-side cache.
1: L1 cache allows all blocks for immediate read requests. Read cache is flushed
after a successful read transaction and write cache is flushed after a successful
write transaction. L1 cache collects the requested data from the L2 cache of the
nodes that contain the data.
L1 cache is the client-side cache. It is the buffer on the node that the client
connects, and is involved in any immediate client data transaction.
The write coalescer collects the write blocks and performs the additional process of
optimizing the write to disk.
L2 Cache
L2 cache.
PowerScale Administration-SSP1
1: L2 cache is also contained in the node RAM. It is fast and available to serve L1
cache read requests and take data handoffs from the write coalescer. L2 cache
interacts with the data that is contained on the specific node. The interactions
between the drive subsystem, the HDDs, and the SSDs on the node go through the
L2 cache for all read and write transactions.
L2 cache is the storage side or node-side buffer. L2 cache stores blocks from
previous read and write transactions.
L3 Cache
L3 cache.
PowerScale Administration-SSP1
1: Extension of L2 cache.
2: SSD access is slower than access to RAM and is relatively slower than L2 cache
but faster than access to data on HDDs. L3 cache is an extension of the L2 read
cache functionality. Because SSDs are larger than RAM, SSDs can store more
cached metadata and user data blocks than RAM. When L3 cache becomes full
and new metadata or user data blocks are loaded into L3 cache, the oldest existing
blocks are flushed from L3 cache. Flushing is based on first in first out, or FIFO. L3
cache should be filled with blocks being rotated as node use requires.
L3 cache provides additional level of storage node-side cache using the SSDs as
read cache.
Good for random, read heavy workflows accessing the same data sets.
PowerScale Administration-SSP1
The graphic shows an eight node cluster that is divided into two node pools with a
detailed view of one of the nodes.
3 1
4
1: Clients connect to L1 cache and the write coalescer. The L1 cache is connected
to the L2 cache on the other nodes and within the same node. The connection to
other nodes occurs over the internal network when data that is contained on those
nodes is required for read or write.
2: The L2 cache on the node connects to the disk storage on the same node. The
L3 cache is connected to the L2 cache and serves as a read-only buffer. The L2
cache on the node connects to the disk storage on the same node.
5: Backend network.
Anatomy of a Read
When a client requests a file, the client-connected node uses the isi get
command to determine where the blocks that comprise the file are located.
PowerScale Administration-SSP1
1: The first file inode is loaded, and the file blocks are read from disk on all other
nodes. If the data is not present in the L2 cache, data blocks are copied in the L2.
The blocks are sent from other nodes through the backend network.
2: If the data is already present in L2 cache, it is not loaded from the hard disks.
OneFS waits for the data blocks from the other nodes to arrive. Otherwise, the
node gets the data load from the local hard disks, and then the file is reconstructed
in L1 cache and sent to the client.
When a client requests a file write to the cluster, the client-connected node
receives and processes the file.
PowerScale Administration-SSP1
2 1
1: Cache writes until write coalescer is full, time limit is reached, or protocol
requests confirmation of delivery.
2: The client-connected node creates a write plan for the file including calculating
Forward Error Correction, or FEC. Data blocks assigned to the node are written to
the journal of that node. Data blocks assigned to other nodes travel through the
internal network to their L2 cache, and then to their journal.
At same time, data blocks that are assigned to other nodes go to L2.
3: Once all nodes have all the data and FEC blocks that are journaled, a commit is
returned to the client. Data blocks assigned to client-connected node stay cached
in L2 for future reads, and then data is written onto the HDDs.
4: The Block Allocation Manager, or BAM, on the node that initiated a write
operation makes the layout decisions. The BAM decides on where best to write the
data blocks to ensure that the file is properly protected. Data is copied to journal.
To decide, the BAM Safe Write, or BSW, generates a write plan, which comprises
all the steps that are required to safely write the new data blocks across the
protection group.
5: Once nodes have the data and FEC is journaled, nodes confirmation is sent to
client-connected node and a commit is sent to client.
6: Once complete, the BSW runs this write plan and guaranties its successful
completion. OneFS does not write files at less than the desired protection level.
Data is written to disks.
PowerScale Administration-SSP1
L3 Cache Settings
L3 cache is enabled by default for all new node pools that are added to a cluster.
L3 cache is either on or off and no other visible configuration settings are available.
File system > Storage pools > SmartPools settings. Enabling and disabling L3 at
the global level and at the node pool level.
2: L3 cache cannot enable if node pool has no unprovisioned SSDs and it cannot
coexit with other SSD strategies.
CLI Commands
The following command are used to disable globally and to enable at the node pool
level.
L3 Cache Considerations
PowerScale Administration-SSP1
• L3 cache cannot co-exist with other SSD strategies58 on the same node pool.
• SSDs in an L3 cache enabled node pool cannot participate as space used for
GNA.
• L3 acts as an extension of L2 cache regarding reads and writes59 on a node.
• You cannot enable L3 cache in all-flash nodes60.
• You cannot disable L3 cache in archive-type nodes (A200, A2000, NL410,
HD400).
• If changing the L3 cache behavior, migrating data and metadata from the SSDs
to HDDs can take hours.
The example shows the command to query historical statistics for cache. The first
command lists the keys that are related to cache.
A use case is, running the command to determine the L3 hit and miss stats to
indicate if the node pool needs more SSDs.
60 On Gen 6x nodes all data drives are SSDs in the F800, F810, F600, and F200.
PowerScale Administration-SSP1
1: The command lists the keys that are related to cache. The number and
granularity of available keys is numerous. The keys give administrators insight to
the caching efficiency and can help isolate caching related issues.
2: The command shows the key to list the L1 metadata read hits for node 2, the
node that is connected over SSH.
Challenge
IT Manager:
Open participation question:
Question: What does L1, L2, and L3 cache provide?
PowerScale Administration-SSP1
SMB Shares
Scenario
IT Manager: The first thing that I would like you to configure is an SMB
share for the Windows users. I want you to create a single share for
now, and ensure that the Windows users have access.
Your Challenge: The IT manager has tasked you to create a share that
the Windows users can access. Before creating the shares, you must
know a few things. The manager wants you ensure you can describe
SMB Continuous Availability, enable SMB sharing, and create shares
and home directories.
Protocol Overview
Configure and create SMB shares for Windows users - created at the zone
level
PowerScale Administration-SSP1
Network or Node
failure
Old behavior: If this node goes down or a network interruption, the client
needs to reconnect to the cluster manually.
SMB shares provide Windows clients network access to file system resources on
the cluster.
Too many disconnections prompt the clients to open help desk tickets with their
local IT department to determine the nature of the data unavailability.
Clients using SMB 1.0 and SMB 2.x use a time-out service.
PowerScale Administration-SSP1
Server-side copy offloads copy operations to the server when the involvement of
the client is unnecessary.
File data no longer traverses the network for copy operations that the server can
perform.
The server-side copy feature is enabled by default. To disable the feature, use the
CLI.
61Advanced algorithms are used to determine the metadata and user data blocks
that are cached in L3. L3 cached data is durable and survives a node reboot
without requiring repopulating.
PowerScale Administration-SSP1
/ifs/finance/
Enabled by default
Server-side copy is data
disabled
Network
Copied data
traverses the
network Server-side copy is
To enable SMB, in the WebUI, go to the Protocols > Windows sharing (SMB) > SMB server
settings tab.
PowerScale Administration-SSP1
The SMB server settings page contains the global settings that determine how the
SMB file sharing service operates.
These settings include enabling or disabling support for the SMB service.
A case62 for disabling the SMB service is when testing disaster readiness.
This video demonstrates the process of creating an SMB share, mapping the
share, and verifying access. See the student guide for a transcript of the video.
62 The organization fails over the production cluster or directory to a remote site.
When the remote data is available and users write to the remote cluster, all SMB
traffic should be halted on the production site. Preventing writes on the production
site prevents data loss when the remote site is restored back to the production site.
PowerScale Administration-SSP1
Link:
https://edutube.emc.com/html5/videoPlayer.htm?vno=aMwue+nqUbFdOFoqKa98F
g
This demonstration shows the steps to configure SMB shares. Log in to the WebUI
as admin. The dashboard shows all the cluster nodes are healthy. The cluster is
running OneFS 8.2. Navigate to Protocols, Windows sharing. The SMB share will
be in the marketing access zone. Select Create an SMB share. The share I am
creating is called “general purpose”. I will add a description. The path
/ifs/marketing/GeneralPurpose does not exist so I will ensure it is created. This is a
Windows only share that did not previously exist so I will select Apply Windows
default ACLs. In the Members table I will give Everyone full control and then Create
share. The next step is to access the share from a Windows client. From the
Windows client, I will open Windows Explorer and map the share. Good. Now as a
simple test I am creating a text document. I will write some content and save. And
then I will open the document. This demonstration stepped through configuring,
mapping, and accessing an SMB share.
PowerScale Administration-SSP1
Share Creation
Settings Section
The CLI equivalent are the isi smb shares create or isi smb shares modify commands.
Type the full path of the share in the path field, beginning with /ifs.
You can also browse to the share. If the directory does not exist, the Create SMB
share directory if it does not exist creates the required directory.
PowerScale Administration-SSP1
Directory ACLs
Use caution when applying the default ACL settings as it may overwrite existing
permissions in cases where the data has been migrated onto the cluster.
When a cluster is set up, the default permissions on /ifs may or may not be
appropriate for the permissions on your directories.
PowerScale Administration-SSP1
Summary63
OneFS supports the automatic creation of SMB home directory paths for users.
631) If adding a share to an existing directory structure, you likely do not want to
change the ACL, so select the Do not change existing permissions. 2) If creating a
share for a new directory, you will likely be changing permissions to the ACL to
grant Windows users rights to perform operations. Set the Apply Windows default
ACLs and then once the share is created, go into the Windows Security tab and
assign permissions to users as needed.
PowerScale Administration-SSP1
Variables:
• %L64
• %D65
• %U66
• %Z67
67%Z expands to the access zone name. If multiple zones are activated, this
variable is useful for differentiating users in separate zones.
PowerScale Administration-SSP1
The graphic shows the permissions that are changed to Full control.
Adjustments made to Advanced settings override the default settings for this
share only.
PowerScale Administration-SSP1
You can make access zone global changes to the default values in the Default
share settings tab. Changing the default share settings is not recommended.
In the CLI, you can create shares using the isi smb shares create
command. You can also use the isi smb shares modify to edit a share and
isi smb shares list to view the current Windows shares on a cluster.
The share name can contain up to 80 characters, and can only contain
alphanumeric characters, hyphens, and spaces. The description field contains
basic information about the share. There is a 255-character limit. Description is
optional but is helpful when managing multiple shares.
Example for directory ACLs: Say that /ifs/eng is a new directory that was created
using the CLI. Windows users can create and delete files in the directory. When
creating the share, if the Do not change existing permissions is set and then users
attempt to save files to the share, an access denied occurs because Everyone has
read access. Even as an administrator you cannot modify the security tab of the
directory to add Windows users because the mode bits limit access to only Root.As
an example, /ifs/eng is and NFS export and you explicitly want the /ifs/eng mode bit
rights set based on UNIX client application requirements. Selecting the Apply
Windows default ACLs option as shown in the graphic, overwrites the original
ACLs, which can break the application. Thus, there is risk that is associated with
using Apply Windows default ACLs with an existing directory.
Example for home directories: To create a share that automatically redirects users
to their home directories, select the Allow variable expansion box. To automatically
create a directory for the user, check the Auto-create directories box. You may also
set the appropriate flags by using the isi smb command in the command-line
interface. In the graphic, 1) set up user access to their home directory by mapping
to /ifs/finance/home. Users are automatically redirected to their home directory
/ifs/finance/home/. 2) Expansion variables are used to automatically create a path
where the users store the home directory files. After the creation, users connecting
to this share are automatically redirected to their home directory according to the
used path variables. The access zone is implied, because all access for Active
Directory is done per access zone and each access zone has its own home
directory path.
PowerScale Administration-SSP1
Challenge
Lab Assignment: Now log in to the cluster and create home directories
and a general purpose share.
PowerScale Administration-SSP1
NFS Exports
Scenario
IT Manager: Now that you have the Windows users able to access the
cluster, you configure access for the linux users. I want you to create an
export that the linux users can access. Have a good understanding of
NFS exports before implementing into the lab.
NFS Overview
1
3
1: NFS relies upon remote procedure call (RPC) for client authentication and port
mapping.
2: NFS is native to UNIX clients. You can configure NFS to enable UNIX clients to
access content stored on PowerScale clusters.
PowerScale Administration-SSP1
Exporting a directory enables accessing the data that is hosted on the cluster.
Node or network
issue
CA is enabled by default.
Clients transparently fail over to another node when a network or node fails.
To enable and disable NFS using the WebUI, click Protocols > UNIX sharing (NFS)
> Global settings tab.
PowerScale Administration-SSP1
1 2
If changing a value in the Export settings, that value changes for all NFS exports in
the access zone. Modifying the access zone default values is not recommended.
You can change the settings for individual NFS exports as you create them, or edit
the settings for individual exports as needed.
2: Enabling NFSv4 requires entering the domain in the Zone settings page.
If NFSv4 is enabled, specify the name for the NFSv4 domain in the NFSv4 domain
field on the Zone setting page.
You can customize the user/group mappings, and the security types (UNIX and/or
Kerberos), and other advanced NFS settings.
The NFS global settings determine how the NFS file sharing service operates. The
settings include enabling or disabling support for different versions of NFS.
Enabling NFSv4 is nondisruptive, and it runs concurrently with NFSv3. Enabling
NFSv4 does not impact any existing NFSv3 clients.
PowerScale Administration-SSP1
Configuration steps on the UNIX sharing (NFS) page have the possibilities to
reload the cached NFS exports configuration to ensure that any DNS or NIS
changes take effect immediately.
Create and manage NFS exports using either the WebUI or the CLI. For the CLI,
use the isi nfs exports command.
Protocols > UNIX sharing (NFS) > NFS exports page, Create an export option.
Highlighted are the paths to export.
3: Specifying no clients allows all clients on the network access to the export.
4: Rule order of precedence: Root clients, always read/write clients, Always read-
only clients, and then clients.
PowerScale Administration-SSP1
You can enter a client by host name, IPv4 or IPv6 address, subnet, or netgroup.
Client fields:
• Clients - allowed access to the export
• Always read-write clients - allowed read/write access regardless of export's
access restriction setting
• Always read-only clients - allowed read-only access regardless of export's
access restriction setting
• Root clients - map as root
OneFS can have multiple exports with different rules that apply the same directory.
A network hostname, an IP address, a subnet, or a netgroup name can be used for
reference. The same export settings and rules that are created here apply to all the
listed directory paths. If no clients are listed in any entries, no client restrictions
apply to attempted mounts.
When multiple exports are created for the same path, the more specific rule takes
precedence. For example, if the 192.168.3 subnet has read-only access and
192.168.3.3 client has read/write access.
PowerScale Administration-SSP1
PowerScale Administration-SSP1
Permissions settings can restrict access to read-only and enable mount access to
subdirectories. Other export settings are user mappings.69
NFS Considerations
Challenge
Lab Assignment: Now that you have learned how to create an export,
you are ready to create the NFS directory, export the directory, and
mount it to the Centos client.
69The "root user mapping" default is to map root users to nobody, and group is
none. The default Security type is "UNIX (system)". Scrolling down in the "Create
an export" window shows the "Advanced settings".
PowerScale Administration-SSP1
S3 Buckets
Scenario
S3 Overview
OneFS namespace
Objects stored in buckets
Amazon Simple Storage Service (S3) is an AWS service that provides object
storage through a web interface. OneFS 9.0.x and later support S3 as a tier 1
protocol. OneFS S3 value:
PowerScale Administration-SSP1
• Multi-protocol access71
• Multi-tenancy - access zone aware
• Latency and IOPs equivilent to other OneFS protocols
• Evolve the PowerScale data lake story:
• Single namespace and multi-protocol access
• Concurrent access72 to objects and files
• Interoperability with OneFS data services such as snapshots, WORM, quotas,
SnycIQ, and others
Enable S3 Service
71
Support interoperability between all OneFS supported protocols. File system
mapping: Object to file, object to directory, and bucket to base directory.
PowerScale Administration-SSP1
Default ports
WebUI Protocols > Object storage (S3) page, Global settings tab. Click the image to enlarge.
Zone Settings
You can create buckets using the Object storage (S3) page or using the isi s3
buckets create command.
PowerScale Administration-SSP1
WebUI Protocols > Object storage (S3) page. Click the image to enlarge.
Create Bucket
PowerScale Administration-SSP1
The graphic shows the Create a Bucket fields completed and the command to view
a created bucket.
S3 Bucket Table
PowerScale Administration-SSP1
Key Management
A key must be created to authenticate the access. Key management from WebUI
facilitates generation of secret keys and access ID. The example show key creation
using the CLI.
PowerScale Administration-SSP1
Considerations
Services
PowerScale Administration-SSP1
Challenge
PowerScale Administration-SSP1
Hadoop Introduction
Requires license
The Hadoop Distributed File System (HDFS) protocol enables a cluster to work
with Apache Hadoop, a framework for data-intensive distributed applications.
Swift Overview
OneFS supports Swift, an object storage interface compatible with the OpenStack
Swift 1.0 API. Swift is a hybrid between the two storage types, storing Swift
metadata as an alternative data stream. Through Swift, users can access file-
based data that is stored on the cluster as objects. The Swift API is implemented
as Representational State Transfer, or REST, web services over HTTP or HTTPS.
Since the Swift API is considered a protocol, content and metadata can be ingested
as objects and concurrently accessed through protocols that are configured on the
cluster. The cluster must be licensed to support Swift.
PowerScale Administration-SSP1
through the OneFS HDFS. Swift benefits include secure multitenancy for
applications through access zones while protecting the data with capabilities such
as authentication, access control, and identity management. Manage data through
enterprise storage features such as deduplication, replication, tiering, performance
monitoring, snapshots, and NDMP backups. Swift balances the workload across
the cluster nodes through SmartConnect and stores object data more efficiently
with FEC instead of data replication.
Swift client
access
PowerScale Administration-SSP1
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 173
Foundations of Data Protection and Data Layout
File Striping
Scenario
IT Manager: I am not sure how the cluster does striping. I want you to
do some research and let me know how the operating system stripes a
file.
Your Challenge: The IT manager wants you to describe how files are
broken up for file stripes and diagram the high-level file striping steps.
OneFS protects files as the data is being written. Striping protects the cluster data
and improves performance. To understand OneFS data protection, the first step is
grasping the concept of data and forward error correction or FEC stripes.
PowerScale Administration-SSP1
• File Stripes - files are logically segmented into 128 KB stripe units to calculate
protection
• FEC stripe unit - FEC stripe unit is the calculated piece of data protection
• Data stripe units + FEC stripe units = Stripe width.In the graphic, the stripe
width is 12 (eight data [1 MB file data] + 4 FEC)
• 16 data stripe units + 4 FEC = Maximum Stripe width of 20.
• 16 data stripe units = 2 MB. Files larger than 2 MB have multiple data stripe
units.
The data stripe units and protection stripe units are calculated for each file stripe by
the Block Allocation Manager (BAM) process73.
1
2
73The BAM process calculates 128-KB FEC stripe units to meet the protection
level for each file stripe. The higher the protection level, the more FEC stripes units
are calculated.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 175
Foundations of Data Protection and Data Layout
16 X 8K = 128-KB
3: The protection is calculated based on the requested protection level for each file
stripe using the data stripe units that are assigned to that file stripe.
4: The combined 128-KB stripe units are called the Stripe Width. A single file stripe
width can contain up to 16, 128-KB data stripe units for a maximum size of 2 MB as
the files data portion. A large file has thousands of file stripes per file that is
distributed across the node pool.
The steps shows a simple example of the write process. The client saves a file to
the node it is connected to. The file is divided into data stripe units. The data stripe
units are assembled into the maximum stripe widths for the file. FEC stripe units
are calculated to meet the Requested Protection level. Then the data and FEC
stripe units are striped across nodes.
Step 1
OneFS stripes the data stripe units and FEC stripe units across the node pools.
Some protection schemes74 use more than one drive per node.
74OneFS uses advanced data layout algorithms to determine data layout for
maximum efficiency and performance. Data is evenly distributed across nodes in
the node pool as it is written. The system can continuously reallocate where the
data is stored and make storage space more usable and efficient. Depending on
the file size and the stripe width, as the cluster size increases, the system stores
PowerScale Administration-SSP1
File
Client
Graphic shows Gen 6 cluster with a simple example of the write process.
Step 2
If the file is greater than 128 KB, then the file is divided into data stripe units.
large files more efficiently. Every disk within each node is assigned both a unique
GUID (global unique identifier) and logical drive number. The disks are subdivided
into 32-MB cylinder groups that are composed of 8-KB blocks. Each cylinder group
is responsible for tracking, using a bitmap, whether its blocks are used for data,
inodes or other metadata constructs. The combination of node number, logical
drive number, and block offset make the block or inode address, which the Block
Allocation Manager controls.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 177
Foundations of Data Protection and Data Layout
Step 3
The node that the client connects to is the node that performs the FEC calculation.
PowerScale Administration-SSP1
Step 4
The data stripe units are assembled to maximum stripe width for the file. Also, here
the protection level that is configured is N+1n75.
Step 5
Depending on the write pattern, the data and FEC stripes might be written to one
drive per node or two drives per node. The important take away is that files
segment into stripes of data, FEC is calculated and this data distributes across the
cluster.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 179
Foundations of Data Protection and Data Layout
PowerScale Administration-SSP1
Challenge
IT Manager:
Open participation questions:
Question: What does OneFS consider a small file and how are
small files put on disks for protection?
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 181
Foundations of Data Protection and Data Layout
Data Protection
Scenario
Data protection is one of the variables that are used to determine how data is laid
out. OneFS is designed to withstand multiple simultaneous component failures
while still affording access to the entire file system and dataset.
• OneFS uses the Reed-Solomon algorithm
• The data can be protected up to an N+4n scheme
• In OneFS, protection is calculated per individual files
PowerScale Administration-SSP1
In Gen 6.5 nodes, the journal is stored on an NVDIMM that is battery protected.
N+Mn
76Smaller neighborhoods improve efficiency by the fact that the fewer devices you
have within a neighborhood, the less chance that multiple devices will
simultaneously fail.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 183
Foundations of Data Protection and Data Layout
• Mn79
• N+Mn80
• N=M81
• N>M82
The number of sustainable drive failures are per disk pool. Multiple drive failures on
a single node are equivalent to a single node failure. The drive loss protection level
is applied per disk pool.
79 The “Mn” is the number of simultaneous drive or node failures that can be
tolerated without data loss.
80 The available N+Mn Requested Protection levels are plus one, two, three, or four
“n” (+1n, +2n, +3n, and +4n). With N+Mn protection, only one stripe unit is written
to a single drive on the node.
82N must be greater than M to gain efficiency from the data protection. If N is less
than M, the protection results in a level of FEC calculated mirroring.
PowerScale Administration-SSP1
N+Md:Bn Protection
N + Md : Bn
The “d” is the number of drives and “n” is the number of nodes. So N+3d:1n reads
as N+3 drives or 1 node.
Unlike N+Mn, N+Md:Bn has different values for the number of drive loss and node
losses that are tolerated before data loss may occur. When a node loss occurs,
multiple stripe units are unavailable from each protection stripe and the tolerable
drive loss limit is reached when a node loss occurs.
• M83
• d84
• Colon (:)85
83In this protection level, M is the number of drives per node onto which a stripe
unit is written.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 185
Foundations of Data Protection and Data Layout
• B86
• n87
With Gen 6x, for better reliability, better efficiency, and simplified protection, using
+2d:1n, +3d:1n1d, or +4d:2n is recommended.
N is replaced in the actual protection with the number of data stripe units for each
protection stripe. If there is no / in the output, it implies a single drive per node.
Mirrored file protection is represented as 2x to 8x in the output.
86 The B value represents the number of tolerated node losses without data loss.
PowerScale Administration-SSP1
Drives per
node
N+2/2
The graphic shows viewing the output showing Actual protection on a file from the isi get command.
The output displays the number of data stripe units plus the number of FEC stripe units that are
divided by the number of disks per node the stripe is written to
The protection overhead for each protection level depends on the file size and the
number of nodes in the cluster. The percentage of protection overhead declines as
the cluster gets larger. In general, N+1n protection has a protection overhead equal
to the capacity of one node, N+2n to the capacity of two nodes, N+3n to the
capacity of three nodes, and so on.
Data mirroring requires significant storage overhead and may not always be the
best data-protection method. Example89
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 187
Foundations of Data Protection and Data Layout
The table shows the relative protection overhead associated with each FEC requested protection
level. Indicators include when the FEC protection would result in mirroring.
MTTDL
MTTDL deals with how long you can go without losing data. MTTDL is used to
calculate the OneFS suggested protection.
• Accommodate failures90
• Disk pools91
• MTBF92
91Disk pools improve MTTDL because they create more failure domains, improving
the statistical likelihood of tolerating failures over the lifetime of the equipment.
PowerScale Administration-SSP1
Quorum
There are six data stripe units to write a 768-KB file. The desired protection
includes the ability to sustain the loss of two hard drives.
93For a quorum, more than half the nodes must be available over the internal,
backend network to allow writes. An eight-node Gen 6 cluster, for example,
requires a five-node quorum.
94 If there is no node quorum, reads may occur, depending upon where the data
lies on the cluster but for the safety of new data, no new information will be written
to the cluster. So, if a cluster loses its quorum, the OneFS file system becomes
read-only and will allow clients to access data but not to write to the cluster.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 189
Foundations of Data Protection and Data Layout
1 2 3
1: Using N+2n protection, the 768-KB file will be placed into three separate data
stripes, each with two protection stripe units. Six protection stripe units are required
to deliver the requested protection level for the six data stripe units. The protection
overhead is 50 percent.
2: Using N+2d:1n protection the same 768-KB file requires one data stripe, two
drives wide per node and only two protection stripe units. The eight stripe units are
written to two different drives per node. The protection overhead is the same as the
eight node cluster at 25 percent.
3: If there is a eight node cluster, two FEC stripe units would be calculated on the
six data stripe units using an N+2n protection level. The protection overhead in this
case is 25 percent.
PowerScale Administration-SSP1
Mirroring is used to protect the file metadata and some system files that exist under
/ifs in hidden directories. Mirroring can be explicitly96 set as the requested
protection level in all available locations.
Use Case97
2
2x to 8x
x
Mirroring. Original file plus
- The protection blocks are copies of the original set of data 1 to 7 copies.
3
blocks.
x
5
-The protection is explicitly set and the required mirroring is x
selected.
6
-Actual protection is applied for other Requested Protection
x
Levels
7
x
8
x
96 Mirroring is set as the actual protection on a file even though another requested
protection level is specified under certain conditions. If the files are small, the FEC
protection for the file results in a mirroring. The loss protection requirements of the
requested protection determine the number of mirrored copies. Mirroring is also
used if the node pool is not large enough to support the requested protection level.
For example, five nodes in a node pool with N+3n Requested Protection, saves the
file at 4X mirror level, the actual protection.
97 One particular use case is where the system is used to only store small files. A
file of 128 KB or less is considered a small file. Some workflows store millions of 1
KB to 4-KB files. Explicitly setting the requested protection to mirroring can save
fractions of a second per file and reduce the write ingest time for the files.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 191
Foundations of Data Protection and Data Layout
Stripe
Some protection schemes use a single drive per node per protection stripe. The
graphic shows only a single data stripe unit, or a single FEC stripe unit is written to
each node. These protection levels are N+M or N+Mn.
PowerScale Administration-SSP1
The table shows each requested N+Mn Requested Protection level over the
minimum number of required nodes for each level. The data stripe units and
protection stripe units98 can be placed on any node pool and in any order.
FEC Node 8
FEC Node 9
The number of data stripe units depends on the size of the file and the size of the
node pool up to the maximum stripe width. As illustrated, N+1n has one FEC stripe
98The number of data stripe units depends on the size of the file and the size of the
node pool up to the maximum stripe width. N+1n has one FEC stripe unit per
protection stripe, N+2n has two, N+3n has three, and N+4n has four. N+2n and
N+3n are the two most widely used Requested Protection levels for larger node
pools, node pools with around 15 nodes or more. The ability to sustain both drive or
node loss drives the use when possible.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 193
Foundations of Data Protection and Data Layout
unit per protection stripe, N+2n has two, N+3n has three, and N+4n has four. N+2n
and N+3n are the two most widely used Requested Protection levels for larger
node pools, node pools with around 15 nodes or more. The ability to sustain both
drive or node loss drives the use when possible.
N+M:B or N+Md:Bn protection protection schemes use multiple drives per node.
The multiple drives contain parts of the same protection stripe. Multiple data stripe
units and FEC stripe units are placed on a separate drive on each node.
Protection:
N+2d:1n
Stripe
The graphic shows an example of a 1 MB file with a Requested Protection of +2d:1n. Four stripe
units, either data or protection stripe units are placed on separate drives in each node. Two drives
on different nodes per disk pool can simultaneously be lost or a single node without the risk of data
loss.
PowerScale Administration-SSP1
Advanced Protection
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 195
Foundations of Data Protection and Data Layout
level of node loss protection. Besides the drive loss protection, the node loss
protection is increased.
The table shows examples of the advanced N+Md:Bn protection schemes100. Two
drives per node per protection stripe. The number of FEC stripe units does not
equal the number of drives that are used for the protection stripe. Even if one node
is lost, there is still a greater level of protection available.
100 Like other protection levels, the data stripe units and FEC stripe units are placed
on any node in the node pool and on any drive. N+3d:1n1d is the minimum
protection for node pools containing 6-TB drives. The use of N+4d:2n is expected
to increase especially for smaller to middle sized node pools as larger drives are
introduced.
PowerScale Administration-SSP1
Protection Overhead
The protection overhead for each protection level depends on the file size and the
number of nodes in the cluster. The percentage of protection overhead declines as
the cluster gets larger.
• N+1n101
• N+2n102
• N+3n103
• Data Mirroring104
For better reliability, better efficiency, and simplified protection, use N+2d:1n,
N+3d:1n1d, or N+4d:2n, as indicated with a red box.
101 N+1n protection has a protection overhead equal to the capacity of one node.
102 N+2n protection has a protection overhead equal to the capacity two nodes.
103N+3n is equal to the capacity of three nodes, and so on. OneFS also supports
optional data mirroring from 2x-8x, enabling from two to eight mirrors of the
specified content.
104 Data mirroring requires significant storage overhead and may not always be the
best data-protection method. For example, if you enable 3x mirroring, the specified
content is explicitly duplicated three times on the cluster. Depending on the amount
of content being mirrored, the mirrors can require a significant amount of capacity.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 197
Foundations of Data Protection and Data Layout
The table shows the relative protection overhead that is associated with each FEC requested
protection level available in OneFS. Indicators include when the FEC protection would result in
mirroring.
Considerations
As the cluster scales, the default protection may need adjusting. You may not want
to apply a higher protection to the entire cluster. Although you get better protection,
it is less efficient. Listed are areas to consider.
106Because the system is doing more work to calculate and stripe the protection
data – impact is approximately linear.
PowerScale Administration-SSP1
Challenge
109The customer may want to protect some repositories at a higher level than the
cluster default.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 199
Foundations of Data Protection and Data Layout
Protection Management
Scenario
2 3
4: Actual is the level of protection OneFS applies to data. It can be more than
requested protection but never less.
PowerScale Administration-SSP1
Requested Protection
Directory path
File
H600
A200
Cluster-wide settings
The cluster-wide default data protection setting is made using the default file
pool110 policy.
110The View default policy details window displays the current default file pool
policy settings. The current protection is displayed under requested protection. The
default setting is to use the requested protection setting at the node pool level as
highlighted in the Edit default policy details window.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 201
Foundations of Data Protection and Data Layout
To view or edit the default setting, go to File system > Storage pools > File pool policies, and
click View / Edit on the Default policy. isi file pool policy modify finance --set-
requested-protection +3:1, sets the requested protection for the file pool policy at +3d:1n.
The default file pool policy protection setting uses the node pool or tier setting.
When a node pool is created, the default requested protection111 that is applied to
the node pool is +2d:1n.
The current requested protection for each node pool is displayed in the Tiers and
node pools section.
PowerScale Administration-SSP1
To view and edit the requested protection setting for the node pools in the WebUI, go to the File
system > Storage pools > SmartPools page. isi storagepool nodepools modify
v200_25gb_2gb --protection-policy +2n, sets the requested protection of a node pool to
+2n.
OneFS stores the properties for each file. To view the files and the next level
subdirectories, click the specific directory.
Manual settings112
112 Manual settings can be used to modify the protection on specific directories or
files. The settings can be changed at the directory, subdirectory, and file level. Best
practices recommend against using manual settings, because manual settings can
return unexpected results and create management issues as the data and cluster
age. Once manually set, reset the settings to default to use automated file pool
policy settings, or continue as manually managed settings. Manual settings
override file pool policy automated changes. Manually configuring is only
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 203
Foundations of Data Protection and Data Layout
To view directories and files on the cluster, go to File System > File system explorer.
recommended for unique use cases. Manual changes are made using the WebUI
File system explorer or the CLI isi set command.
PowerScale Administration-SSP1
H600
A200
The graphic shows a workflow that moves data to an archive tier of storage.
Suggested Protection
Suggested protection refers to the visual status and CELOG event notification
when node pools are set below the calculated suggested protection level.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 205
Foundations of Data Protection and Data Layout
Not using the suggested protection does not mean that data loss occurs, but it
does indicate that the data is at risk. Avoid anything that puts data at risk. What
commonly occurs is a node pool starts small and then grows beyond the configured
requested protection level. The once adequate +2d:1n requested protection level
becomes no longer appropriate, but is never modified to meet the increased
protection requirements.Not using the suggested protection does not mean that
data loss occurs, but it does indicate that the data is at risk. Avoid anything that
puts data at risk. What commonly occurs is a node pool starts small and then
grows beyond the configured requested protection level.
PowerScale Administration-SSP1
The Suggested protection feature provides a method to monitor and notify users
when the requested protection setting is different than the suggested protection for
a node pool.
SmartPools module health status - suggested To modify the settings, click on View/Edit
protection is part of the reporting in the tab
Indicates v200_24gb_2gb node pool with a requested protection level that is different
than the suggested
The notification shows the suggested setting and node pools that are within suggested protection
levels are not displayed.
Actual Protection
The actual protection114 applied to a file depends on the requested protection level,
the size of the file, and the number of node pool nodes.
114 The actual protection level is the protection level OneFS sets. Actual protection
is not necessarily the same as the requested protection level.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 207
Foundations of Data Protection and Data Layout
Orange - mirroring,
low minimum size for
requested protection
Bold - actual
requested protection
Red - actual
The chart indicates the actual protection that is applied to a file according to the number of nodes in the node pool. If protection changes
actual protection does not match the requested protection level, it may have been changed to be more efficient given the from requested
file or number of nodes in the node pool. protection
115A requested protection of +2d:1n and there is a 2-MB file and a node pool of at
least 18 nodes, the file is laid out as +2n.
116A 128-KB file is protected using 3x mirroring, because at that file size the FEC
calculation results in mirroring.
117 In both cases, the actual protection applied to the file exceeds the minimum
drive loss protection of two drives and node loss protection of one node. The
exception to meeting the minimum requested protection is if the node pool is too
small and unable to support the requested protection minimums. For example, a
node pool with four nodes and set to +4n requested protection. The maximum
supported protection is 4x mirroring in this scenario.
PowerScale Administration-SSP1
N+2/2
Drives per node
Output
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 209
Foundations of Data Protection and Data Layout
isi get
The isi get command provides detailed file or directory information. The primary
options are –d <path> for directory settings and –DD <path>/<filename> for
individual file settings.
The graphic shows the isi get –DD output. The output has three primary
locations containing file protection. The locations are a summary in the header, line
item detail settings in the body, and detailed per stripe layout per drive at the
bottom.
Challenge
IT Manager:
Open participation questions:
Question: What is a use case for setting requested protection at
the cluster level? At the node pool level? At the directory level?
PowerScale Administration-SSP1
Data Layout
Scenario
IT Manager: You are doing a great job. Now, examine how OneFS lays
out the data on disks.
4
1
3
2
2
3
1
4
1: The number of nodes in a node pool affects the data layout because data
spreads across all nodes in the pool. The number of nodes in a node pool
determines how wide the stripe can be.
2: The nomenclature for the protection level is N+Mn, where N is the number of
data stripe units and Mn is the protection level. The protection level also affects
data layout. You can change the protection level down to the file level, and the
protection level of that file changes how it stripes across the cluster.
3: The file size also affects data layout because the system employs different
layout options for larger files than for smaller files to maximize efficiency and
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 211
Foundations of Data Protection and Data Layout
performance. Files smaller than 128 KB are treated as small files. Due to the way
that OneFS applies protection, small files are triple mirrored.
4: The access pattern modifies both prefetching and data layout settings that are
associated with the node pool. Disk access pattern can be set at a file or directory
level so you are not restricted to using only one pattern for the whole cluster.
There are four variables that combine to determine how OneFS lays out data.
The variables make the possible outcomes almost unlimited when trying to
understand how the cluster behaves with varying workflow with differing variables.
You can manually define some aspects of how it determines what is best, but the
process is automated.
An administrator can optimize layout decisions that OneFS makes to better suit the
workflow. The data access pattern influences how a file is written to the drives
during the write process.
2: Use Streaming for large streaming workflow data such as movie or audio files.
Streaming prefers to use as many drives as possible, within the given pool, when
writing multiple protection stripes for a file. Each file is written to the same sub pool
within the node pool. Streaming maximizes the number of active drives per node as
the streaming data is retrieved. Streaming also influences the prefetch caching
algorithm to be highly aggressive and gather as much associated data as possible.
The maximum number of drives for streaming is five drives per node across the
node pool for each file.
PowerScale Administration-SSP1
3: A random access pattern prefers using a single drive per node for all protection
stripes for a file, like a concurrency access pattern. With random however, the
prefetch caching request is minimal. Most random data does not benefit from
prefetching data into cache.
A 1 MB file is divided into eight data stripe units and three FEC units. The data is
laid out in three stripes. With a streaming access pattern, more spindles are
preferred. 1 MB file split into eight stripe unit and three stripes - streaming uses
spindles.
Streaming
N +1n
1024 KB
file
8 X 128
KB chunk
3 stripes and 3
drives wide
The graphic is a representation of a Gen 6 chassis with four nodes. Each node has five drive sleds.
Each drive sled has three disks. The orange disk represents a neighborhood. The disk that is used
is in the same neighborhood (orange), do not traverse to disks in the other neighborhoods (gray)
A 1-MB file is divided into eight data stripe units and three FEC units. The data is
laid out in three stripes, one drive wide.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 213
Foundations of Data Protection and Data Layout
Concurrency
N+ 1
1024 KB
file
8 X 128
KB chunk
3 stripes and 1
drive wide
The graphic is a representation of a Gen 6 chassis with four nodes. Each node has five drive sleds.
Each drive sled has three disks. The orange disk represents a neighborhood.
Configuring the data access pattern is done on the file pool policy, or manually at
the directory and file level. Set data access patterns using the WebUI or use isi
set for directory and file level or isi file pool policy for file pool policy
level.
PowerScale Administration-SSP1
For WebUI Administration, go to File systems > Storage pools > File pool policies.
Challenge
IT Manager:
Open participation questions:
Question: What is the preferred file layout with a streaming
access pattern?
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 215
Configuring Storage Pools
PowerScale Administration-SSP1
Storage Pools
Scenario
IT Manager: Before you configure file policies and tiering data, I want
you to explain the components of storage pools.
Storage pools monitor the health and status at the node pool level. Using storage
pools, multiple tiers of nodes (node pools) can all co-exist within a single file
system, with a single point of management.
Node Pool
Node Pool
PowerScale Administration-SSP1
Gen 6 drive sleds have three, four, or six drives whereas the F200 has 4 drive bays
and the F600 has 8 drive bays.
The graphic shows a Gen 6 node pool that has two chassis, eight nodes, and each node having five
drive sleds with three disks.
Exploring the building blocks and features of storage pools helps understand the
underlying structure when moving data between tiers. The storage pool
components, SmartPools, File Pools and CloudPools, are covered in detail in other
topics.
PowerScale Administration-SSP1
Disk Pool
Neighborhood Tier
2
Neighborhoods
are a group of disk pools and can span from 4
up to 19 nodes for Gen 6 nodes. Nodes have a
single neighborhood from 1-to-19
nodes. Neighborhoods are
automatically assigned and not configurable.
119Not spanning disk pools the granularity at which files are striped to the cluster.
Disk pool configuration is automatic and cannot be configured manually. Removing
a sled does not cause data unavailability as only one disk per disk pool is
temporarily lost.
PowerScale Administration-SSP1
Gen 6 Neighborhood
A Gen 6 node pool splits into two neighborhoods when adding the 20th node 120.
One node from each node pair moves into a separate neighborhood.
Each
Single neighborh neighborhood
ood, 3 disk pools has 3 disk pools
in a 3 diskper
drive sled
example
At 40 nodes,protection against
chassis failure
120After the 20th node added up to the 39th node, no 2 disks in a given drive sled
slot of a node pair share a neighborhood. The neighborhoods split again when the
node pool reaches 40 nodes.
PowerScale Administration-SSP1
The graphic shows a 40 node cluster used to illustrate a chassis failure. Once the
40th node is added, the cluster splits into four neighborhoods, labeled NH 1
through NH 4.
PowerScale Administration-SSP1
Node Pool
SmartPools
SmartPools Basic121
121The basic version of SmartPools supports virtual hot spares, enabling space
reservation in a node pool for re protection of data. OneFS implements SmartPools
PowerScale Administration-SSP1
SmartPools Advanced122
File Pools
File pools are the SmartPools logical layer, at which file pool policies are applied.
User created, and defined policies are set on the file pools.
CloudPools
Moving the cold archival data to the cloud, lowers storage cost and optimizes
storage resources.
CloudPools offers the flexibility of another tier of storage that is off-premise and off-
cluster.
basic by default. You can create multiple node pools, but only a single tier and only
a single file pool. A single tier has only one file pool policy that applies the same
protection level and I/O optimization settings to all files and folders in the cluster.
122More advanced features are available in SmartPools with a license. With the
advanced features you can create multiple tiers and file pool policies that direct
specific files and directories to a specific node pool or a specific tier. Advanced
features include the ability to create multiple storage tiers, multiple file pool policy
targets, and multiple file pool policies.
PowerScale Administration-SSP1
Node Loss: A loss of a node does not automatically start reprotecting data. Many
times a node loss is temporary, such as a reboot. If N+1 data protection is
configured on a cluster, and one node fails, the data is accessible from every other
node in the cluster. If the node comes back online, the node rejoins the cluster
automatically without requiring a rebuild. If the node is physically removed, it must
also be smartfailed. Only Smartfail nodes when needing to remove from the cluster
permanently.
The graphic shows the isi storagepool settings view command with user
configured settings highlighted.
Serviceability
Listed are the CLI options that can help get information about storage pools.
• isi status -p
PowerScale Administration-SSP1
Challenge
Lab Assignment: Go to the lab and verify the storage pool settings.
PowerScale Administration-SSP1
File Pools
Scenario
IT Manager: Our media team needs their storage on disks that do not
compete with the other disk.
Your Challenge: The IT manager has tasked you to segregate data into
different node pools.
H400
F200
PowerScale Administration-SSP1
File pool policies automate file movement, enabling users to identify and move
logical groups of files.
• User-defined filters123
• File-based, not hardware-based124
• User-defined or default protection and policy settings125
The example shows that each policy has a different optimization and protection
level. A file that meets the policy criteria for tier 3 is stored in the tier 3 node pool
with +3d:1n1d protection. Also, the file is optimized for streaming access.
The default file pool policy is defined under the default policy.
123Files and directories are selected using filters and apply actions to files
matching the filter settings. The policies are used to change the storage pool
location, requested protection settings, and I/O optimization settings.
124
Each file is managed independent of the hardware, and is controlled through the
OneFS operating system.
125 Settings are based on the user-defined and default storage pool policies. File
pool policies add the capability to modify the settings at any time, for any file or
directory.
PowerScale Administration-SSP1
2
1
3
4
1: The individual settings in the default file pool policy apply to files without settings
that are defined in another file pool policy that you create. You cannot reorder or
remove the default file pool policy.
2: To modify the default file pool policy, click File system, click Storage pools,
and then click the File pool policies tab. On the File pool policies page, next to
the Default policy, click View/Edit.
3: You can choose to have the data that applies to the Default policy target a
specific node pool or tier or go anywhere. Without a license, you cannot change
the anywhere target. If existing file pool policies direct data to a specific storage
pool, do not configure other file pool policies with anywhere.
4: You can define the SSD strategy for the Default policy.
5: You can specify a node pool or tier for snapshots. The snapshots can follow the
data, or go to a different storage location.
6: Assign the default requested protection of the storage pool to the policy, or set a
specified requested protection.
PowerScale Administration-SSP1
8: In the Data access pattern section, you can choose between Random,
Concurrency, or Streaming.
PowerScale Administration-SSP1
This example is a use case where a media-orientated business unit wants greater
protection and an access pattern that is optimized for streaming.
A tier that is called media_tier with a node pool has been created.
The business unit targets their mp4 marketing segments to the media_tier where
the hosting application can access them.
PowerScale Administration-SSP1
Create the filters in the File matching criteria section when creating or editing a
file pool policy.
Filter elements:
• Filter type126
126 File pool policies with path-based policy filters and storage pool location actions
are run during the write of a file matching the path criteria. Path-based policies are
first started when the SmartPools job runs, after that they are started during the
matching file write. File pool policies with storage pool location actions, and filters
that are based on other attributes besides path, write to the node pool with the
highest available capacity. The initial write ensures that write performance is not
sacrificed for initial data placement.
PowerScale Administration-SSP1
• Operators127
• Multiple criteria128
SSD Options
With the exception of F-Series nodes, if a node pool has SSDs, by default the L3
cache is enabled on the node pool. To use the SSDs for other strategies, first
disable L3 cache on the node pool. Manually enabling SSD strategies on specific
files and directories is not recommended.
SSDs for Metadata Read Acceleration is the recommended setting. The setting
uses one metadata mirror, other mirrors and data on HDDs.
Pros Cons
127Operators can vary according to the selected filter. You can configure the
comparison value, which also varies according to the selected filter and
operator. The Ignore case box should be selected for files that are saved to the
cluster by a Windows client.
128The policy requires at least one criterion, and allows multiple criteria. You can
add AND or OR statements to a list of criteria. Using AND adds a criterion to the
selected criteria block. Files must satisfy each criterion to match the filter. You can
configure up to three criteria blocks per file pool policy.
PowerScale Administration-SSP1
Helps Job Engine - all random Usually shows small SSD utilization:
lookups and treewalks are faster as clients may ask “Where is the value”
one copy of metadata is always on or complain it was over configured
SSD.
Metadata read/write acceleration requires more SSD space. Writes all metadata
mirrors to SSDs and can consume up to six times more SSD space.
Pros Cons
Metadata updates hit SSDs - speeds Overfilling SSDs can have significant
up creates, writes, and deletes impact – manage with care.
including SnapShot deletes.
Does not show the full utilization until
the file system capacity is high.
Use SSDs for data and metadata requires the most space. Writes all data and
metadata for a file on SSDs.
Pros Cons
PowerScale Administration-SSP1
Use file pool policies designating Must manage total SSD capacity
specific path for the data on SSDs. utilization - can push metadata from
SSD, which has a wide impact.
Avoid SSDs
Using the avoid SSDs option affects performance. This option writes all file data
and all metadata mirrors to HDDs. Typically, use this setting when implementing L3
cache and GNA in the same cluster. You create a path-based file pool policies that
targets an L3 cache enabled node pool. The data SSD strategy and snapshot SSD
strategy for this L3 cache enabled node pole should be set to ‘Avoid SSD’.
PowerScale Administration-SSP1
The FilePolicy job on the WebUI Cluster management > Job operations > Job types page.
129 The SetProtectPlus job applies the default file pool policy.
130
When SmartPools is licensed, the SmartPools job processes and applies all file
pool policies. By default, the job runs at 22:00 hours every day at a low priority.
PowerScale Administration-SSP1
Policy Template
Policy templates on the WebUI File system > Storage pools > File pool policies page.
Template settings are preset to the name of the template along with a brief
description. You can change the settings.
Template considerations:
• Opens a partially populated, new file pool policy.
• You must rename the policy.
• You can modify and add criteria and actions.
• Use in web administration interface only.
131 Uses a file system index database on the file system instead of the file system
itself to find files needing policy changes. By default, the job runs at 22:00 hours
every day at a low priority. The FilePolicy job was introduced in OneFS 8.2.0.
132The SmartPoolsTree job is used to apply selective SmartPools file pool policies.
The job runs the "isi filepool apply" command. The Job Engine manages the
resources that are assigned to the job. The job enables for testing file pool policies
before applying them to the entire cluster.
PowerScale Administration-SSP1
Plan to add more node capacity when the cluster reaches 80% so that it does not
reach 90%. The cluster needs the extra capacity for moving around data, and for
the VHS space to rewrite data when a drive fails. Listed are more considerations.
• Avoid overlapping file policies where files may match more than one rule. If data
matches multiple rules, only the first rule is applied.
• File pools should target a tier and not a node pool within a tier.
• You can use the default policy templates as examples.
PowerScale Administration-SSP1
Serviceability
Example output of the 'isi filepool apply <path/file> -n -v -s' command with truncated output.
Listed here are the CLI options that can help get information about file pools.
• If file pool policy rules are not being applied properly, check the policy order.
• Test file pool policy before applying.
Challenge
PowerScale Administration-SSP1
SmartPools
Scenario
SmartPools Overview
SmartPools enables the grouping of nodes into storage units that include node
pools, CloudPools, and tiers.
With SmartPools, you can segregate data based on its business value, putting data
on the appropriate tier of storage with appropriate levels of performance and
protection.
133Node pool membership changes through the addition or removal of nodes to the
cluster. Typically, tiers are formed when adding different node pools on the cluster.
PowerScale Administration-SSP1
SmartPools Licensing
Because of the availability to have multiple data target locations, some additional
target options are enabled in some global settings.
PowerScale Administration-SSP1
SmartPool Settings
Cache Statistics
GNA
SmartPools can automatically transfer data among tiers with different performance
and capacity characteristics.
Global namespace acceleration, or GNA, enables the use of SSDs for metadata
acceleration across the entire cluster.
PowerScale Administration-SSP1
GNA Aspects
Pros Cons
Allows metadata read acceleration Difficult to manage and size the disk.
for non-SSD nodes - need some
nodes with SSDs Hard rules and limits
Helps Job Engine and random reads Links expansion of one tier to another
tier to adhere to the limits
L3Cache
L3 cache is enabled by default for all new node pools that are added to a cluster.
L3 cache is either on or off and no other visible configuration settings are available.
Any node pool with L3 cache enabled is excluded from GNA space calculations
and do not participate in GNA enablement.
PowerScale Administration-SSP1
The left graphic shows global setting. The right graphic shows L3 cache enable or disable on each
node pool separately. Click image to enlarge.
VHS
Virtual hot spare, or VHS, allocation enables space to rebuild data when a drive
fails.
When selecting the option to reduce the amount of available space, free-space
calculations exclude the VHS reserved space.
OneFS uses the reserved VHS free space for write operations unless you select
the option to deny new data writes.
Command example that reserves 10% capacity for VHS: isi storagepool
settings modify --virtual-hot-spare-limit-percent 10
PowerScale Administration-SSP1
Spillover
With the licensed SmartPools module, you can direct data to spillover to a specific
node pool or tier group.
Actions
If you clear the box (disable), SmartPools does not modify or manage settings on
the files.
PowerScale Administration-SSP1
Protection example: If a +2d:1n protection is set and the disk pool suffers three
drive failures, the data that is not lost can still be accessed. Enabling the option
ensures that intact data is still accessible. If the option is disabled, the intact file
data is not accessible.
GNA can be enabled if 20% or more of the nodes in the cluster contain SSDs and
1.5% or more of the total cluster storage is SSD-based. The recommendation is
that at least 2.0% of the total cluster storage is SSD-based before enabling GNA.
Going below the 1.5% SSD total cluster space capacity requirement automatically
disables GNA metadata. If you SmartFail a node that has SSDs, the SSD total size
percentage or node percentage containing SSDs could drop below the minimum
requirement, disabling GNA. Any node pool with L3 cache enabled is excluded
from GNA space calculations and do not participate in GNA enablement.
GNA also uses SSDs in one part of the cluster to store metadata for nodes that
have no SSDs. The result is that critical SSD resources are maximized to improve
performance across a wide range of workflows.
PowerScale Administration-SSP1
VHS example: If specifying two virtual drives or 3%, each node pool reserves
virtual drive space that is equivalent to two drives or 3% of their total capacity for
VHS, whichever is larger. You can reserve space in node pools across the cluster
for this purpose, equivalent to a maximum of four full drives. If using a combination
of virtul drives and total disk space, the larger number of the two settings
determines the space allocation, not the sum of the numbers.
SmartPools Considerations
PowerScale Administration-SSP1
• Disk pools are not user configurable, and a disk drive is only a member on one
disk pool or neighborhood.
• Node pools must have at least four nodes for Gen 6 and at least three nodes for
the F200/600. The default is one node pool per node type and configuration.
• The file pool policy default is all files are written anywhere on cluster. To target
more node pools and tiers, activate the SmartPools license.
Challenge
PowerScale Administration-SSP1
CloudPools
Scenario
IT Manager: Next, take the file pool policies to the CloudPools level. For
some of the long-term archive data, the group is looking at cloud
options.
CloudPools offers the flexibility of another tier of storage that is off-premise and off-
cluster. Essentially what CloudPools do is provide a lower TCO134 for archival-type
data. Customers who want to run their own internal clouds can use a PowerScale
installation as the core of their cloud.
The video provides a CloudPools overview and use case. See the student guide for
a transcript of the video.
PowerScale Administration-SSP1
Link:
https://edutube.emc.com/html5/videoPlayer.htm?vno=wx4VTLcN32kSlHGFwGLE1
Q
Shown is an Isilon cluster with twelve nodes. A key benefit of CloudPools is the
ability to interact with multiple cloud vendors. Shown in the graphic are the
platforms and vendors that are supported as OneFS 8.1.1.
Let us look at an example, each chassis in the cluster represents a tier of storage.
The topmost chassis is targeted for the production high-performance workflow and
PowerScale Administration-SSP1
may have node such as F800s. When data is no longer in high demand,
SmartPools moves the data to the second tier of storage. The example shows the
policy moves data that is not accessed and that is over thirty days old. Data on the
middle tier may be accessed periodically. When files are no longer accessed for
more than 90 days, SmartPools archives the files to the lowest chassis or tier such
as A200 nodes.
The next policy moves the archive data off the cluster and into the cloud when data
is not accessed for more than 180 days. Stub files that are also called SmartLinks
are created. Stub files consume approximately 8 KB space on the Isilon cluster.
Files that are accessed or retrieved from the cloud, or files that are not fully moved
to the cloud, have parts that are cached on the cluster and are part of the stub file.
The storing of CloudPools data and user access to data that is stored in the cloud
is transparent to users.
CloudPools files undergo a compression algorithm and then are broken into their 2
MB cloud data objects or CDOs for storage. The CDOs conserve space on the
cloud storage resources. Internal performance testing does note a performance
penalty for a plane compression and decompressing files on read. Encryption is
applied to file data transmitting to the cloud service. Each 128 KB file block is
encrypted using a AES 256 encryption. Then transmitted as an object to the cloud.
Internal performance testing notes a little performance penalty for encrypting the
data stream.
CloudPools Considerations
CloudPools uses the SmartPools framework to move data and state information to
off-cluster storage while retaining the ability to read, modify, and write to data.
PowerScale Administration-SSP1
CloudPools Administration
Configure and manage CloudPools from the WebUI File system, Storage pools
page, CloudPools tab. Managing CloudPools using the CLI is done with the isi
cloud command.
135 In OneFS 8.2, CloudPools compress data before sending it over the wire.
PowerScale Administration-SSP1
CloudPools Tab
Once the SmartPools and CloudPools licenses are applied, the WebUI shows the
cloud storage account options.
After a cloud storage account is defined and confirmed, the administrator can
define the cloud pool itself.
The file pool policies enable the definition of a policy to move data out to the cloud.
PowerScale Administration-SSP1
Must be unique
The graphic shows the window for creating a cloud storage account.
After creating a storage account, create a CloudPool and associate or point it to the
account.
PowerScale Administration-SSP1
CloudPools SmartLink
Run the isi get -D command to see files archived to the cloud using
CloudPools.
The example checks to see if the local version on the cluster is a SmartLink file.
PowerScale Administration-SSP1
SmartPools file pool policies are used to move data from the cluster to the selected
CloudPools storage target.
When configuring a file pool policy, you can apply CloudPools actions to the
selected files.
CloudPools Settings
You may want to modify the settings for the file pool policy based on your
requirements. Modifications are not necessary for most workflows. You can elect to
encrypt and compress data.
1
2
3
4
7
5
8
6
9
10
1: The default CloudPools setting allows you to archive files with snapshot
versions, but you can change the default setting.
2: You can encrypt data prior to archiving it to the cloud. Cloud data is decrypted
when accessed or recalled.
3: You can compress data prior to archiving to the cloud. Cloud data is
decompressed when accessed or recalled.
PowerScale Administration-SSP1
4: Set how long to retain cloud objects after a recalled file replaces the SmartLink
file. After the retention period, the cloud objects garbage collector job cleans up the
local resources allocated for the SmartLink files, and removes the associated cloud
objects.
5: If a SmartLink file has been backed up and the original SmartLink file is
subsequently deleted, associated cloud objects are deleted only after the retention
time of the backed-up SmartLink file has expired.
6: If a SmartLink file has been backed up and the original SmartLink file is
subsequently deleted, associated cloud objects are deleted only after the original
retention time, or a longer incremental or full backup retention period, has expired.
7: Specifies how often SmartLink files modified on the cluster are written to their
associated cloud data objects.
8: Determines whether cloud data is cached when a file is accessed on the local
cluster.
9: Specifies whether cloud data is fully or partially recalled when you access a
SmartLink file on the cluster.
10: Specifies how long the system retains recalled cloud data that is in the cache of
associated SmartLink files.
The graphic shows various default advanced CloudPool options that are configured.
The output of the isi cloud command shows the actions that you can take.
PowerScale Administration-SSP1
1 3 5 7 10
2 4 6 8 9
1: Use to grant access to CloudPool accounts and file pool policies. You can add
and remove cloud resource, list cluster identifiers, and view cluster details.
2: Used to manage CloudPool accounts. You can create, delete, modify, and
view a CloudPool account, and list the ClouldPool accounts.
3: Use to archive or recall files from the cloud. Specify files individually, or use a file
matching pattern. Files that are targeted for archive must match the specified file
pool policy, or any file pool policy with a cloud target.
4: Use to manage CloudPools TLS client certificates. You can delete, import,
modify, view, and list certificates.
6: Use to configure and manage a CloudPool pool. You can create, delete,
modify, list, and view pools. OneFS no longer accesses the associated cloud
storage account when it is deleted. If a file pool policy references the CloudPool,
OneFS does not allow the delete.
7: Use to manage network proxies. You can create, delete, modify, list, and
view proxies. CloudPools prevents deletion of a proxy that is attached to a cloud
storage account.
8: Files that are stored in the cloud can be fully recalled using the isi cloud
recall command. Recall can only be done using the CLI. When recalled, the full
file is restored to its original directory. The file may be subject to the same file pool
policy that originally archived it, and rearchive it to the cloud on the next
SmartPools job run. If re-archiving is unintended, the recalled file should be moved
to a different, unaffected, directory. The recalled file overwrites the stub file. You
can start the command for an individual file or recursively for all files in a directory
path.
PowerScale Administration-SSP1
9: Use to manage CloudPool top-level settings. You can list and modify
CloudPool settings, and regenerate the CloudPool master encryption key.
10: Use to restore the cloud object index (COI) for a cloud storage account on the
cluster. The isi cloud access add command also restores the COI for a cloud
storage account.
• Support137
137C2S support delivers full CloudPools functionality for a target endpoint, and
supports the use with C2S Access Portal (CAP), and X.509 client certificate
PowerScale Administration-SSP1
• Integration138
• No Internet connection139
CloudPools Limitations
In a standard node pool, file pool policies can move data from high-performance
tiers to storage tiers and back as defined by their access policies. However, data
that moves to the cloud remains stored in the cloud unless an administrator
explicitly requests data recall to local storage. If a file pool policy change is made
that rearranges data on a normal node pool, data is not pulled from the cloud.
Public cloud storage often places the largest fees on data removal, thus file pool
policies avoid removal fees by placing this decision in the hands of the
administrator.
The connection between a cluster and a cloud pool has limited statistical features.
The cluster does not track the data storage that is used in the cloud, therefore file
spillover is not supported. Spillover to the cloud would present the potential for file
recall fees. As spillover is designed as a temporary safety net, once the target pool
capacity issues are resolved, data would be recalled back to the target node pool
and incur an unexpected fee.
authority. C2S also provides support (from AIMA) to securely store certificates,
validate, and refresh if needed.
139This service is 'air gapped' which means it has no direct connection to the
Internet.
PowerScale Administration-SSP1
Statistic details, such as the number of stub files on a cluster or how much cache
data is stored in stub files and would be written to the cloud on a flush of that
cache, is not easily available. No historical data is tracked on the network usage
between the cluster and cloud either in writing traffic or in read requests. These
network usage details should be viewed from the cloud service management
system.
Challenge
PowerScale Administration-SSP1
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 261
Configuring Data Services
File Filtering
Scenario
Your Challenge: The IT manager wants you to explain file filtering and
configure the shares to filter unnecessary files.
The graphic shows that .avi files are prevented from writing to
the finance access zone.
PowerScale Administration-SSP1
File filtering enables administrators to deny or allow file access on the cluster that is
based on the file extension.
• Denies writes for new files.
• Prevents accessing existing files.
• Explicit deny lists.140
• Explicit allow lists.141
• No limit to extension list.
• Per access zone.142
• Configurable for the SMB defaults143.
• No license is required.
140Explicit deny lists are used to block only the extensions in the list. OneFS
permits all other file types to be written. Administrators can create custom
extension lists based on specific needs and requirements.
141
Explicit allow list permits access to files only with the listed file extensions.
OneFS denies writes for all other file types.
142 The top level of file filtering is set up per access zone. When you enable file
filtering in an access zone, OneFS applies file filtering rules only to files in that
access zone.
143OneFS does not take into consideration which file sharing protocol was used to
connect to the access zone when applying file filtering rules. However, you can
apply additional file filtering at the SMB share level.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 263
Configuring Data Services
If enabling file filtering on an access zone with existing shares or exports, the file
extensions determine access to the files.
• User denied access.144
• Administrator access.145
144 Users cannot access any file with a denied extension. The extension can be
denied through the denied extensions list, or because the extension was not
included as part of the allowed extensions list.
145 Administrators can still access existing files. Administrators can read the files or
delete the files. Administrators with direct access to the cluster can manipulate the
files.
PowerScale Administration-SSP1
146How the file filtering rule is applied to the file determines where the file filtering
occurs. If a user or administrator accesses the cluster through an access zone or
SMB share without applying file filtering, files are fully available.
147 File filters are applied only when accessed over the supported protocols.
149 With the compliance considerations today, organizations struggle to meet many
of the requirements. For example, many organizations are required to make all
emails available for litigation purpose. To help ensure that email is not stored
longer than wanted, deny storing .pst.
150Another use case is to limit the cost of storage. Organizations may not want
typically large files, such as video files, to be stored on the cluster, so they can
deny .mov or .mp4 file extension.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 265
Configuring Data Services
When you enable file filtering in an access zone, OneFS applies file filtering rules
only to files in that access zone.
151
An organizational legal issue is copyright infringement. Many users store their
.mp3 files on the cluster and open a potential issue for copyright infringement.
152 Another use case is to limit an access zone for a specific application with its
unique set of file extensions. File filtering with an explicit allow list of extensions
limits the access zone or SMB share for its singular intended purpose.
PowerScale Administration-SSP1
1. Select access
zone
3. Select to add or
deny
2. Enable -
unchecked by
default
Access zone level: Web UI: Access > File filter > File filter settings.
You can configure file filters on the Protocols > Windows sharing (SMB) >
Default share settings page153.
File filtering settings can be modified by changing the filtering method or editing file
extensions.
• Browse to Access > File Filter, and select the access zone that needs to be
modified from the Current Access Zone drop down list.
• Clear Enable file filters check box to disable file filtering in access zone.
153 Configuring file filters on individual SMB shares enables more granular control.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 267
Configuring Data Services
• Select to deny or allow and then enter the extension of the file, and click submit.
• Click the Remove Filter button next to the extension to remove a file name
extension.
CLI: isi smb shares create and isi smb shares modify commands. If
using RBAC, the user must have the ISI_PRIV_FILE_FILTER privilege.
Challenge
PowerScale Administration-SSP1
SmartQuotas
Scenario
This video provides an overview for SmartQuotas. See the student guide for a
transcript of the video.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 269
Configuring Data Services
Link:
https://edutube.emc.com/Player.aspx?vno=tCIE1bGAUz6k3W1ic8tZfw==&autoplay
=true
SmartQuotas is a software module that is used to limit, monitor, thin provision, and
report disk storage usage at the user, group, and directory levels. Administrators
commonly use file system quotas for tracking and limiting the storage capacity that
a user, group, or project can consume. SmartQuotas can send automated
notifications when storage limits are exceeded or approached.
Quotas are a useful way to ensure that a user or department uses only their share
of the available space. SmartQuotas are also useful for enforcing an internal
chargeback system. SmartQuotas contain flexible reporting options that can help
administrators analyze data usage statistics for their Isilon cluster. Both
enforcement and accounting quotas are supported, and various notification
methods are available.
Before OneFS 8.2, SmartQuotas reports the quota free space only on directory
quotas with a hard limit. For user and group quotas, SmartQuotas reports the size
of the entire cluster capacity or parent directory quota, not the size of the quota.
OneFS 8.2.0 includes enhancements to report the quota size for users and groups.
The enhancements reflect the true available capacity that is seen by the user.
SmartQuotas Implementation
You can choose to implement accounting quotas or enforcement quotas. The table
below displays the difference between the types.
PowerScale Administration-SSP1
Enforcement Quotas
Quota Types
2: User and default user quotas: User quotas are applied to individual users, and
track all data that is written to a specific directory. User quotas enable the
administrator to control the capacity any individual user consumes in a particular
directory. Default user quotas are applied to all users, unless a user has an
explicitly defined quota for that directory. Default user quotas enable the
administrator to apply a quota to all users, instead of individual user quotas.
3: Group and default group quotas: Group quotas are applied to groups and limit
the amount of data that the collective users within a group can write to a directory.
Group quotas function in the same way as user quotas, except for a group of
people and instead of individual users. Default group quotas are applied to all
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 271
Configuring Data Services
groups, unless a group has an explicitly defined quota for that directory. Default
group quotas operate like default user quotas, except on a group basis.
With default directory quotas, you can apply a template configuration to another
quota domain.
The graphic shows an example of creating a 10-GB hard quota, default directory
quota on the /ifs/sales/promotions directory. The directory default quota is not in
and of itself a quota on the promotions directory. Directories below the promotions
directory, such as the /Q1 and /Q2 directories inherit and apply the 10 GB quota.
The /Q1 domain and the /Q2 domain are independent of each other. Sub
directories such as /storage and /servers do not inherit the 10 GB directory
quota.Given this example, if the /Q2 folder reaches 10 GB, that linked quota is
independent of the 10 GB default directory quota on the parent directory.
Modifications to default directory quota, promotions, reflect to inherited quotas
asynchronously. Inheritance is seen when listing quotas, querying inheriting quota
record, or when I/O happen in the sub directory tree.
PowerScale Administration-SSP1
You can use the WebUI to view the created quotas and their links. See the student
guide for information about quota links.
The top example shows creating a template on the Features directory. The
directory has a hard limit of 10 GB, an advisory at 6 GB, and a soft limit at 8 GB
with a grace period of 2 days.
The Unlink option makes the quota independent of the parent, meaning
modifications to the default directory quota no longer apply to the sub directory.
This example shows removing the link on the Screen_shots sub directory and then
modifying the default directory quota on the parent, Quota, directory. Remove the
link using the button on the WebUI or isi quota quotas modify --
path=/ifs/training/Features/Quota/Screen_shots --
type=directory --linked=false. Using the --linked=true option re-links
or links to the default directory quota.
154 The 'isi quota' command is used to create the default directory quota.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 273
Configuring Data Services
Quota Accounting
Count all snapshot data in usage limits - sum of the current directory and any
snapshots of that directory 1 KB file
The quota accounting options are Include snapshots in the storage quota, 155and
155Tracks both the user data and any associated snapshots. A single path can
have two quotas that are applied to it, one without snapshot usage (default) and
one with snapshot usage. If snapshots are in the quota, more files are in the
calculation.
156Enforces the File system logical size quota limits. The default setting is to only
track user data, not accounting for metadata, snapshots, or protection.
PowerScale Administration-SSP1
• Physical size157
• Application logical size158 (OneFS 8.2 and later)
Overhead Calculations
2x data protection
157Tracks the user data, metadata, and any associated FEC or mirroring overhead.
This option can be changed after the quota is defined.
158 Tracks the usage on the application or user view of each file. Application logical
size is typically equal or less than file system logical size. The view is in terms of
how much capacity is available to store logical data regardless of data reduction,
tiering technology, or sparse blocks. The option enforces quotas limits, and reports
the total logical data across different tiers, such as CloudPools.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 275
Configuring Data Services
10 GB for the file and 10 GB for the data-protection overhead. The user has
reached 50% of the 40 GB quota by writing a 10 GB file to the cluster.
150 TB capacity
159 With thin provisioning, the cluster can be full even while some users or
directories are well under their quota limit. Configuring quotas that exceed the
cluster capacity enables a smaller initial purchase of capacity/nodes.
160
Thin provisioning lets you add more nodes as needed, promoting a capacity on-
demand model.
PowerScale Administration-SSP1
• Management reduction.161
• Careful monitoring.162
Quota Nesting
Nesting quotas is having multiple quotas within the same directory structure.
User quota = 25 GB
Directory = 1 TB
Directory can be any size up to 1 TB -
each user can only store 25 GB
Directory structure cannot exceed 1
TB
No quota
Nesting - multiple quotas within same
directory structure
At the top of the hierarchy, the /ifs/sales folder has a directory quota of 1 TB. Any
user can write data into this directory, or the /ifs/sales/proposals directory, up to a
combined total of 1 TB. The /ifs/sales/promotions directory has a user quota
assigned that restricts the total amount that any single user can write into this
directory to 25 GB. Even though the parent directory (sales) is below its quota
restriction, a user is restricted within the promotions directory. The
/ifs/sales/customers directory has a directory quota of 800 GB that restricts the
capacity of this directory to 800 GB. However, if users place 500 GB of data in the
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 277
Configuring Data Services
/ifs/sales/proposals directory, users can only place 500 GB in the other directories.
The parent directory cannot exceed 1 TB.
Create example
Modify example
View example
In OneFS 8.2.0 and later, you can view advisory and soft quota limits as a percent
of the hard quota limit.
A hard limit must exist to set the advisory and soft percentage.
PowerScale Administration-SSP1
Quota Notifications
Rules
Send notifications by email or through a cluster event. See the student guide for
more information.
The email option sends messages using the default cluster settings. You can send
the email to the owner of the event, or to an alternate contact, or both the owner
and an alternate. You can also use a customized email message template. Use a
distribution list to send the email to multiple users.
If using LDAP or Active Directory to authenticate users, the cluster uses the user
email setting that is stored within the directory. If no email information is stored in
the directory, or if a Local or NIS provider authenticates, you must configure a
mapping rule.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 279
Configuring Data Services
The graphic shows one of the available quota templates that are located in the
/etc/ifs directory.
• PAPI support163.
• OneFS 8.2 enhancements164.
164In OneFS 8.2.0, administrators can configure quota notification for multiple
users. The maximum size of the comma-separated email ID list is 1024 characters.
The isi quota command option --action-email-address field accepts multiple
comma-separated values.
PowerScale Administration-SSP1
Template Variables
An email template contains variables. You can use any of the SmartQuotas
variables in your templates.
Considerations
• Increased from 20,000 quota limits per cluster to 500,000 quota limits per
cluster.
• Quota notification daemon optimized to handle about 20 email alerts per
second.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 281
Configuring Data Services
• Support for the rpc.quotad service in the NFS container with some
statistics.
Best Practice:
• Do not enforce quotas on file system root (/ifs).
• Do not configure quotas on SyncIQ target directories.
Challenge
PowerScale Administration-SSP1
SmartDedupe
Scenario
IT Manager: The cluster is hosting home directories for the users. Much
of the data is shared and has multiple copies. Deduplication should help
address the inefficient use of space.
SmartDedupe Overview
Multiple instances of
Single instance of
identical data
data
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 283
Configuring Data Services
SmartDedupe Architecture
4 3 5
1: The SmartDedupe control path consists of PowerScale OneFS WebUI, CLI and
RESTful PAPI, and is responsible for managing the configuration, scheduling, and
control of the deduplication job.
When SmartDedupe runs for the first time, it scans the data set and selectively
samples blocks from it, creating the fingerprint index. This index contains a sorted
list of the digital fingerprints, or hashes, and their associated blocks. Then, if they
PowerScale Administration-SSP1
are determined to be identical, the block’s pointer is updated to the already existing
data block and the new, duplicate data block is released.
3: Shadow stores are similar to regular files but are hidden from the file system
namespace, so cannot be accessed via a path name. A shadow store typically
grows to a maximum size of 2GB, with 32,000 files referring each block. If the
reference count limit is reached, a new block is allocated, which may or may not be
in the same shadow store. Also shadow stores do not reference other shadow
stores. And snapshots of shadow stores are not permitted because the data that is
stored in shadow stores cannot be overwritten.
SmartDedupe Considerations
• SmartDedupe License165
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 285
Configuring Data Services
166
Deduplication is most effective for static or archived files and directories - less
modified files equals less negative effect.
168
In-line data deduplication and in-line data compression is supported in the F810
and H5600 platforms in OneFS 8.2.1.
169Deduplication does not occur across the length and breadth of the entire cluster,
but only on each disk pool individually.
170 Data that is moved between node pools may change what level of deduplication
is available. An example would be a file pool policy that moves data from a high-
performance node pool to nearline storage. The data would no longer be available
for deduplication for the other data on the high-performance node pool, but would
be newly available for deduplication on nearline storage.
171
Metadata is changed more frequently, sometimes in trivial ways, leading to poor
deduplication.
PowerScale Administration-SSP1
173The default size of a shadow store is 2 GB, and each shadow store can contain
up to 256,000 blocks. Each block in a shadow store can be referenced up to
32,000 times.
174 When deduplicated files are replicated to another PowerScale cluster or backed
up to a tape device, the deduplicated files no longer share blocks on the target
cluster or backup device. Although you can deduplicate data on a target
PowerScale cluster, you cannot deduplicate data on an NDMP backup device.
Shadow stores are not transferred to target clusters or backup devices. Because of
this, deduplicated files do not consume less space than non deduplicated files
when they are replicated or backed up. To avoid running out of space, ensure that
target clusters and tape devices have free space to store deduplicated data.
175SmartDedupe will not deduplicate the data stored in a snapshot. However, you
can create snapshots of deduplicated data. If deduplication is enabled on a cluster
that already has a significant amount of data stored in snapshots, it will take time
before the snapshot data is affected by deduplication. Newly created snapshots will
contain deduplicated data, but older snapshots will not.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 287
Configuring Data Services
SmartDedupe Function
A job in the OneFS Job Engine178 runs through blocks that are saved in every disk
pool, and compares the block hash values.179
176Only one deduplication job can run at a time - uses CPU and memory
resources, and you should run at non peak or off hour times.
178 The job first builds an index of blocks, against which comparisons are done in a
later phase, and ultimately confirmations and copies take place. The deduplication
job can be a time consuming, but because it happens as a job the system load
throttles, the impact is absolute. Administrators find that their cluster space usage
has dropped once the job completes.
179If a match is found, and confirmed as a true copy, the block is moved to the
shadow store, and the file block references are updated in the metadata.
PowerScale Administration-SSP1
2 3 4 5
2: Compare 8 KB blocks.
5: Free blocks
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 289
Configuring Data Services
180A home directory scenario where many users save copies of the same file can
offer excellent opportunities for deduplication.
181 Static, archival data is seldom changing, therefore the storage that is saved may
far outweigh the load dedupe places on a cluster. Deduplication is more justifiable
when the data is relatively static.
182 Workflows that create many copies of uncompressed virtual machine images
can benefit from deduplication. Deduplication does not work well with compressed
data, the compression process tends to rearrange data to the point that identical
files in separate archives are not identified as such. Environments with many
unique files do not duplicate each other, so the chances of blocks being found
which are identical are low.
PowerScale Administration-SSP1
SmartDedupe Jobs
Because the sharing phase is the slowest deduplication phase, a dry run, or
DedupeAssessment, returns an estimate of capacity savings.
1: The assessment enables a customer to decide if the savings that are offered by
deduplication are worth the effort, load, and cost.
2: Dedupe works on datasets which are configured at the directory level, targeting
all files and directories under each specified root directory. Multiple directory paths
can be specified as part of the overall deduplication job configuration and
scheduling.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 291
Configuring Data Services
SmartDedupe Administration
The WebUI SmartDedupe management is under the File system menu options.
Enter the paths for deduplication183 from the Settings tab.
From the Deduplication window, you can start a deduplication job and view any generated reports.
Challenge
183 Selecting specific directory gives the administrator granular control to avoid
attempting to deduplicate data where no duplicate blocks are expected, like large
collections of compressed data. Deduplicating an entire cluster without considering
the nature of the data is likely to be inefficient.
PowerScale Administration-SSP1
SnapshotIQ
Scenario
SnapshotIQ Overview
If you modify a file and determine that the changes are unwanted, you can copy or
restore the file from the earlier file version.
You can use snapshots to stage content to export, and ensure that a consistent
point-in-time copy of the data is replicated or backed up.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 293
Configuring Data Services
The graphic represents the blocks for production data and the snapshot of that production data. The
snapshot is preserving the original blocks B and E after they have changed (B' and E').
184 Some OneFS operations generate snapshots for internal system use without
requiring a SnapshotIQ license. If an application generates a snapshot, and a
SnapshotIQ license is not configured, the snapshot can be still accessed. However,
all snapshots that OneFS operations generate are automatically deleted when no
longer needed. You can disable or enable SnapshotIQ at any time. Note that you
can create clones on the cluster using the "cp" command, which does not require a
SnapshotIQ license.
PowerScale Administration-SSP1
Snapshot Operations
Bloc
k D'
Snapshot Usage
copy original
block to Block A
snapshot
Block B
Snapshot File
Block C
Block D
Block D'
Snapshot growth: as the data is modified and only the changed data blocks are
contained186 in snapshots.
185A snapshot is not a copy of the original data, but only an extra set of pointers to
the original data. At the time it is created, a snapshot consumes a negligible
amount of storage space on the cluster. The original file references the snapshots.
186If data is modified on the cluster (Block D’ in the graphic), only one copy of the
changed data is made. With CoW the original block (Block D) is copied to the
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 295
Configuring Data Services
OneFS uses both Copy on Write, or, CoW and Redirect on Write, or, RoW.
CoW are typically user-generated snapshots and RoW are typically system-
generated snapshots.
Both methods have pros and cons, and OneFS dynamically picks the snapshot
method to use to maximize performance and keep overhead to a minimum.
A
COW ROW
A
B
Snapshot
Snapshot
B
C
C
File File System D
System
D D
' B
'
The graphic shows changes that are made to, D. Changes incur a double write penalty, there is less
fragmentation of the HEAD file, which is better for cache prefetch and related file reading functions.
snapshot. The snapshot maintains a pointer to the data that existed at the time that
the snapshot was created.
PowerScale Administration-SSP1
An unordered deletion is the deletion of a snapshot that is not the oldest snapshot
of a directory. For more active data, the configuration and monitoring overhead is
slightly higher, but fewer snapshots are retained.
The benefits of unordered deletions that are compared with ordered deletions
depend on how often the snapshots that reference the data are modified. If the
data is modified frequently, unordered deletions save space. However, if data
remains unmodified, unordered deletions are not likely to save space, and it is
recommended that you perform ordered deletions to free cluster resources.
In the graphic, /ifs/org/dir2 two has two snapshot schedules. If the retention period
on schedule 1 is longer than the retention period on schedule 2, the snapshots for
the directory are deleted out of order. Unordered deletions can take twice as long
to complete and consume more cluster resources than ordered deletions. However,
unordered deletions can save space by retaining a smaller total number of blocks
in snapshots.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 297
Configuring Data Services
Creating Snapshots
188Use shorter expiration periods for snapshots that are generated more
frequently, and longer expiration periods for snapshots that are generated less
frequently.
PowerScale Administration-SSP1
OneFS tracks snapshots in the .snapshot directory. Click each tab for information
about snapshot structure and access.
Snapshot location
Accessing snapshots
190From /ifs all the .snapshots on the system can be accessed, but users can only
open the .snapshot directories for which they already have permissions. Without
access rights users cannot open or view any .snapshot file for any directory.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 299
Configuring Data Services
Preserving Permissions
Snapshots can be taken at any point in the directory tree. Each department or user
can have their own snapshot schedule.
The snapshot preserves193 the file and directory permissions at that point in time of
the snapshot.
191This is a virtual directory where all the snaps listed for the entire cluster are
stored.
192 To view the snapshots on /ifs/eng/media, user can change directory (cd) to
/ifs/eng/media and access /.snapshot
193The snapshot owns the changed blocks and the file system owns the new
blocks. If the permissions or owner of the current file is changed, it does not affect
the permissions or owner of the snapshot version.
PowerScale Administration-SSP1
Restoring Snapshots
Restore Theory
A A
File
B Snapshot
System
Time 1
C
D
D
Snapshot
Time 2
E
Restore Target
Client
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 301
Configuring Data Services
QUESTION: What happens when the user wants to recover block A data that was
overwritten in Time 3 with A’?
Snapshot options
Clients with Windows Shadow Copy Client can restore the data from the snapshot.
PowerScale Administration-SSP1
List point in time copies of the files in To recover a file, use the "mv" or "cp"
the directory command
Clients accessing the export over NFS can navigate using the .snapshot directory.
To recover a deleted file, right-click the folder that previously contained the file,
click Restore Previous Version, and select the required file to recover. To restore a
corrupted or overwritten file, right-click the file itself, instead of the folder that
contains file, and then click Restore Previous Version.
No additional storage is consumed and the restore is instant when restoring the
production file from a snap using RoW. Snapshot Time 2 has preserved A. A
backup snapshot is automatically created before copying A back to the file system.
The backup is a failback or safety mechanism should the restore from the snap be
unacceptable and the user wants to revert to A’.
SnapshotIQ Considerations
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 303
Configuring Data Services
Challenge
PowerScale Administration-SSP1
SyncIQ
Scenario
SyncIQ delivers unique, highly parallel replication performance that scales with the
dataset to provide disaster recovery. The video provides an overview of SyncIQ.
See the student guide for a transcript of the video.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 305
Configuring Data Services
Link:
https://edutube.emc.com/Player.aspx?vno=OZC9t92nwmWVLWNjfT/+5w==&autop
lay=true
Shown is a cluster with the source directory using SyncIQ to replicate data to a
remote target directory. OneFS SyncIQ uses asynchronous replication, enabling
you to maintain a consistent backup copy of your data on another Isilon cluster.
Asynchronous replication is similar to an asynchronous file write.
The target system passively acknowledges receipt of the data and returns an ACK
once the target receives the entire file or update. Then the data is passively written
to the target. SyncIQ enables you to replicate data from one PowerScale cluster to
another. Activate a SyncIQ license on both the primary and the secondary Isilon
clusters before replicating data between them. You can replicate data at the
directory level while optionally excluding specific files and sub-directories from
being replicated.
PowerScale Administration-SSP1
Under each deployment, the configuration could be for the entire cluster or a
specified source directory. Also, the deployment could have a single policy that is
configured between the clusters or several policies, each with different options
aligning to RPO and RTO requirements.
Click the tabs to know more about each type of deployment Typologies.
One-to-one
One-to-many
SyncIQ supports data replication from a single source cluster to many target
clusters, allowing the same dataset to exist in multiple locations, as illustrated in the
graphic below. A one-to-many deployment could also be referenced as a hub-and-
spoke deployment, with a central source cluster as the hub and each remote
location representing a spoke.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 307
Configuring Data Services
Many-to-one
The many-to-one deployment topology is essentially the flipped version of the one-
to-many explained in the previous section. Several source clusters replicate to a
single target cluster as illustrated in the graphic below. The many-to-one topology
may also be referred to as a hub-and-spoke configuration. However, in this case,
the target cluster is the hub, and the spokes are source clusters.
Local Target
A local target deployment allows a single Isilon cluster to replicate within itself
providing the SyncIQ powerful configuration options in a local cluster as illustrated
in the graphic below. If a local target deployment is used for disaster readiness or
archiving options, the cluster protection scheme and storage pools must be
considered.
Cascaded
PowerScale Administration-SSP1
Considerations
Capabilities
194The SyncIQ Job Engine is separate from the cluster maintenance activity Job
Engine in OneFS. SyncIQ runs based on SyncIQ policies that you can schedule or
run as required manually.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 309
Configuring Data Services
197 The SyncIQ process uses snapshots on both the source and target snapshots.
No SnapshotIQ license is required for basic SyncIQ snapshots on either the source
or target clusters. These snapshots are only used for SyncIQ jobs. SyncIQ
snapshots are single-instance snapshots and OneFS only retains the latest or last-
known good version.
198
SyncIQ can support larger maximum transmission units or MTU over the LAN or
WAN. SyncIQ supports auto-negotiation of MTU sizes over WAN connections. The
MTU across the network is negotiated by the network.
PowerScale Administration-SSP1
• Import snapshots199.
• OneFS 8.2 and above provides over-the-wire encryption200 and bandwidth
reservation201 at a policy level.
199SyncIQ has the capability to import manually taken snapshots to use as the
point-in-time reference for synchronization consistency. You can add new nodes
while a sync job runs. There is no requirement to stop the sync job before adding
new nodes. Functionality enables the ability to create a point-in-time report showing
the SyncIQ worker activity.
200In-flight encryption makes data transfer between OneFS clusters secure. The
function benefits customers who undergo regular security audits and/or
government regulations.
201The SyncIQ bandwidth setting at the global level splits the bandwidth
reservation evenly among all policies. Using the CLI, you can make bandwidth
reservations for individual policies.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 311
Configuring Data Services
Limitations
203 Performing a complete failover and failback test on a monthly or quarterly basis
is discouraged. Perform failover testing if quiescing writes to the source (prevent
changing the data) and all SyncIQ policies are successfully run a final time to
assure complete synchronization between source and target. Failing to perform a
final synchronization can lead to data loss.
PowerScale Administration-SSP1
Compatibility
The table shows the versions of OneFS you can synchronize using SyncIQ. Target
cluster running OneFS 7.1.x version of OneFS is no longer supported. For
information about the support and service life-cycle dates for hardware and
software products, see the Isilon Product Availability Guide.
204Retrieving a copy of the data from the target cluster does not require a failover.
The target is a read-only copy of the data. Perform a copy operation to make a
copy of the read-only data on the target cluster to a location outside of the SyncIQ
domain on the target, or to a location on the source cluster, or to the client.
205The 'Whenever the source is modified' option is not for continuous replication.
OneFS does not offer a continuous replication option. This option is for specific
workflows that have infrequent updates and require distribution of the information
as soon as possible.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 313
Configuring Data Services
CloudPools
SyncIQ can synchronize CloudPools data from the CloudPools aware source
cluster to a PowerScale target cluster.
SyncIQ provides data protection for CloudPools data and provides failover and
failback capabilities.
The processes and capabilities of SyncIQ are based on the OneFS version
relationship between the source cluster and the target cluster. This relationship
determines the capabilities and behaviors available for SyncIQ policy replication.
Failover
Failover is the process of changing the role of the target replication directories into
the role of the source directories for assuming client read, write, and modify data
activities.
PowerScale Administration-SSP1
Source
Target
The example shows a failover where the client accesses data on the target cluster.
Failback
Like failover, you must select failback for each policy. You must make the same
network changes to restore access to direct clients to the source cluster.
206A failback can happen when the primary cluster is available once again for client
activities. The reason could be from any number of circumstances including the
natural disasters are no longer impacting operations, or site communication or
power outages have been restored to normal. You must failback each SyncIQ
policy.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 315
Configuring Data Services
Source
Target
The example shows a failback where the client accesses source data.
Failback Preparation
Source cluster
Resync-prep prepares the source cluster to receive the changes made to the data
on the target cluster.
The mirror policy is placed under Data Protection > SyncIQ > Local Targets on the
primary cluster. On the secondary cluster, the mirror policy is placed under Data
Protection > SyncIQ > Policies.
PowerScale Administration-SSP1
Failover Revert
A failover revert undoes a failover job in process207. Use revert before writes
occur208 on the target.
Source
Target
207Failover revert stops the failover job and restores the cluster to a sync ready
state. Failover reverts enables replication to the target cluster to once again
continue without performing a failback.
208Use revert if the primary cluster once again becomes available before any writes
happen to the target. A temporary communications outage or if doing a failover test
scenario are typical use cases for a revert.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 317
Configuring Data Services
SyncIQ
209You create and start replication policies on the primary cluster. A policy
specifies what data is replicated, where the data is replicated to, and how often the
data is replicated.
210The primary cluster holds the source root directory, and the secondary cluster
holds the target directory. There are some management capabilities for the policy
on both the primary and secondary clusters, though most of the options are on the
primary.
211SyncIQ jobs are the operations that do the work of moving the data from one
PowerScale cluster to another. SyncIQ generates these jobs according to
replication policies.
PowerScale Administration-SSP1
The panels describe the files for creating the SyncIQ policy. Refer to the student
guide for more information.
Settings
Creating a SyncIQ policy is done of the Data protection > SyncIQ > Policies page
or using the isi sync policy create command.
Unique name
The graphic shows the SyncIQ policy Settings fields. Click the image to enlarge.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 319
Configuring Data Services
SyncIQ domain root Replicates only listed paths and ignores unlisted -
use with caution
Target Cluster
Advnaced
The final segment of the policy creation are the advanced fields.
PowerScale Administration-SSP1
Prioritize policies
Data details written to /var/log/isi_migrate.log
Settings: In the Settings section, assign a unique name to the policy. Optionally you
can add a description of the policy. The Enable this policy box is checked by
default. If you cleared the box, it would disable the policy and stop the policy from
running. Next designate whether a Copy policy or a Synchronize policy. The
replication policy can be started using one of four different run job options:
Manually, On a Schedule, Whenever the source is modified, or Whenever a
snapshot of the source directory is taken.
Source cluster directories: In the Source Cluster criteria, the Source root directory
is the SyncIQ domain. The path has the data that you want to protect by replicating
it to the target directory on the secondary cluster. Unless otherwise filtered,
everything in the directory structure from the source root directory and below
replicates to the target directory on the secondary cluster.
Includes and excludes: The Included directories field permits adding one or more
directory paths below the root to include in the replication. Once an include path is
listed that means that only paths listed in the include path replicate to the target.
Without include paths all directories below the root are included. The Excluded
directories field lists directories below the root you want explicitly excluded from the
replication process. You cannot fail back replication policies that specify includes or
exclude settings. The DomainMark job does not work for policies with subdrectories
mentioned in Include or Exclude. Using includes or excludes for directory paths
does not affect performance.
File matching criteria: The File matching criteria enables the creation of one or
more rules to filter which files do and do not get replicated. Creating multiple rules
connect them together with Boolean AND or OR statements. When adding a new
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 321
Configuring Data Services
filter rule, click either the Add an “And” condition or Add an “Or” condition links. File
matching criteria says that if the file matches these rules then replicate it. If the
criteria does not match the rules, do not replicate the file.
Target: Snapshots are used on the target directory to retain one or more consistent
recover points for the replication data. You can specify if and how these snapshots
generate. To retain the snapshots SyncIQ takes, select Enable capture of
snapshots on the target cluster. SyncIQ always retains one snapshot of the most
recently replicated delta set on the secondary cluster to facilitate failover,
regardless of this setting. Enabling capture snapshots retains snapshots beyond
the time period that is needed for SyncIQ. The snapshots provide more recover
points on the secondary cluster.
Advanced: The Priority field in the Advanced settings section enables policies to be
prioritized. If more than 50 concurrent SyncIQ policies are running at a time,
policies with a higher priority take precedent over normal policies. If the SyncIQ
replication is intended for failover and failback disaster recovery scenarios,
selecting Prepare policy for accelerated failback performance prepares the
DomainMark for the failback performance. The original source SyncIQ domain
requires a DomainMark. Running a DomainMark during the failback process can
take a long time to complete. You can retain SyncIQ job reports for a specified
time. With an increased number of SyncIQ jobs in OneFS 8.0, the report retention
period could be an important consideration. If tracking file and directory deletions
that are performed during synchronization on the target, you can select to Record
deletions on synchronization.
Deep copy: The Deep copy for CloudPools setting applies to those policies that
have files in a CloudPools target. Deny is the default. Deny enables only stub file
replication. The source and target clusters must be at least OneFS 8.0 to support
Deny. Allow the SyncIQ policy determine if a deep copy should be performed.
Force automatically enforces a deep copy for all CloudPools data that are
contained within the SyncIQ domain. Allow or Force are required for target clusters
that are not CloudPools aware.
PowerScale Administration-SSP1
A SyncIQ policy can copy or synchronize source data to meet organizational goals.
When creating a SyncIQ policy, choose a replication type of either sync 212 or
copy213.
212 If a mirrored copy of the source is the goal, create a sync policy.
213If the goal is to have all source data that is copied and to retain deleted file
copies, then create a copy policy.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 323
Configuring Data Services
Tip: You can always license SnapshotIQ on the target cluster and
retain historic SyncIQ associated snapshots to aid in file deletion
and change protection.
The video details a basic SyncIQ use case, configuring replication between two
clusters. See the student guide for a transcript of the video.
Link:
https://edutube.emc.com/Player.aspx?vno=6cyyA4XvBqkyHJwXs6ltdg==&autoplay
=true
PowerScale Administration-SSP1
Challenge
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 325
Configuring Data Services
SmartLock
Scenario
SmartLock Overview
PowerScale Administration-SSP1
• SyncIQ integration214
• OneFS data services integration215
SmartLock Concepts
Before configuring SmartLock on a cluster, you must familiarize yourself with a few
concepts to fully understand the SmartLock requirements and capabilities.
• Retention Period
• Compliance
• WORM
There are two SmartLock operation modes available to the cluster: SmartLock
compliance mode216 and SmartLock enterprise mode217.
Compliance Enterprise
216
You can create compliance directories only if the cluster has been upgraded to
SmartLock compliance mode.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 327
Configuring Data Services
Only use if SEC 17a-4 must be Does not restrict cluster to follow SEC
followed. 17a-4 rules.
Configured during initial cluster install. Data not modified until retention dates
have passed.
PowerScale Administration-SSP1
2: Enterprise SmartLock directories are data retention directories that do not meet
SEC regulatory compliance requirements. Enterprise directories are the most
commonly used directories in a SmartLock configuration. Enterprise SmartLock
directories enable administrators or RBAC enabled users the ability to delete files,
which are known as privileged deletes. You can enable or turn on, temporarily
disable or turn off, or permanently disable privileged deletes. The Enterprise
directory may be fully populated with data or empty when creating or modifying.
3: Compliance SmartLock directories are data retention directories that meet SEC
regulatory compliance requirements. Set up the cluster in Compliance mode to
support Compliance SmartLock directories.
When using SmartLock, there are two types of directories: enterprise and
compliance. A third type of directory is a standard or non-WORM218 directory.
If using the compliance clock, you must copy data into the Compliance SmartLock
directory structure before committing the data to a WORM state.
SmartLock Configuration
In this use case the administrator wants to create a WORM directory where files
are locked down for a month. Once moved into the folder, the files are committed to
WORM.
Create a WORM domain from the WebUI File system > SmartLock page and select
Create domain or using the CLI "isi worm domains command.
218
OneFS supports standard non-WORM directories on the same cluster with
SmartLock directories.
219When you upgrade, privileged deletes are disabled permanently and cannot be
changed back.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 329
Configuring Data Services
5
6
1: Setting to "On" enables the root user to delete files that are currently committed
to a WORM state.
3: The default retention period is assigned when committing a file to a WORM state
without specifying a day to release the file from the WORM state.
4: The minimum retention period ensures that files are retained in a WORM state
for at least the specified period of time. The maximum retention period ensures that
files are not retained in a WORM state for more than the specified period of time.
5: After a specified period, a file that has not been modified is committed to a
WORM state.
6: Files committed to a WORM state are not released from a WORM state until
after the specified date, regardless of the retention period.
Use case:
• The administrator requires a WORM directory where files are in a WORM state
for at least 30 days and are removed from the WORM state after 60 days.
PowerScale Administration-SSP1
CLI:
For a file to have a file retention date applied, and set to a read-only state, you
must commit the file to WORM.
Until the files are committed to WORM, files that are in a SmartLock directory act
as standard files that you can move, modify, or delete.
First set the retention date on the file, then Set per SmartLock domain
commit the file to WORM.
Sets a time period from when
the file was last modified on a
directory
Commit files to WORM state using Windows After the time period expires,
controls or UNIX commands the file is automatically
Example: # chmod ugo-w committed to WORM.
/ifs/finance/worm/JulyPayroll.xls
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 331
Configuring Data Services
SmartLock Considerations
Challenge
PowerScale Administration-SSP1
Monitoring Tools
PowerScale Administration-SSP1
PowerScale HealthCheck
Scenario
PowerScale Administration-SSP1
HealthCheck Overview
WebUI, Cluster management > HealthCheck page. Click the image to enlarge.
The OneFS HealthCheck tool is a service that helps evaluate the cluster health
status and provides alerts to potential issues.
You can use HealthCheck to verify the cluster configuration and operation,
proactively manage risk, reduce support cycles and resolution times, and improve
uptime.
CLI example to view the checklist items: isi healthcheck checklists list
PowerScale Administration-SSP1
For the CLI equivalent output use the "isi healthcheck checklists view cluster_capacity" command.
Click on the image to enlarge.
The graphic shows that the checklist items for the cluster_capacity check. The
HealthCheck terms and their definition are:
You can use the CLI to view the parameters of a checklist item. The example
shows viewing the node_capacity item parameters.
PowerScale Administration-SSP1
Running a HealthCheck
By default, a HealthCheck evaluation runs once a day at 11:00 AM. You can run a
HealthCheck using the WebUI.
The example shows selecting the Run option for the cluster_capacity checklist. The
HealthCheck table shows the status of the checklist.
PowerScale Administration-SSP1
HealthCheck Schedule
You can manage the HealthCheck schedules of the checklists. By default, the
basic checklist is scheduled.
PowerScale Administration-SSP1
Viewing an Evaluation
Evaluation showing
failures
You can view the evaluation from the HealthChecks tab or the Evaluations tab. For
a failed evaluation, the file will show the checklist items that failed.
PowerScale Administration-SSP1
HealthCheck Resources
Challenge
PowerScale Administration-SSP1
InsightIQ
Scenario
InsightIQ Overview
isi_stat_d
InsightIQ host
http
Client
FSA datastore
http
Datastore
InsightIQ focuses on PowerScale data and performance. Listed are key benefits for
using InsightIQ. Refer to the student guide for more information.
• Determine whether a storage cluster is performing optimally.
• Compare changes in performance across multiple metrics, such as CPU usage,
network traffic, protocol operations, and client activity.
• Correlate critical storage cluster events with performance changes.
• Determine the effect of workflows, software, and systems on storage cluster
performance over time.
PowerScale Administration-SSP1
InsightIQ Dashboard
PowerScale Administration-SSP1
Aggregated view
Metrics
Monitoring 3 clusters
Cluster health
Cluster-by-cluster breakout
You can modify the view to represent any time period where InsightIQ has
collected data. Also, breakouts and filters can be applied to the data. In the
Aggregated Cluster Overview section, you can view the status of all monitored
clusters as a whole. There is a list of all the clusters and nodes that are monitored.
Total capacity, data usage, and remaining capacity are shown. Overall health of the
clusters is displayed. There are graphical and numeral indicators for connected
clients, active clients, network throughput, file system throughput, and average
CPU usage. Depending on the chart type, preset filters enable you to view specific
data. For example, In/Out displays data by inbound traffic compare with outbound
traffic.
You can also view data by file access protocol, individual node, disk, network
interface, and individual file or directory name. If displaying the data by the client
only, the most active clients are represented in the displayed data. Displaying data
by event can include an individual file system event, such as read, write, or lookup.
Filtering by operation class displays data by the type of operation being performed.
PowerScale Administration-SSP1
Capacity Analysis
The capacity analysis pie chart is an estimate of usable capacity is based on the
existing ratio of user data to overhead220.
220 There is an assumption that data usage factors remain constant over more use.
If a customer uses the cluster for many small files and then wants to add some
large files, the result is not precisely what the system predicts.
PowerScale Administration-SSP1
Default Reports
You can monitor clusters through customizable reports that display detailed cluster
data over specific periods of time.
• Performance reports
• File system reports
• Live reporting
PowerScale Administration-SSP1
You can drill down to file system reporting to get a capacity reporting interface that
displays more detail about usage, overhead and anticipated capacity.
The graphic shows the Capacity Forecast, displaying the amount data that can be
added to the cluster before the cluster reaches capacity.
The administrator can select cluster information and use that as a typical usage
profile to estimate when the cluster reaches 90% full. The information is useful for
planning node/cluster expansion ahead of time to avoid delays around procurement
and order fulfillment.
The Plot data shows the granularity of the reporting available. The Forecast data
shows the breakout of information that is shown in the forecast chart. Depending
on the frequency and amount of variation, outliers can have a major impact on the
accuracy of the forecast usage data.
Create custom live performance reports by clicking Performance Reporting > Create a New
Performance Report. Click the image to enlarge.
There are three types of reports On the Create a New Performance Report page.
PowerScale Administration-SSP1
221In the Create a New Performance Report area, in the Performance Report
Name field, type a name for the live performance report. Select the Live
Performance Reporting checkbox. In the Select the Data You Want to See area,
specify the performance modules that you want to view in the report. You can add
a performance module or modify an existing one. Repeat this step for each
performance module that you want to include. Save the report.
PowerScale Administration-SSP1
InsightIQ collects the FSA data from the cluster for display to the administrator.
PowerScale Administration-SSP1
Enable FSA
Monitored Clusters page, Settings > Monitored Clusters. Click the image to enlarge.
Before you can view and analyze data usage and properties through InsightIQ, you
must enable the FSA feature.
222 Unlike InsightIQ datasets, which are stored in the InsightIQ datastore, FSA
result sets are stored on the monitored cluster in the /ifs/.ifsvar/modules/fsa
directory.
223The job collects information across the cluster, such as the number of files per
location or path, the file sizes, and the directory activity tracking.
PowerScale Administration-SSP1
To enable FSA, Open the Monitored Clusters page by clicking Settings > Monitored
Clusters. In the Actions column for the cluster that you want to enable or disable
FSA, click Configure. The Configuration page displays. Click the Enable FSA tab.
To enable the FSA job, select Generate FSA reports on the monitored cluster. To
enable InsightIQ for FSA report, select View FSA reports in InsightIQ.
If there are long time periods between the FSAnalyze job runs, the snapshot can
grow very large, possibly consuming much of the cluster's space. To avoid large
snapshot, you can disable the use of snapshots for FSAnalyze. Disabling snapshot
use means that the jobs may take longer to run.
Considerations
PowerScale Administration-SSP1
Challenge
PowerScale Administration-SSP1
DataIQ v1
Scenario
Your Challenge: The IT manager has asked you to explain DataIQ and
its available monitoring capabilities.
DataIQ Overview
1 2 3 4 5 6 7
1: DataIQ eliminates the problem of data silos by proving a holistic view into
heterogeneous storage platforms on-premises and in the cloud. A single pane of
glass view gives users a file-centric insight into data and enables intuitive
navigation.
2: DataIQ optimized near real-time scan, and high-speed file indexing deliver
immediate project and user information. Powerful search capabilities across
heterogeneous storage can locate data in seconds, no matter where it resides.
High-speed search and indexing scans and organizes files in "look aside" mode.
PowerScale Administration-SSP1
3: DataIQ can ‘tag’ an attribute and use that tag to query millions of files across any
storage system. Tags enable business users, and IT, to view data in a true
business context. Tags give organizations the ability to see their data in the right
context, and to optimize their storage environment costs.
4: DataIQ enables data mobility with bi-directional movement between file and
object storage. The use of self-service archive capabilities to move files to the most
appropriate storage tier, such as archive or the cloud, empowers business owners.
Self-service enables content owners to move data from high-performance file
storage to an object archive.
6: DataIQ quickly scans file and object storage of all types. It can classify data
according to customer specification and provide instant rollup information. For
example, total tree size, average age of subtree data, 'last modified' date at any
point of folder structure. DataIQ generates fast and granular reports with business-
specific views and metrics, enabling rapid issue isolation. DataIQ integrates with IT
infrastructures to provide rights for AD and LDAP for users and group, as well as
APIs to enhance and extract business data. DataIQ plug-ins enable users to gain
additional insights. Plug-ins extend the GUI and launch internal scripts such as
Data Mover, Previewer, Audited Delete, Send to QA, and other custom scripts.
DataIQ Implementation
The DataIQ server scans the managed storage, saves the results in an index, and
provides access to the index.
Access is available from one or more GUI clients, CLI clients, and through the API
for application integration.
PowerScale Administration-SSP1
DataIQ
Windows
Clients
Linux
Clients
MAC
Clients
After logging in to the DataIQ WebUI, the landing page is the Data Management
page.
Settings - Pages
Use the left and right arrows to view the Settings pages.
PowerScale Administration-SSP1
Local settings
The Local settings page allows you to personalize the theme of the DataIQ WebUI.
• Client maps224
• Viewable files and folders225
General management
You can configure email alerts and SRS on the General management page.
If a volume has the minimum free space threshold configured, an email is sent
when the threshold is triggered.
224 Client maps enable you to map the DataIQ path to the path that the client sees.
225 You can view or hide the hidden-type files and folders. You can also set how the
files and folders are viewed in the tables and lists.
PowerScale Administration-SSP1
The Access and permissions page is where you can configure groups, add roles to
the groups, set authentication providers, and add users.
PowerScale Administration-SSP1
The Other settings include file type class and configuration files.
Licensing
From the Licensing page, you can manage and upload licenses generated from the
Software Licensing Central online portal.
PowerScale Administration-SSP1
Shown is an overview of the actions a user with the role of data owner can
perform. The actions are performed from the Data management settings page.
Use the left and right arrows to view the panels.
Volumes Panel
Text
Edits apply globally, settings at the volume Configure volume type, scan management, Volumes added to the scan group adopt the Change the scan management, delete the
level have precedence and hard link handling scan group settings, scan group settings have volume
precedence over volume settings
Click to enlarge.
From the Data management configuration page, Volumes panel, you can set
volume defaults, add and edit volumes, and create scan groups.
PowerScale Administration-SSP1
S3 Endpoints Panel
Click to enlarge.
226DataIQ enables you to setup the endpoint as a volume for scanning. To delete
an endpoint, go to the view breakout for the endpoint.
PowerScale Administration-SSP1
PowerScale Administration-SSP1
Click to enlarge.
From the Data management settings page, Other settings panel, you can configure
file type classes.227
The example shows the configuration files and a breakout of the clientmap file. Click to enlarge.
The Data management settings has four configurations files that you can edit. The
files are listed in the Others setting panel:
• Clientmap configuration files allows you to view file paths as they are seen by
the user.
• Data management configuration file allows you to change DataIQ settings
• Viewfilter configuration file allows you to restrict the view of folders by group
• Autotagging configuration file allows you to setup and define tags
227File type classes allow you to scan the volumes by a file type class. For
example, you can make a class called images and then add .jpeg, .png, and .gif
extensions to the class.
PowerScale Administration-SSP1
Scroll through the carousel to view each of the volume management areas. You
can double click the images to enlarge.
Volume defaults
Set a capacity
threshold. When
triggered, flags the
volume
Provides more
accurate reports on
volumes with hardlinks
The volume defaults are applied to new volumes and volumes without configured
settings.
The settings on volumes that are configured take precedence over the default
values.
Add Volume
The Add new volume window consists of three panels, the general settings, scan
configuration, and advanced settings.
PowerScale Administration-SSP1
Scan Groups
You can create scan groups and add volumes with the same scan, TCO, and
minimum free space trigger to the group.
Settings in the scan group have precedence to the settings on the volume.
Editing Volumes
If the volume belongs to a scan group and the scan group settings no longer apply,
you can remove the volume from the scan group and edit the volume settings.
PowerScale Administration-SSP1
settings for each configuration file. Select each page for an overview and use case
for the configurations.
228For example, a class that is called Video and a class that is called Image are
configured. The IT manager requests a report on the cost of video-type files and
the cost of image-type files. You can use the DataIQ Analyze feature to view the
storage consumption and cost of each class.
PowerScale Administration-SSP1
Clientmap Configuration
Format
Example mappings
Use the clientmap file to map virtual DataIQ paths to valid paths on a client.
Convert229 from virtual to client and from client path to virtual path.
229Conversion from virtual paths to client paths occurs when copying paths to the
system clipboard. Conversion from client paths to DataIQ virtual paths occurs when
a client path is entered into a field such as a search field.
PowerScale Administration-SSP1
Data Management
Format
Option definition
230Modifying settings can impact DataIQ functionality. The defaults are typically
used. The file has a description of each setting.
PowerScale Administration-SSP1
Viewfilter Configuration
Example filters
The Viewfilter configuration file231 enables you to create rules to restrict groups
from viewing folders.
231 Viewfilter uses regular expressions (RE). If a volume or folder matches the RE
for the user's group, then that volume and folder are viewable for the user. If a user
is a member of more than one group, the user is only restricted from folders that
are restricted in all their groups.
PowerScale Administration-SSP1
Autotagging Configuration
Use auto-tagging232 to tag and track items. A use case is applying a tag to project
paths for use when determining a work order for a customer.
Use the left and right arrows to view the Data Management pages.
PowerScale Administration-SSP1
Browse
Configure limits
and actions on
selected item
The main functions of the Browse page are searching233, a panel that shows the
volumes in a tree view, a directory breakdown panel, a table that shows the files
within the selected folder, and an article details panel.
Flagging items in the table makes them reflective in the other data management
components.
233The search bar uses characters similar to Java regular expression (regex) such
as ^ for the beginning of filenames and $ for the ending of filenames.
PowerScale Administration-SSP1
Browse Details
234However, if data changes, updated files may not appear in file searches. Go to
the Actions panel and perform a scan on a volume or path to make sure you are
getting the latest information.
PowerScale Administration-SSP1
Analyze
The Analyze page235 allows you to analyze volumes from a business context.
Flagged items
The Flagged items page lists the items the user marks as flagged.
235
The page enables you to view multi-dimensional project oriented data by cost
and size.
PowerScale Administration-SSP1
Tag management
Business rules configuration, also called auto-tagging, is used to tag tracked items
during a scan.
The Tag management page shows the results of scan when auto-tagging is
configured.
Jobs
The Jobs page shows a table of the jobs and their status as well as a details panel
for the selected job.
PowerScale Administration-SSP1
Logs - Scan
The Logs page has two tabs, the Scan logs and the Error logs. The Scan logs table
show the generated logs from completed scan jobs.
Logs - Error
236A full scans is done the first time a storage file system is indexed. DataIQ walks
the entire file system, indexing every folder. This initial baseline scan ensures that
everything about the file system is known.
237An optimized scan is an incremental scan that only scans the folders where
there have been changes since the last full scan.
PowerScale Administration-SSP1
Auto-Tagging Example
The installer does not create the autotagging configuration file, but you can use the
sample file /usr/local/dataiq/etc/autotag.cfg.sample as a starting
point. Auto-tagging generally occurs when DataIQ scans a file system.
First make a copy of the existing Autotagging configuration file as a backup. The
graphic shows the location of the Autotagging configuration file on the Settings,
Data management configuration page.
PowerScale Administration-SSP1
2. Reference Path
Enter the path examples on their own line, preceded by comment (#).
3. Auto-Tagging Rule
PowerScale Administration-SSP1
Enter the corresponding rule below each reference path. Having the commented
path makes it easier to understand the rule later and provides a reference for other
administrators.
Tags are automatically removed if the rule that created it no longer matches and
the tag has not been altered.
4. Simulate
Once the auto-tagging rules are configured, Simulate and report, and then view
the results. The results panel lists each rule and the number of times it matched. If
the results look reasonable, Save and run the new rules.
The Simulate and report will indicate rules that are invalid.
PowerScale Administration-SSP1
5. Analyze
Go to the Data Management page and watch the auto-tab job details to see when
it completes. View the counts in the details window. Go to the Analyze page to
verify the generated tag sets and view the report.
PowerScale Administration-SSP1
• Put the RE from an existing rule or rule fragment in the Stanford Analyzer to
understand it (select Java). Modify the RE in the analyzer until it meets your
needs.
• Test in an RE tester (search for "Java regular expression tester"), and then
put into DataIQ and run in the simulator.
Plug-In Overview
Plugins provide functions such as data transfer and audited delete to enable
administrators to manage data resources across storage platforms such as
PowerScale and ECS.
PowerScale Administration-SSP1
The plug-ins DataIQ supports are listed. Click each plug-in for a brief description.
• Data Mover
• Audited Deletes
• Duplicate Finder
• Previewer
Plug-in Examples
The graphics show WebUI excerpts of the plug-ins that are installed on a DataIQ
instance.
PowerScale Administration-SSP1
Challenge
Lab Assignment: Go to the lab and add the PowerScale cluster to the
DataIQ application.
PowerScale Administration-SSP1
isi statistics
Scenario
The three main commands that enable you to view the cluster from the command
line are isi status, isi devices, and isi statistics.
isi statistics
The isi statistics command provides protocol, drive, hardware, and node
statistics238.
238Other services such as InsightIQ, the WebUI, and SNMP gather information
using the "isi statistics" command.
PowerScale Administration-SSP1
The output shows the operations by protocol. The example shows that NFS clients
are connected to node 6 with 278.5k bytes per second input rate.
Output for the general cluster statistics in a top-style display where data is continuously overwritten
in a single table.
isi devices
The isi devices command displays information about devices in the cluster and
changes their status. There are multiple actions available including adding drives
and nodes to the cluster. Use the isi devices command for drive states,
hardware condition, node management, and drive replacement management.
isi status
The isi status command displays information about the current status of the
cluster, alerts, and jobs. The example of the isi status output gives a general node
status, performance metrics, critical alerts, and Job Engine status.
PowerScale Administration-SSP1
The --quiet option omits the alerts and Job Engine status output.
Tip: See the CLI Reference guide for a complete list of the
command options and output definitions.
The isi statistics command dumps all collected stats, and you can run the
"query" subcommand on a specific statistic.
• You can build a custom isi statistics query that is not in the provided
subcommands
• Cluster and node statistics from kernel counters
• isi_stats_d
PowerScale Administration-SSP1
The isi statistics command within a cron job239 gathers raw statistics over a
specified time period.
239 A cron job can run on UNIX-based systems to schedule periodic jobs.
PowerScale Administration-SSP1
The example output shows the isi statistics drive command for the SSD
drives on node 6.
The examples shows isi statistics heat, which uses --long to include
more columns.
The head -10 option displays the first 10 most active most accessed files and
directories.
The example node 6 output shows the Timestamp in Epoch timestamp format,
Ops as protocol operations, the Event type and Class (getattr is a namespace
read), and LIN for the file or directory associated with the event.
PowerScale Administration-SSP1
Practical Skills
Combining large sets of collected data with log analysis can help identify long-term
trends and sources of trouble.
2: isi Statistics can fill the gaps. Skillful use of isi statistics can
produce equivalent information to what InsightIQ offers and contains many
performance-related options.
PowerScale Administration-SSP1
Challenge
Lab Assignment: Now that you know which CLI commands are
available for monitoring, go to the lab and run the isi statistics
command.
PowerScale Administration-SSP1
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 389
Appendix
Course Materials
• Participant Guide
• Instructor laptop
• Projector and Speakers
• Internet access
• Whiteboard and markers
PowerScale Administration-SSP1
Course Agenda
Lunch
Depending on course pace and student knowledge, module and lab exercise
schedule may be altered
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 391
Appendix
Introductions
• Name
• Company
• Job Role
• Experience
• Expectations
PowerScale Administration-SSP1
DNS Primer
1: A FQDN is the DNS name of an object in the DNS hierarchy. A DNS resolver
query must resolve an FQDN to its IP address so that a connection can be made
across the network or the Internet. If a computer cannot resolve a name or FQDN
to an IP address, the computer cannot make a connection, establish a session or
exchange information. An example of an FQDN looks like sales.isilon.xattire.com.
2: A single period (.) represents the root domain, and is the top level of the DNS
architecture.
3: Below the root domain are the top-level domains. Top-level domains represent
companies, educational facilities, nonprofits, and country codes such as *.com,
*.edu, *.org, *.us, *.uk, *.ca, and so on. A name registration authority manages the
top-level domains.
4: The secondary domain represents the unique name of the company or entity,
such as EMC, Isilon, Harvard, MIT.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 393
Appendix
5: The last record in the tree is the hosts record, which indicates an individual
computer or server.
PowerScale Administration-SSP1
What is an A record?240
For example, a server that is named centos would have an A record that mapped
the hostname centos to the IP address assigned to it: centos.dees.lab A
192.168.3.3 Where centos is the hostname, dees.lab is the domain name, and
centos.dees.lab is the FQDN.
240
An A-record maps the hostname to a specific IP address to which the user
would be sent for each domain or subdomain. It is simple name-to-IP resolution.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 395
Appendix
The Name Server Record, or NS records, indicate which name servers are
authoritative for the zone or domain.
241Companies that want to divide their domain into sub domains use NS records.
Sub domains indicate a delegation of a portion of the domain name to a different
group of name servers. You create NS records to point the name of this delegated
sub domain to different name servers.
PowerScale Administration-SSP1
You must create an address (A) record in DNS for the SmartConnect service IP.
Delegating to an A record means that if you failover the entire cluster, you can do
so by changing one DNS A record. All other name server delegations can be left
alone. In many enterprises, it is easier to update an A record than a name server
record, because of the perceived complexity of the process.
Delegationtion recommendation.242
242 The recommendation is to create one delegation for each SmartConnect zone
name or for each SmartConnect zone alias on a cluster. This method permits
failover of only a portion of the workflow—one SmartConnect zone—without
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 397
Appendix
affecting any other zones. This method is useful for scenarios such as testing
disaster recovery failover and moving workflows between data centers.
PowerScale Administration-SSP1
The graphic shows how SmartConnect uses the X-Attire DNS server to provide a
layer of intelligence within the OneFS software application.
6
5
4 7
1
2
3: All clients are configured to make requests from the resident DNS server using a
single DNS hostname. Because all clients reference a single hostname,
isilon.xattire.com, it simplifies the management for large numbers of clients.
4: The resident DNS server forwards the delegated zone lookup request to the
delegated zone server of authority, here the SIP address of the cluster.
6: SmartConnect then returns this information to the DNS server, which, in turn,
returns it to the client.
7: The client then connects to the appropriate cluster node using the wanted
protocol.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 399
Appendix
NFS Connectivity
NFS relies upon remote procedure call (RPC) for client authentication and port
mapping. RPC is the NFS method that is used for communication between a client
and server over a network. RPC is on Layer 5 of the OSI model. Because RPC
deals with the authentication functions, it serves as gatekeeper to the cluster.
NFS connectivity
PowerScale Administration-SSP1
Let us look at the flow of a request by a client. When the RPC services start up on
the cluster, it registers with portmapper. The service tells portmapper what port
number it is listening on, and what RPC program numbers it is prepared to serve.
244 When the server receives the CALL, it performs the service that is requested
and sends back the REPLY to the client. During a CALL and REPLY, RPC looks for
client credentials, that is, identity and permissions.
245
If the server is not running a compatible version of the RPC protocol, it sends an
RPC_MISMATCH. If the server rejects the identity of the caller, it sends an
AUTH_ERROR.
246It acts as a gatekeeper by mapping RPC ports to IP ports on the cluster so that
the right service is offered.
247Clients calling for an RPC service need two pieces of information, the number of
the RPC program it wants to call and the IP port number.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 401
Appendix
HDFS Topic
• Data Lakes and Analytics
• HDFS Overview Video
• OneFS with Hadoop
• OneFS vs. Hadoop
• HDFS Administration
• Best Practices Resources
• Troubleshooting Resources
PowerScale Administration-SSP1
Swift Topic
• File and Object Storage Differences
• Accounts, Containers, and Objects
• Configuring Isilon Swift Accounts
• Storage URL
• Isilon Swift Considerations and Limitations
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 403
Appendix
When a node boots, it first checks its own vault resources before querying its
paired node. This way if the node can recover its journal from its own resources,
there is no need to query the paired node. But, if the journal is bad, the node can
identify the journal condition from its node state block data, and recovery should be
possible. There is a consequence to the nodes running in pairs. If a node runs
unpaired, it is under-protected.
PowerScale Administration-SSP1
Concurrency Examples
The process of striping spreads all write operations from a client248 across the
nodes of a cluster. Each tab illustrates a file that is broken down into chunks, after
which it is striped across disks249 in the cluster along with the FEC.
Concurrency
N+1n
256 KB file
128 KB chunk
128 KB chunk
128 KB FEC
248 A client is connected to only one node at a time. However when that client
requests a file from the cluster, the client connected node does not have the entire
file locally on its drives. The client-connected node retrieves and rebuilds the file
using the back-end network.
249 Even though a client is connected to only one node, when that client saves data
to the cluster, the write operation occurs in multiple nodes. The scheme is true for
read operations also.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 405
Appendix
All files 128 KB or less are mirrored. For a protection strategy of N+1 the 128 KB
file has 2 instances, the original data and one mirrored copy.
Concurrency
N+1n
128 KB file
128 KB FEC
The example shows a file that is not evenly distributed in 128 KB chunks. Blocks in
the chunk that are not used are free for use in the next stripe unit. Unused blocks in
a chunk are not wasted.
128 KB chunk
64 KB used
The example shows +2d:1n protection of a 1 MB file. The file is divided into eight
data stripe units and three FEC units. The data is laid out in two stripes over two
drives per node to achieve the protection.
PowerScale Administration-SSP1
Concurrency
N+2d:1n
1 MB file
8 x 128 KB chunk
Blocks within the same stripe (stripe 1) are written to separate drives on
each node
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 407
Appendix
A Data Lake is a central data repository that enables organizations to access and
manipulate the data using various clients and protocols. The flexibility keeps IT
from managing and maintaining a separate storage solution (silo) for each type of
data such as SMB, NFS, Hadoop, SQL, and others.
Click the i buttons in the graphic for information about ingest and OneFS storage.
1 2
2: Utilizing Isilon to hold the Hadoop data gives you all of the protection benefits of
the OneFS operating systems. You can select any of the data protection levels that
OneFS offers giving you both disk and node fault tolerance.
PowerScale Administration-SSP1
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 409
Appendix
URL:
https://edutube.emc.com/Player.aspx?vno=wZCty171ec2RjiMSRZZe9g==&autopla
y=true
Shown is an Isilon cluster with twelve nodes. A key benefit of CloudPools is the
ability to interact with multiple cloud vendors. Shown in the graphic are the
platforms and vendors that are supported as OneFS 8.1.1.
PowerScale Administration-SSP1
Let us look at an example, each chassis in the cluster represents a tier of storage.
The topmost chassis is targeted for the production high-performance workflow and
may have node such as F800s. When data is no longer in high demand,
SmartPools moves the data to the second tier of storage. The example shows the
policy moves data that is not accessed and that is over thirty days old. Data on the
middle tier may be accessed periodically. When files are no longer accessed for
more than 90 days, SmartPools archive the files to the lowest chassis or tier such
as A200 nodes.
The next policy moves the archive data off the cluster and into the cloud when data
is not accessed for more than 180 days. Stub files that are also called SmartLinks
are created. Stub files consume approximately 8 KB space on the Isilon cluster.
Files that are accessed or retrieved from the cloud, or files that are not fully moved
to the cloud, have parts that are cached on the cluster and are part of the stub file.
The storing of CloudPools data and user access to data that is stored in the cloud
is transparent to users.
CloudPools files undergo a compression algorithm and then are broken into their 2
MB cloud data objects or CDOs for storage. The CDOs conserve space on the
cloud storage resources. Internal performance testing does note a performance
penalty for a plane compression and decompressing files on read. Encryption is
applied to file data transmitting to the cloud service. Each 128 KB file block is
encrypted using a AES 256 encryption. Then transmitted as an object to the cloud.
Internal performance testing notes a little performance penalty for encrypting the
data stream.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 411
Appendix
To recap the overview, all production data resides on PowerScale. This removes
the task of exporting it from your production applications and importing it as with a
traditional Hadoop environment. The MapReduce continues to run on dedicated
Hadoop compute nodes. PowerScale requires this Hadoop front end to do the data
analysis. PowerScale holds the data so that Hadoop, applications, or clients can
manipulate it.
PowerScale Administration-SSP1
250 Hadoop requires a landing zone to stage data before using tools to ingest data
to the Hadoop cluster. PowerScale enables cluster data analysis by Hadoop.
Consider the time that it takes to push 100 TB across the WAN and wait for it to
migrate before any analysis can start. PowerScale does in place analytics so no
data moves around the network.
251Hadoop assumes that all members of the domain are trusted. PowerScale
supports integrating with AD or LDAP, and gives you the ability to safely segment
access.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 413
Appendix
252Each physical HDFS cluster can only support one distribution of Hadoop.
PowerScale can co-mingle physical and virtual versions of any Apache standards-
based distributions.
253Hadoop pairs the storage with the compute, so adding more space may require
you to pay for more CPU that may go unused. If you need more compute, you end
up with a lot of overhead space. With PowerScale you scale compute as needed or
storage as needed, aligning your costs with your requirements.
PowerScale Administration-SSP1
HDFS Administration
The graphic shows the WebUI Protocols, Hadoop (HDFS), Settings page, and
the corresponding isi hdfs settings command output.
6 7
4
5
1: The Default block size determines how the HDFS service returns data upon
read requests from Hadoop compute client. The server-side block size determines
how the OneFS HDFS daemon returns data to read requests. Leave the default
block size at 128 MB. If the customer runs an older version of HDFS, consider a 64
MB block size. If the block size is set to high, many read/write errors and
performance problems occur. Tune on setup.
2: Default checksum type is used for old HDFS workflows. Because OneFS uses
forward error correction, checksums for every transaction are not used, as it can
cause a performance issue.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 415
Appendix
5: Odp version - on updates, the Hortonworks version must match the version that
is seen in Ambari. Version conflict is common when customer upgrades
Hortonworks. Can cause jobs not to run. Installation also fails when Odp version
does not match.
6: Proxy users for secure impersonation can be created on the Proxy users tab.
For example, create an Apache Oozie proxy user to securely impersonate a user
called HadoopAdmin. Enable the Oozie user to request that the HadoopAdmin user
perform Hadoop jobs. Apache Oozie is an application that can automatically
schedule, manage, and run Hadoop jobs.
7: On the Virtual racks tabs, nodes can be preferred along with an associated
group of Hadoop compute clients to optimize access to HDFS data.
PowerScale Administration-SSP1
• Visit the Using Hadoop with Isilon - Isilon Info Hub web page for documentation.
• Use the Isilon Hadoop tools to create users and groups in the local provider.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 417
Appendix
Troubleshooting Resources
There are several guides that are dedicated to troubleshooting an HDFS solution.
PowerScale Administration-SSP1
Object storage combines the data with richly populated metadata to enable
searching for information by file content. Instead of a file that tells you the create or
modified date, file type, and owner, you can have metadata that tells you the
project name, formula results, personnel assigned, location of test and next run
date. The rich metadata of an object store enables applications to run analytics
against the data.
Object storage has a flat hierarchy and stores its data within containers as
individual object. An object storage platform can store billions of objects within its
containers, and you can access each object with a URL. The URL associated with
a file enables the file to be located within the container. Hence, the path to the
physical location of the file on the disk is not required. Object storage is well suited
for workflows with static file data or cloud storage.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 419
Appendix
Metadata:
Rich metadata:
File Name: Formula 5Xa
Object ID: 98765 Level: xxx
Created by: M.Smith
File Type: .doc
Created on: 9/9/14 Test date: xxx
Lab facility: Atlanta
File type: Word Patient trial: xxx
Building: 7
Patent: xxx
Lead Scientist: M. Smith
Approval ID: xxx
Description: xxx
Risk Assessment: xxx
PowerScale Administration-SSP1
Object1
Administrative
Container1 Object2
control point
Contain user
data
Object3
Account
Object4
Container1
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 421
Appendix
Administrators must provision the accounts before users can use the service. The
general steps are enable Swift license, decide upon file system user or group
ownership, create accounts using the isi swift command, and then assign
users access to account. Make any necessary file system permission changes if
you are relocating data into the account.
The example shows creating a Swift account in the sales access zone and using
an Active Directory user and group. The isi swift accounts list shows the
accounts that are created in the access zone. The isi swift accounts view
shows the account details.
PowerScale Administration-SSP1
Storage URL
Shown is what a Swift Storage URL looks like. URIs identify objects in the form
http://<cluster>/v1/account/container/object. In the example shown,
192.168.0.1 identifies the cluster. HTTP requests are sent to an internal web
service listening on port 28080. This port is not configurable. HTTPS requests are
proxied through the Apache web server listening on port 8083. This port is not
configurable. OpenStack defines the protocol version /v1. The reseller prefix
/AUTH_bob, where /AUTH is a vestige of the OpenStack implementation's internal
details. The _bob portion of the URL is the account name used. The container /c1
is the container in which an object is stored and the object /obj1 is the object.
Web service
Cluster Protocol version Reseller prefix Account Container Object
listening port
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 423
Appendix
Pre OneFS 8.0 Swift accounts are deactivated when upgrading to OneFS 8.0 and
later. After the upgrade, Swift no longer uses home directories for accounts. The
upgrade plan should determine which users are using Swift. Create new accounts
under the new Swift path, and then move the data from the old accounts into the
newly provisioned accounts. Swift is not compatible with the auditing feature.
PowerScale Administration-SSP1
Cache - L2
Storage side or node-side buffer. Buffers write transactions and L2 writes to disk
and prefetches anticipated blocks for read requests, sometimes called read ahead
caching. For write transactions, L2 cache works with the journaling process to
ensure protected committed writes. As L2 cache becomes full, it flushes according
to the age of the data. L2 flushes the least recently used, or LRU, data.
Chimer Nodes
By default, if the cluster has more than three nodes, three of the nodes are
selected as chimers. If the cluster has four nodes or less, only one node is selected
as a chimer. If no external NTP server is set, nodes use the local clock. Chimer
nodes are selected by the lowest node number that is not excluded from chimer
duty.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 425
DataIQ Previewer Plug-in
The Preview plug-in shows a preview image of the file in the WebUI for common
file types. The supported graphic file extensions are: ".jpg", ".jpeg", ".tiff", ".tif",
".dpx", ".bmp", ".png", ".gif", ".tga", ".targa", ".exr", ".pcx", ".pict", ".ico". The
supported video file extensions are: ".mov", ".mp4", ".mpeg", ".mpg", ".ts", ".avi",
".mkv", ".wmf", ".wmv", ".mxf", ".ogv". The plug-in does not work with object stores
such as S3, GCP, or ECS.
File Provider
A file provider enables you to supply an authoritative third-party source of user and
group information to a clustr. A third-party source is useful in UNIX and Linux
environments that synchronize /etc/passwd, /etc/group, and etc/netgroup
files across multiple servers.
Generation 6 Hardware
The Gen 6 platforms reduce the data center rack footprints with support for four
nodes in a single 4U chassis. It enable enterprise to take on new and more
demanding unstructured data applications. The Gen 6 can store, manage, and
protect massively large datasets with ease. With the Gen 6, enterprises can gain
new levels of efficiency and achieve faster business outcomes.
PowerScale Administration-SSP1
Groupnet
The groupnet is a top-level networking container that manages hostname resolution
against DNS nameservers and contains subnets and IP address pools. Every
subnet is assigned to a single groupnet. Each cluster has a default groupnet
named groupnet0. Groupnet0 contains an initial subnet, subnet0, an initial IP
address pool, pool0, and an initial provisioning rule, rule0. Groupnets are how the
cluster communicates with the world. DNS client settings, such as name servers
and a DNS search list, are properties of the groupnet. If the cluster communicates
to another authentication domain, it must find that domain. To find another
authentication domain, you need a DNS setting to route to that domain. With
OneFS 8.0 and later releases, groupnets can contain individual DNS settings,
whereas prior OneFS versions had a single global entry.
Hadoop
Hadoop is designed to scale up from a single server to thousands of servers.
Hadoop clusters dynamically scale up and down based on the available resources
and the required services levels. Performance varies widely for processing, and
queries can take a few minutes to multiple days depending on how many nodes
and the amount of data requested.
Home Directory
Home directory provisioning creates a single home share that redirects users to
their SMB home directories. If one does not exist, a directory is automatically
created.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 427
File system reports include data about the files that are stored on a cluster. The
reports have use if, for example, you want to identify the types of data being stored
and where that data is stored. Before applying a file system report, enable InsightIQ
File System Analytics for that cluster.
isi get
The isi get command displays the protection settings on an entire directory path or,
as shown, a specific file without any options. The POLICY or requested protection
policy, the LEVEL or actual protection, the PERFORMANCE or data access pattern
are displayed for each file. Using with a directory path displays the properties for
every file and subdirectory under the specified directory path. Output can show files
where protection is set manually. Mirrored file protection is represented as 2x to 8x
in the output.
Job - Schedule
With the Schedule options, you can start the job manually or set to run on a
regularly scheduled basis.
PowerScale Administration-SSP1
Layers of Access
• Protocol Layer - The first layer is the protocol layer. Protocols may be Server
Message Block, or SMB, Network File System, or NFS, File Transfer Protocol,
or FTP, or some other protocol.
• Authentication Layer - The authentication layer identifies a user using a system
such as NIS, local files, or Active Directory.
• Identity Assignment Layer - The third layer is identity assignment. This layer is
straightforward and based on the results of the authentication layer, but there
are some cases that need identity mediation within the cluster, or where roles
are assigned within the cluster that are based on user identity.
• Authorization Layer - Finally, based on the established connection and
authenticated user identity, the file and directory permissions are evaluated. The
evaluation determines whether the user is entitled to perform the requested data
activities.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 429
Leaf-Spine
Leaf-Spine is a two level hierarchy where nodes connect to leaf switches, and leaf
switches connects to spine switches. Leaf switches do not connect to one another,
and spine switches do not connect to one another. Each leaf switch connects with
each spine switch and all leaf switches have the same number of uplinks to the
spine switches.
Local Provider
Local authentication is useful when Active Directory, LDAP, or NIS directory
services are not configured or when a specific user or application needs access to
the cluster. Local groups can include built-in groups and Active Directory groups as
members
MTTDL
MTTDL is a statistical calculation that estimates the likelihood of a hardware failure
resulting in data loss. MTTDL is a system view of reliability and asks the question
“What happens when hardware does fail, and will I lose any data when it does?”
NAS
NAS is an IP-based, dedicated, high-performance file sharing and storage device.
NFS
Network File System, or NFS, is an open standard that UNIX clients use. The NFS
protocol enables a client computer to access files over a network. NFS clients
mount the OneFS export that is accessible under a client mountpoint. The
mountpoint is the directory that displays files from the server. The NFS service
enables you to create as many NFS exports as needed.
OneFS CLI
The command-line interface runs "isi" commands to configure, monitor, and
manage the cluster. Access to the command-line interface is through a secure shell
(SSH) connection to any node in the cluster.
PowerScale Administration-SSP1
PaaS
PaaS combined with approaches like continuous integration and deployment can
measure application development cycles in the days and weeks rather than months
or years. The combinations can dramatically reduce the time it takes from having
an idea to identifying insight, to action, and creating value.
PAPI
The PAPI is divided into two functional areas: one area enables cluster
configuration, management, and monitoring functionality, and the other area
enables operations on files and directories on the cluster. A chief benefit of PAPI is
its scripting simplicity, enabling customers to automate their storage administration.
PowerScale A200
The A200 is an ideal active archive storage solution that combines near-primary
accessibility, value and ease of use.
PowerScale A2000
The A2000 is an ideal solution for high density, deep archive storage that
safeguards data efficiently for long-term retention.
PowerScale F200
Ideal for low-cost all-flash node pool for existing Gen6 clusters. Ideal for small,
remote clusters.
PowerScale F600
Ideal for small, remote clusters with exceptional system performance for small
office and remote office technical workloads.
PowerScale F800
Use the F800 for workflows that require extreme performance and efficiency.
PowerScale F810
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 431
Use the F810 for workflows that require extreme performance and efficiency. The
F810 also provides high-speed inline data deduplication and in-line data
compression. It delivers up to 3:1 efficiency, depending on your specific dataset
and workload.
PowerScale H400
The H400 provides a balance of performance, capacity and value to support a wide
range of file workloads. It delivers up to 3 GB/s bandwidth per chassis and provides
capacity options ranging from 120 TB to 720 TB per chassis.
PowerScale H500
The H500 is a versatile hybrid platform that delivers up to 5 GB/s bandwidth per
chassis with a capacity ranging from 120 TB to 720 TB per chassis. It is an ideal
choice for organizations looking to consolidate and support a broad range of file
workloads on a single platform.
PowerScale H5600
The H5600 combines massive scalability – 960 TB per chassis and up to 8 GB/s
bandwidth in an efficient, highly dense, deep 4U chassis. The H5600 delivers inline
data compression and deduplication. It is designed to support a wide range of
demanding, large-scale file applications and workloads.
PowerScale H600
The H600 is Designed to provide high performance at value, delivers up to 120,000
IOPS and up to 12 GB/s bandwidth per chassis. It is ideal for high performance
computing (HPC) workloads that don’t require the extreme performance of all-flash.
Quotas - Accounting
Accounting quotas monitor, but do not limit, disk storage. With accounting quotas,
you can review and analyze reports to help identify storage usage patterns.
Accounting quotas assist administrators to plan for capacity expansions and future
storage requirements. Accounting quotas can track the amount of disk space that
various users or groups use.
Quotas - Advisory
Advisory quotas do not deny writes to the disk, but they can trigger alerts and
notifications after the threshold is reached.
PowerScale Administration-SSP1
Quotas - Enforcement
Enforcement quotas include the functionality of accounting quotas and enable the
sending of notifications and the limiting of disk storage. Enforcement quotas include
the functionality of accounting quotas and enable the sending of notifications and
the limiting of disk storage.
Reed-Solomon
OneFS uses the Reed-Solomon algorithm, which is an industry standard method to
create error-correcting codes, or ECC, at the file level.
Reed-Solomon
OneFS uses the Reed-Solomon algorithm, which is an industry standard method to
create error-correcting codes, or ECC, at the file level.
Scale-out Solution
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 433
Not all clustered NAS solutions are the same. Some vendors overlay a
management interface across multiple independent NAS boxes. This gives a
unified management interface, but does not unify the file system. While this
approach does ease the management overhead of traditional NAS, it still does not
scale well.
With scale-out, a single component (node) of a system or cluster contains the
performance, compute, and capacity. As the need for capacity or compute power
increases, you add more nodes to the cluster. The node is not equivalent to a
scale-up controller as disk capacity is not added to a node. The cluster scales out
as nodes you add nodes, making it a much more scalable solution than a scale-up
implementation.
Scale-up Solution
The two controllers can run active/active or active-passive. For more capacity, add
another disk array. Each of these components is added individually. As more
systems are added, NAS sprawl becomes an issue.
Scale-up Storage
Scale-up storage is the traditional architecture that is dominant in the enterprise
space. High performance, high availability single systems that have a fixed capacity
ceiling characterize scale-up.
Serial Console
The serial console is used for initial cluster configurations by establishing serial
access to the node designated as node 1.
SmartDedupe
OneFS deduplication saves a single instance of data when multiple identical
instances of that data exist, in effect, reducing storage consumption. Deduplication
can be done at various levels: duplicate files, duplicate blocks in files, or identical
extents of data within files. Stored data on the cluster is inspected, block by block,
and one copy of duplicate blocks is saved, thus reducing storage expenses by
reducing storage consumption. File records point to the shared blocks, but file
metadata is not deduplicated.
SmartLock Compliance
PowerScale Administration-SSP1
SmartLock WORM
SmartLock provides WORM (write-once/read-many) status on files. In a WORM
state, files can be read but not modified. "Committing" a file is changing a file from
a read/write state to a WORM state that has a retention expiration date. Files are
committed to a WORM state when using SmartLock.
SmartPools
SmartPools is a software module that enables administrators to define and control
file management policies within a cluster.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 435
A single tier has only one file pool policy that applies the same protection level and
I/O optimization settings to all files and folders in the cluster. The basic version of
SmartPools supports virtual hot spares, enabling space reservation in a node pool
for reprotection of data. OneFS implements SmartPools basic by default.
PowerScale Administration-SSP1
Snapshot Schedule
The most common method is to use schedules to generate the snapshots. A
snapshot schedule generates snapshots of a directory according to a schedule. A
benefit of scheduled snapshots is not having to manually create a snapshot every
time wanted. An expiration period should be assigned to the snapshots that are
generated, automating the deletion of snapshots after the expiration period.
SnapshotIQ
OneFS snapshots are used to protect data against accidental deletion and
modification. Because snapshots are available locally, users can restore their data
without administrative intervention.
Stateless Connection
A stateless connection maintains the session or “state” information about the client
side. If a node goes down, the IP address that the client is connected to fails over
to another node in the cluster. The client would not know that their original node
had failed.
WebUI
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 437
The browser-based OneFS web administration interface provides secure access
with OneFS-supported browsers. This interface is used to view robust graphical
monitoring displays and to perform cluster-management tasks.
Windows ACL
A Windows ACL is a list of access control entries, or ACEs. Each entry contains a
user or group and a permission that allows or denies access to a file or folder.
PowerScale Administration-SSP1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 439
LAB GUIDE -
POWERSCALE
ADMINISTRATION
Version [1]
LAB GUIDE
INSTRUCTOR GUIDE
Internal Use - Confidential
Dell Confidential and Proprietary
Copyright © 2020 Dell Inc. or its subsidiaries. All Rights Reserved. Dell, EMC and other
trademarks are trademarks of Dell Inc. or its subsidiaries. Other trademarks may be
trademarks of their respective owners.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 1
Exploring the Lab Environment
Note: Chrome is the preferred browser and delivers the best experience. If
you do not have Chrome, you can use the browser of your choice.
5. A Welcome video opens, providing a lab orientation. If this is the first time you
are using the VLP, view the video. If the video does not automatically play, in
the upper right corner of your browser, click Help > Tips to review the EduLab
Orientation Video. The orientation video shows you how to deploy the lab,
access the console, and view the lab guide.
6. To learn more about the VLP, watch the EduLab Orientation video mentioned
in the previous step.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 3
Exploring the Lab Environment
Instructor Note: Ensure that all the students are able to access the
VLP, assist if otherwise.
You are John and have worked at BMC for just over a month and have been
assigned the storage administrator for the cluster. Based on your storage
experiences, you acknowledge that the configuration and day-to-day maintenance
of storage infrastructure is complex and may require several administrators to
manage it effectively.
The IT manager has told you the previous NAS platform had challenges with
scaling the storage. Those systems were limited by their single file system
(namespace) size (16TB typical). BMC will need namespace capabilities of a much
larger scale and the ability to quickly adjust to unexpected growth as their customer
base expands.
The cluster has been racked, stacked, powered, and added to DNS. The IT
manager along with senior storage administrators have put together the necessary
documentation and steps required to configure the cluster. As a part of the team,
configure the cluster based on the documentation and record your progress using
the lab exercises. The initial implementation is targeted for three business units in
the BMC organization: Sales, Engineering and Marketing.
Your goal is to implement the PowerScale cluster so that the users from each
business unit can access the data. Once done, you will configure services
such as snapshots and quotas on the data.
• InsightIQ
• DataIQ v1
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 5
Exploring the Lab Environment
Note that for usability purposes, the password for ALL systems is
Dees12345!
Note that the URL for the PowerScale WebUI used throughout this lab is
https://192.168.3.11:8080, the IP address for node-1. You can manage the
cluster through any active cluster node by using the IP address of that node.
Although most lab exercises are broken down into the three methods, not all lab
exercises can be strictly accomplished using only the WebUI or only the CLI. Some
of the lab exercises will not have a skills-based structure.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 7
Exploring the Lab Environment
a. The first page of the lab exercise is the PEQ. If you are experienced
or if you prefer a more challenging experience, try using the PEQ
method.
b. Following the PEQ pages are the step-by-step instructions for the
WebUI. Remember, you do not have to use this method if you are
configuring using the PEQ or CLI method.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 9
Exploring the Lab Environment
Introduction to PowerScale
Scenario:
With the Implementation Engineer onsite, watch the Implementation Engineer run
the Configuration Wizard.
Your 4-node cluster is ready. Each node is named boston-<#>. Next, you will
verify that you can connect to the cluster via the OneFS WebUI and the CLI. Once
connected, check the basic health of the cluster and the available storage space.
For the introduction to the PowerScale, there is no skills-based exercises. You will
perform each task in a step-by-step manner.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 11
Introduction to PowerScale
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 13
Introduction to PowerScale
2. Login to the WebUI for your new PowerScale cluster. Click on the Log in
button after entering the following credentials:
b. Password: Dees12345!
3. On the Dashboard menu, the Cluster status tab is shown as default. You
may see failed events in the New event groups table. You can ignore the
events for the purposes of this lab.
4. Scroll down to the Status table. In the Status table, mouse over the icon in the
Status column for nodes with ID 1,2,3 and 4 to see their status description.
Note: A node may have a status other than healthy (green) because of the
number of installed drives. Healthy status is not required for the purposes of
these lab exercises.
TIP: You can click the ID number of each node to go to the Node status page.
a. Node 1 IP:
b. Node 2 IP:
c. Node 3 IP:
d. Node 4 IP:
Instructor Note: Prompt the students to find and view the IP address
for each node.
6. Examine the Cluster size chart presented in the Cluster Status tab and note
the available hard disk space. Check the chart to see the growth of available
space as you add additional nodes to your cluster.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 15
Introduction to PowerScale
Note: The lab image used does not reflect capacities you will see on a
production system.
Tip: Click the shortcut on the desktop or select from the task bar.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 17
Introduction to PowerScale
3. Login as root, press ENTER, then type the password Dees12345! and press
ENTER. Note the PowerScale OneFS version for your documentation.
4. According to the Implementation Engineer, the first thing you should always do
when getting to the CLI is to check the health of the cluster. To display
information about the cluster’s status, at the boston-1 prompt, press ENTER
after typing the isi status command.
Tip: Use pipe to more (| more) or to less (| less) to prevent from needing
to scroll to view the output. Use the spacebar to forward the output. The less
command is a more powerful version of more. When reading through large
text files, the less command loads each page into memory a page at a time,
making it more efficient.
Tip: You can also use the isi stat command, which has identical output.
Remember, all commands are case sensitive.
5. Notice that the health of the cluster and the nodes is OK. Also, note the initial
available storage size with just the first four nodes in the cluster. The numbers
on your virtual cluster may not match the screenshots exactly but the location
of the information, and where you should look, will match the highlighted fields
in the screenshots throughout this document.
Question: What is the avail capacity for the cluster? How does this compare
to what the WebUI reported?
Instructor Note: The WebUI and CLI report the same storage
capacity.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 19
Adding Nodes to the Cluster
Scenario:
Now that you have validated the initial cluster build, you need to add the two
remaining nodes to the cluster. You have been given the procedures for how to do
this using the CLI and the WebUI.
The objective of the lab is to add two nodes to the cluster. If needed, use the
WebUI or the CLI step-by-step walk through portion of the exercise for guidance.
a. Add one of the two remaining nodes to the cluster using the
Configuration Wizard. Open the node that you want to add from the
console.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 21
Adding Nodes to the Cluster
2. Since this node is not part of the cluster, you will see the Configuration Wizard
as shown:
3. To join Node 5 to the boston cluster using the Configuration Wizard, perform
the following actions:
a. At the Wizard prompt, type 2 and press Enter to join the node to an
existing cluster.
b. In Join cluster prompt, type 1 and press Enter to join Node 5 to the
boston cluster.
Note: This process may take a few minutes due to the nature of the
virtual environment.
b. At the boston-1 prompt, run the isi status command to verify that
node 5 has joined the cluster. Node 5 appears at the bottom of the list
with ID set to 5.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 23
Adding Nodes to the Cluster
3. The Add node to cluster window shows the nodes available to add. There
will be one node displayed. When multiple nodes are available, there is no way
to distinguish node order prior to adding them to the cluster. In a production,
real world implementation, when adding multiple nodes into a cluster, ensure
to get the serial number off the back of the node. You can join nodes to the
cluster in the order you want to join them, or the order they appear in the rack.
a. Select the node and then click Add node. The node joins the cluster.
Note: Do not add another node to the cluster until the first node is joined because
the nodes may join out of order and break your sequential IP numbering.
4. Now verify that the node is added, go back to the Dashboard, Cluster
overview page of the WebUI. On the Cluster status tab, confirm that the
node is added to the cluster. You may need to refresh the screen or click on
another option if the virtual environment is not responding.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 25
Adding Nodes to the Cluster
b. View the Node status of Node 6 by clicking the node ID in the cluster
status tab. Notice that the node is assigned the next IP address (ext-
1) from the configuration range so that the IP addresses are
sequential. The configuration range was created when the
Implementation Engineer initially set up the cluster.
Note: If the Status column is red, wait a minute or two for the node to
finish joining. The status turns green and the values populate once
the node fully joins the cluster.
a. Under Event Groups, look at the list of event notifications along with
the level in the Severity. If there are issues with the cluster or events
that you should be aware of, the Severity will be Warning or Critical.
You may not have any events such as those shown in the graphic.
The graphic shows event you may see because of the virtual
environment. Check the box at the top of the ID column to select all
the events.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 27
Adding Nodes to the Cluster
Scenario:
Now you have a 6 node cluster. Next, you need to validate the configuration of the
newly built cluster. The IT manager has asked you to use specific isi commands
to become familiar with the cluster, understand the hardware layout, and use the
built-in help options. OneFS relies on the backend network to communicate across
the nodes, so take a moment to see how this works.
This lab explores the OneFS isi command structure and syntax. You will perform
each task in a step-by-step manner.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 29
Using CLI commands
isi commands
In this task, use the built-in man isi and isi --help, to view the commands
and command structure.
2. To view the isi command structure with all of its subcommands and options,
at the boston-1 prompt, press ENTER after typing the isi --help or isi -
h command.
Tip: You must use the vertical scroll bar to view the entire content.
Alternatively, typing a pipe | and the word more (| more) is a common
addition to commands with a lot of output. | more enables you to view the
output one page at a time. Press ENTER to view more of the output.
3. Pick one of the commands from the output and type it at the prompt. For
example: isi job or isi sync and see what type of output or assistance
you get. Shown in the graphic is the output for running the isi ntp
command.
Tip: To view a basic description of any command and its available options,
type -h or --help option after the command and press ENTER. For
example: isi ntp --help
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 31
Using CLI commands
Tip: Type q to exit from the man command and return to the boston-1 prompt.
5. The tab complete is a useful command line feature that can auto-fill partially
typed commands, file names, options, paths, subcommands and more. To
perform tab complete, press TAB while typing a command and notice that it
will autocomplete the command you are typing or suggest options.
Instructor Note: Demonstrate the use of the tab complete feature for
autocomplete and option suggestions.
1. To view the hard drives for node 1, at the boston-1 prompt, press ENTER
after typing the isi devices drive list | grep dev command.
Note: The use of | grep in the command. The addition to the command
enables you to view only the output fields with the word "dev". Running the
command without grep will output many lines that show an EMPTY bay. Try
running without using | grep.
The output shows HEALTHY drives in node 1. Also, the JOURNAL drive in
the first drive bay is required because this virtual cluster does not have an
internal SSD to store it.
a. To view the journal drive for all nodes, run the isi_for_array isi
device drive list | grep -w da1 command. The grep -w
switch is used to find an exact match. Similarly, you can view all the
drives for all nodes by running the isi_for_array isi device
drive list | grep dev command.
2. To get the list of subcommands for statistics, at the boston-1 prompt, press
ENTER after typing the isi statistics --help command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 33
Using CLI commands
3. To view all the drives on this virtual node, at the boston-1 prompt, press
ENTER after typing the isi statistics drive | more command. Notice
in the output that there is a single SSD and the rest are SAS and UNKNOWN
drives. You can use grep to filter only what you want.
Note: grep'ing the Drive string in the command is to display the header. The -
e is used to display multiple strings.
4. To view the hardware in every bay, you can run isi devices for all six
nodes in the cluster. At the boston-1 prompt, press ENTER after typing the
isi_for_array 'isi devices drive list' | grep -ie Lnn -e
dev | more command.
Note the use of | grep and | more in the command. The string Lnn is used
to output the header. Try the isi_for_array 'isi devices drive
list' command without piping.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 35
Using CLI commands
Note: You can only run the console commands while in the Configuration
console. You must leave the Configuration console to run any other cluster
command.
Tip: You do not need to type isi before commands while in the Configuration
console.
3. To view options for the interface command, press ENTER after typing the
help interface command.
Question: What are the two valid interfaces shown in the output?
4. To view the internal networks on the cluster, press ENTER after typing the
interface command.
5. To view the configuration table with the network information for int-a and int-b,
press ENTER after typing the status command.
Question: What is the number of allocated IP address and the number of free
IP addresses available on the cluster?
6. Exit the Configuration console. Press ENTER after typing the exit command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 37
Designing a OneFS Directory Structure
Scenario:
The first thing to configure is the directory structure. The three BMC business units
to access the boston cluster are sales, engineering, and marketing. In this
exercise you will create a unique base directory structure for the engineering and
marketing business units. You will create the sales base directory as part of
another exercise.
Given in the lab exercise is the information needed to create the base directories. If
needed, use the WebUI or the CLI step-by-step walk through portion of the
exercise for guidance.
Base Directories
In this task, you will create two of the three required base directories. Note that you
will create the sales directory is a later lab exercise.
b. Verify that the base directories have been created by either looking
into the File system explorer page or by using the ls command.
Instructor Note: Yes. The WebUI default is 777 whereas the CLI is
755.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 39
Designing a OneFS Directory Structure
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 41
Designing a OneFS Directory Structure
d. Run the ls -la /ifs command and verify the base directory is
present.
NTP
Scenario:
Your goal is to provide user access to the cluster. Now that you have a 6-node
cluster, you need to synchronize all the cluster nodes to the same time source.
Time synchronization is critical for cluster functionality.
The IT manager has instructed you to synchronize the cluster’s Network Time
Protocol (NTP) services with the dees.lab Active Directory server. Time
synchronization is your first step to access management, you need to
configure the date and time reported by the cluster.
Given in the lab exercise is the information needed to configure NTP. If needed,
use the WebUI or the CLI step-by-step walk through portion of the exercise for
guidance.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 43
NTP
NTP
The Active Directory domain controller is on the Windows Server 2012R2 VM. This
VM is designated as the time server that the cluster will sync its time to.
Authentication will fail if the cluster time is more than 5 minutes off from the AD
domain controller.
b. Verify that the date, time, and timezone on the nodes is the same as
that on the domain controller. Change the timezone if they do not
match. To change the time and timezone on the nodes, navigate to
the Date & Time tab on the General settings page of the WebUI.
2. Navigate to the Cluster management, General settings page, and then click
the NTP tab.
a. In the NTP Servers field, enter 192.168.3.2, and then click Save
Changes. The server with IP address 192.168.3.2 is the time server
that you want to synchronize the cluster to. The IP address of the
NTP server that you have added is displayed on the NTP Servers
field.
3. Select the Date and time tab. Compare the node’s Date and Time and Time
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 45
NTP
Zone to the system time on the DC server and the Jumpstation. You will
need to launch the DC console and login as administrator/Dees12345!.
Notice that the node time is in 24-hour format.
a. Change the cluster time if the DC time zone is different from the time
zone set on the nodes. Use the Time zone drop-down list to
configure the cluster time and time zone to match the DC server
using 24-hour format. No matter where you are in the world, set the
time zone to America/New_York - EDT.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 47
NTP
2. At the boston-1 prompt, run the isi ntp servers create 192.168.3.2
command to add the NTP server. To verify that the server has been added,
run the isi ntp servers list command, and you will see the server
listed as shown below:
3. Verify that the date, time and timezone on the nodes are the same as that on
the DC server. You will need to launch the DC console and login as
administrator/Dees12345! Notice that the node time is in 24-hour format.
a. To check the date, time and timezone on the nodes, open the
Configuration console by running the isi config command at the
boston-1 prompt. Notice that the prompt changes to boston>>>
i. To check the current date and time, run the date at the
boston>>> prompt. Similarly, run the timezone command to
check the timezone set on the nodes. Verify that the timezone
is set to Eastern Time Zone as shown below.
ii. If the timezone on the DC server and the nodes do not match,
change the timezone on the nodes by running the timezone
Eastern Time Zone command at the boston>>> prompt.
Run the commit command to save changes.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 49
Authentication Providers
Authentication Providers
Scenario:
BMC uses a mix of NFS and SMB clients. The next step for achieving client access
is to configure your cluster to use both LDAP and Active Directory services to
authenticate clients. OneFS best practice is to always connect LDAP first before
joining the cluster to an AD domain. To complete the cluster authentication setup,
join the cluster to Active Directory (AD).
Given in the lab exercise is the information needed to configure the Authentication
Providers. If needed, use the WebUI or the CLI step-by-step walk through portion
of the exercise for guidance.
Authentication Providers
Configure LDAP and Active Directory.
1. Use the following information to configure your cluster to use LDAP for
authentication:
2. Use the following information to configure Active Directory for the cluster:
a. Before configuring Active Directory, ensure that the time on the nodes
and time on the Jumpstation has a time difference of less than five
minutes. You can configure Active Directory using the WebUI,
Authentication providers page or the isi auth ads command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 51
Authentication Providers
2. Navigate to Access, Authentication providers, and then click the LDAP tab.
Click Add an LDAP provider.
4. To verify that the LDAP provider has been added, go to the LDAP tab, in the
LDAP providers table, ensure the new LDAP provider is listed and that it has
a green dot in the Status column as shown below.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 53
Authentication Providers
2. To configure LDAP, at the boston-1 prompt, run the isi auth ldap
create Isilon-LDAP --base-dn="dc=dees,dc=lab" --server-
uris="ldap://centos" --groupnet=groupnet0 command.
a. At the boston-1 prompt, run the isi auth status command. This
command shows which providers are currently active on the cluster.
Verify that the LDAP provider created is listed and online.
b. At the boston-1 prompt, run isi auth ldap list to view all the
LDAP providers configured for the cluster.
4. Query the LDAP server and output the various attributes for users. At the
boston-1 prompt, run the ldapsearch -x -h centos -b
dc=dees,dc=lab command.
4. On the Join a Domain page, enter the fields highlighted in the graphic and
then click Join.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 55
Authentication Providers
a. Joining the domain may take several minutes. After the cluster
successfully joins the domain, you will get a success notification and
the status should be green in the Active Directory providers table.
2. To verify the date and time on the nodes, enter the Configuration console by
running the isi config command at the boston-1 prompt. Run the date
command at the boston>>> prompt. Ensure that the time on the nodes and
the Jumpstation have a time difference of less than five minutes. If not, change
the time at the nodes by running the date YYYY/MM/DD HH:MM:SS
command at the boston>>> prompt. Replace the values for the date and time
in the command. Exit the Configuration console by running the exit
command.
3. To configure Active Directory for the cluster, at the boston-1 prompt, run the
isi auth ads create dees.lab administrator --
password=Dees12345! --groupnet=groupnet0 command.
a. At the boston-1 prompt, run the isi auth status command. This
command shows which providers are currently active on the cluster.
Verify that the Active Directory provider that was just created is listed
and online.
b. At the boston-1 prompt, run isi auth ads list to view all the
Active Directory providers configured for the cluster.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 57
Authentication Providers
5. At the boston-1 prompt, run the isi auth users list --domain dees
command. The output shows the Active Directory users.
Scenario:
✓ Configured NTP
✓ Configured authentication
✓ Configured two of three base directories
The base directories are the root of each access zone you create. Now you will
create the access zones.
Given in the lab exercise is the information needed to configure your access zones.
If needed, use the WebUI or the CLI step-by-step walk through portion of the
exercise for guidance.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 59
Configuring Access Zones
Access Zones
The intent is to have three BMC organizations in unique authentication zones.
Configure three access zones, one for each BMC business unit that will access the
cluster.
1. Use the following information to configure the access zones on the cluster:
Hint: For the sales access zone, you need to create the base
directory while creating the access zone.
Question: What are the default permissions for the sales base
directory?
Instructor Note: The System access zone is the default access zone
configured for IFS.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 61
Configuring Access Zones
b. Enter the fields as shown in the graphic and then click Create zone.
Note: As the base directory for sales is not created, the checkbox for
Create zone base directory if it does not exist is checked.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 63
Configuring Access Zones
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 65
Configuring SmartConnect
Configuring SmartConnect
Scenario:
You have created the base directory structure, connected to the authentication
providers, and created the access zones. The next step is to configure
SmartConnect. You have enlisted the help of the DNS Administrator to get
SmartConnect configured and name resolution working for the cluster.
SmartConnect
1. Use the following information to configure the SmartConnect service IP (SSIP):
a. You can use the WebUI, Network configuration page or the isi
network subnets modify command to configure the SSIP.
Tip: If you need to configure a single SSIP for the cluster, enter the
same IP for the start and end IP values in the range.
i. Subnet: subnet0
2. Configure the IP address pools for the three SmartConnect zones. You can
use the WebUI, Network configuration page or the isi network pools
create command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 67
Configuring SmartConnect
3. Once the IP address pools are configured for each access zone, you need to
create the delegation records in DNS. Use the following information to create a
delegation record for each SmartConnect zone.
b. Create two new hosts in the forward lookup zone dees.lab named
sip9 and sip10. sip9 IP address is 192.168.3.9 and sip10 IP address
is 192.168.3.10.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 69
Configuring SmartConnect
a. Click Edit and update the fields as shown in the graphic and click on
Save changes.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 71
Configuring SmartConnect
Tip: In a hybrid cluster environment (mix of different node types), you may
only want to choose the interfaces of nodes that the business unit will access.
For example, in an environment with a mix of F800 and H400 nodes, the Sales
unit may need access only to the F800 nodes. In that case, choose only the
interfaces of the F800 nodes while creating the IP address pool for the Sales
unit.
3. Navigate to the Cluster management, Licensing page. Scroll down and click
on Manage trials. Select SmartConnect Advanced and click Start trial.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 73
Configuring SmartConnect
a. Click More on subnet0 and select Add pool. Enter the following to
create the engineering IP address pool:
• Add all the external interfaces of ext-1 to manage the incoming traffic
5. Click More on subnet0 and select Add pool. Enter the following to create the
marketing IP address pool:
• Add all the external interfaces of ext-1 to manage the incoming traffic
6. Verify that the IP address pools are successfully created and listed under
subnet0 as shown in the graphic.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 75
Configuring SmartConnect
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 77
Configuring SmartConnect
1. From the VLP, open the DC console. The DC virtual machine is the domain
controller and the DNS server. Login using administrator/Dees12345!
3. In the New Host window, enter the fields as shown and then select Add Host.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 79
Configuring SmartConnect
4. Add a second New Host using the name sip10 and the IP address
192.168.3.10
b. For the Delegated domain, type sales, click Next, and then click
Add on the Name Servers window.
d. On the Name Servers page, click Next, and then click Finish.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 81
Configuring SmartConnect
c. In the navigation pane, right-click dees.lab and select Reload. Log off
the DC virtual machine.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 83
RBAC and ZRBAC
Scenario:
Before continuing with configuring user access, the IT manager has asked you to
grant administrative access control to other members of the team who will perform
various administrative activities on the cluster. For each member, create a user
account and assign the user to a suitable role. Create a user with auditing
privileges for the cluster and a user with administrative privileges for the
Engineering business unit.
Given in the lab exercise is the information needed to create and assign users to
roles. If needed, use the step-by-step walk through portion of the exercise for
guidance.
a. View the default engineering and sales users. You can use the
WebUI, Users tab in the Membership and roles page or you can
use the isi auth users command.
b. View the default administrative roles. Use the WebUI, Roles tab in
the Membership and roles page or the isi auth roles
command.
Instructor Note: The access zones have their own distinct roles.
c. View the privileges for a role in the WebUI. Use the View/Edit button
next to the role name or use the isi auth roles view command.
Question: Is the list of the privileges different for each role? What
privilege enables login to the WebUI? What privilege enables SSH
login?
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 85
RBAC and ZRBAC
2. Create a Local user. You can use the WebUI, Users tab in the Membership
and roles page or the isi auth users create command.
Instructor Note: Add the AD provider to the System zone and choose
the user from the AD users.
Hint: To select the zone, use the --zone option in the CLI
command. For the WebUI, select the zone from the dropdown
menu.
3. Add users to roles. You can use the WebUI, Roles tab in the Membership
and roles page or the isi auth roles modify command.
a. User: Dante
i. Role: AuditAdmin
Instructor Note: No, because the AuditAdmin role does not have the
privileges.
b. User: engAdmin
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 87
RBAC and ZRBAC
2. To view the OneFS users that are created by default, navigate to the Access,
Membership and roles page. Click on the Users tab.
a. To view the users in the System zone, set the Current access zone
to System from the dropdown menu, and select the Providers from
the dropdown menu. The users under FILE: System are the core
accounts created by OneFS.
b. To view the users in the engineering zone, set the Current access
zone to engineering, and change the providers to LDAP: Isilon-
LDAP
c. Similarly view the users for different access zones and providers by
changing the Current access zone and Providers field.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 89
RBAC and ZRBAC
a. View the built-in roles for any zone by selecting that zone in the
Current access zone dropdown menu. Shown below is an example
of the built-in roles in the engineering zone. The engineering zone
has 2 built-in roles: ZoneAdmin and ZoneSecurityAdmin
b. To view the privileges for the ZoneAdmin role, click the View/Edit
button. In the View role details window, note the privileges.
i. You may view the privileges of any role in any zone in a similar
manner. Also, view the privileges of a role in the CLI and
notice the difference in the naming from that shown in the
WebUI.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 91
RBAC and ZRBAC
2. To list the OneFS users that are created by default, at the boston-1 prompt,
run the following commands:
b. Non-System zone:
ii. sales zone local users: isi auth users list --zone
sales
iv. Similarly, view the users for other access zones by replacing
the zone name.
3. To list the built-in administrator roles, at the boston-1 prompt, run the following
commands:
iii. Similarly, view the roles for other access zones by replacing
the zone name.
4. To view the privileges for the built-in roles, at the boston-1 prompt, run the
following commands:
a. For the SystemAdmin role in the System zone: isi auth roles
view SystemAdmin
b. For the ZoneAdmin role in the engineering zone: isi auth roles
view ZoneAdmin --zone engineering
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 93
RBAC and ZRBAC
c. Similarly, view the privileges for any role in any zone by replacing the
role and zone names.
2. Create a user account for user Dante in the System access zone.
b. On the Create a user dialog box, enter the new user information as
shown in the graphic. After entering the user information, click the
Create user button.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 95
RBAC and ZRBAC
c. Verify that the user account for Dante is created successfully and
shown below the user Guest in the Users tab.
3. Create a user account for user engAdmin in the engineering access zone.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 97
RBAC and ZRBAC
2. Create a user account for user Dante in the System access zone.
a. At the boston-1 prompt, run the isi auth users create Dante
--enabled yes --set-password --password-expires no
command.
Note: The text you type at the password and confirm prompt will not be visible.
b. Verify that the user account for Dante has been created by running
the isi auth users list command.Note that without using the
--provider option, the output shows users from all the configured
providers.
3. Create a user account for user engAdmin in the engineering access zone.
b. Verify the user is created by running the isi auth users list -
-zone engineering command.
c. View the user default setting by running the isi auth users view
engAdmin --zone engineering command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 99
RBAC and ZRBAC
b. In the View role details window, click on Edit role. In the Edit role
details window, click on Add a member to this role.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 101
RBAC and ZRBAC
a. Log out of the WebUI and login as Dante using the credentials
Dante/Dees12345!
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 103
RBAC and ZRBAC
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 105
RBAC and ZRBAC
a. To add to the ZoneAdmin role, at the boston-1 prompt, run the isi
auth roles modify ZoneAdmin --add-user engAdmin --
zone engineering command.
e. You will be unable to SSH into the cluster as engAdmin because the
ZoneAdmin and ZoneSecurityAdmin roles are not assigned the
ISI_PRIV_LOGIN_SSH privilege which is require for SSH acces.
On-disk Identity
Scenario:
The On-disk identity stores an authoritative version of the identity for users and
groups. This becomes important when accessing restored files replicated to a
remote system. The IT manager has asked you to use the OneFS CLI to identify
the UID, GID, SID, and the On-disk identity for user accounts.
Given in the lab exercise is the information required to identify identities. If needed,
use the CLI step-by-step walk through portion of the exercise for guidance.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 107
On-disk Identity
On-disk Identity
1. Use the following information to help identify the UID, GID, SID, and the on-
disk identity for user accounts. This step can only be performed using the
OneFS CLI.
a. Ensure that the global setting for the On Disk Identity is set as
native. Use the WebUI Access > Settings page or the isi auth
settings global command.
b. View the access token for the marketing zone user sera. To do this,
use the isi auth mapping token command.
c. View the access token for the marketing zone user sera by
referencing the domain as dees\\sera or [email protected]
Question: Is the SID the same for user sera when referenced as
sera and dees\\sera? Why or Why not?
Instructor Note: Yes, because the user sera account exists in both
Active Directory and LDAP and the account has a real SID as well as
a real UID.
d. View the access token for the sales zone user john.
Instructor Note: The On Disk identity is the SID. User john only
exists in Active Directory. If an object lacks a real UID or GID, then its
SID is set as the On Disk identity. OneFS only uses a SID for the on
disk identity when no configured authentication sources return a real
UID or GID.
e. View the mapping information for user with UID 10. To do this, use
the isi auth mapping view command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 109
On-disk Identity
2. Verify that the global setting for the On Disk Identity is set as native. To do
this, at the boston-1 prompt, run the isi auth settings global view
command and note the setting.
a. Change the setting, if needed, at the boston-1 prompt run the isi
auth settings global modify --on-disk-identity
native command.
3. View the access token for the marketing zone user sera. At the boston-1
prompt, run the isi auth mapping token sera --zone marketing
command.
a. Locate the UID and SID for user sera. Notice that the SID is an
algorithmic SID or 'UNIX SID', where the common S-1-22 is
appended with the UID to create a SID for an account that requires
one. Locate the GID and SID for the Primary Goup. Notice that the
SID is a UNIX SID.
b. Notice that the On Disk identity is the same as the UID. User sera is
an LDAP user. An object that has an explicit UNIX identity maintains
that identity as the On Disk identity.
4. At the boston-1 prompt, run the isi auth mapping token dees\\sera
--zone marketing command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 111
On-disk Identity
a. The On Disk identity setting is the UID. As the user account sera
exists in both Active Directory and LDAP, the account has a real SID
as well as a real UID.
5. View the access token for the sales zone user john. At the boston-1 prompt,
run the isi auth mapping token dees\\john --zone sales
command.
a. Note the On Disk identity for user john. The On Disk identity is the
SID. User john only exists in Active Directory. If an object lacks a real
UID or GID, then its SID is set as the On Disk identity. OneFS only
uses a SID for the on-disk identity when no configured authentication
sources return a real UID or GID.
6. To display the mapping information for a user with UID 10, at the boston-1
prompt, run the isi auth mapping view UID:10 command. You can
also view the mapping information for a user by using the SID or GID in the
command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 113
Permissions
Permissions
Scenario:
Before configuring the shares and exporting the file systems, the IT manager wants
you to become familiar with the way OneFS depicts and enforces permissions. For
this lab, view and modify the permissions on the cluster, and demonstrate how to
make a file or directory using advanced ACL.
Given in the lab exercise is the information needed to view, change and understand
file system permissions in OneFS. If needed, use the WebUI or the CLI step-by-
step walk through portion of the exercise for guidance.
Permissions
1. Use the following information to view, change the permissions and ownership
of the sales access zone base directory.
a. View the permissions and ownership of the sales base directory. You
can use the WebUI, File system explorer page or the ls command.
b. Change the owner of the sales base directory to the Active Directory
administrator. You can use the file system explorer of the WebUI or
the chown command.
c. Change the group owner of the sales base directory to the domain
administrators group. You can use the files system explorer of the
WebUI or the chgrp command.
d. Ensure that all individuals have read, write and execute access to the
sales base directory. You can use the files system explorer of the
WebUI or the chmod command.
2. Use the following information to change the authoritative permission type, add
and remove ACLs, for the sales base directory. This step can only be
performed only using the CLI.
a. Change the authoritative permission type for the sales base directory
to ACL. Use the chown command and its options.
b. Add an advanced ACL to the sales base directory for a user. Use the
chmod command and its options.
i. User: nobody
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 115
Permissions
c. Add an advanced ACL to the sales base directory for a group. Use
the chmod command and its options.
d. Remove the everyone ACL entry for the sales base directory. Use
the chmod command and its options.
3. Use the following information to understand the impact of changing the ACL
policy settings. This step can be performed only using the WebUI.
a. View the default ACL policy settings. Use the Access menu in the
WebUI.
b. Change the environmental variable with the different options and note
the changes to different settings.
Question: For each option, what is the impact of using chmod on files
that have existing ACLs?
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 117
Permissions
2. Examine the permissions and ownership of the sales access zone base
directory.
a. Navigate to the File system, File system explorer page. Set the
Current access zone to sales. On the sales directory, click View
details.
b. In the View directory property details window, you can see the
permissions and ownership set for the directory.
3. Change the permissions and ownership for the sales access zone base
directory.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 119
Permissions
d. The IT manager wants the permissions for the base directory writable
for all individuals. Ensure all the Permissions check boxes are
checked. Click Save changes, and then Close.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 121
Permissions
2. Examine the permissions and ownership of the sales access zone base
directory. At the boston-1 prompt, run the ls -led /ifs/sales command.
Note the permissions and ownership for the directory.
Note: The ls command lists the directory contents. The -l option is to list files
in the long format. The -e option prints the Access Control List (ACL). The -d
option lists on the directory and not its contents.
3. Change the permissions and ownership for the sales access zone base
directory.
a. Change the owner of the sales base directory to the Active Directory
administrator. At the boston-1 prompt, run the chown
[email protected] /ifs/sales command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 123
Permissions
2. View the authoritative permission type for the sales access zone base
directory. At the boston-1 prompt, run the ls -led /ifs/sales command.
Note that the permission type is set to POSIX as the ACL for the sales
directory is the Synthetic ACL.
Note: Although the owner of the directory was the Active Directory
administrator, the -s option uses an external authentication provider
and in this example, makes the directory ACL authoritative.
b. Run the ls -led /ifs/sales command and note the entry of the
added user.
Note: The nobody account is used for mapping root users over NFS to a different
UID so they are not treated as root on the cluster. This is known as root squash.
The generic_all permission represents Full Control permissions in Windows.
b. Run the ls -led /ifs/sales command and note the entry of the
added group.
6. Remove an ACL.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 125
Permissions
Note: With POSIX authoritative, changing the mode bits also changes
the synthetic ACLs.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 127
Permissions
Note: In the given scenario, you have created access zone sales that
authenticates using Active Directory, engineering that authenticates
using LDAP, and marketing that uses both authentication providers.
The ideal setting for such a setup is Balanced.
SMB Shares
Scenario:
The Marketing business unit has requested two Windows shares for their
environment. The first share is to accommodate home directories for all the
marketing users and the second share is for general purpose file sharing. Once the
shares are created, verify accessibility from an SMB client.
Given in the lab exercise is the information needed to configure SMB shares. If
needed, use the WebUI or the CLI step-by-step walk through portion of the
exercise for guidance.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 129
SMB Shares
SMB Shares
1. Create a share to accommodate home directories for all marketing users.
a. Enable SMB service. Use the WebUI, Windows sharing (SMB) page
or the isi services command.
Hint: Ensure to set the access zone before creating the share.
Hint: Use the --wellknown option when using the isi smb
shares permission command. In the WebUI, you may
change the permissions by clicking the View/Edit button.
c. Map the user home directory using the File Explorer on the
Jumpstation.
i. Folder: \\marketing\home
d. Verify that the home directory for user John is created. Use the File
system explorer page of the WebUI or the ls command.
a. Create a share. You can use the WebUI, Windows sharing (SMB)
page or the isi smb shares create command.
Hint: Ensure to set the access zone before creating the share.
b. Map the user home directory using the File Explorer on the
Jumpstation.
i. Folder: \\marketing\mktg-share
iii. Change the permissions for the mktg-share directory. You can
use the File system explorer page of the WebUI or the
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 131
SMB Shares
3. Navigate to the Protocols, Windows sharing (SMB) page. Select the SMB
shares tab. Set the Current access zone to marketing. Click on Create an
SMB share.
4. In the Create an SMB share window, enter the fields as shown and click
Create share.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 133
SMB Shares
Note: When joining the Active Directory domain, the Domain Users group is
automatically added to the cluster's user group by default. This can be seen by
selecting Add member and choosing an AD user.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 135
SMB Shares
6. Go back to the WebUI and navigate to the File system, File system explorer
page. Set the Current access zone to marketing. Click on home. Verify that
the john home directory has been created.
a. Minimize mRemoteNG.
b. Open the File Explorer on the Jumpstation by clicking the icon on the
task bar. Right-click on This PC and click on Map network drive.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 137
SMB Shares
5. Verify that the john home directory has been created. Go back to
mRemoteNG, at the boston-1 prompt, run ls /ifs/marketing/home and
ensure that john is listed.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 139
SMB Shares
b. In the Create an SMB share window, enter the fields as shown and
click Create share.
3. Minimize the WebUI. Go to the File Explorer on the Jumpstation to map the
share. Map the mktg-share share. Refer to the previous task if required.
a. In the mapped mktg-share, right click in the right panel, select New
and then select Folder. Why do you get an access denied message?
This is because you do not have the needed permissions.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 141
SMB Shares
Note: When a share is created with the ACL policy settings set to
Balanced and the Directory ACLs setting set to Apply Windows
default ACLs, the share is ACL authoritative.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 143
SMB Shares
2. Create a general share for all the marketing users. At the boston-1 prompt,
run the isi smb shares create mktg-share --zone=marketing --
path=/ifs/marketing/mktg-share --create-path command.
a. In the mapped mktg-share, right click in the right panel, select New
and then select Folder. Why do you get an access denied message?
This is because you do not have the needed permissions.
b. At the boston-1 prompt, run the isi smb share view mktg-
share --zone marketing | grep -ie Permission -e
Everyone command. Notice that the permission is set as allow read
for Everyone.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 145
SMB Shares
Note: When a share is created with the ACL policy settings set to
Balanced and the Directory ACLs setting set to Apply Windows
default ACLs, the share is ACL authoritative.
NFS Exports
Scenario:
Now that you have configured the Windows shares, export the directories to the
Linux users. The IT manager has asked that you create, configure, export, and
mount a general purpose directory for use by the marketing members that use
Linux clients. Although the dominant access to the files is using Linux, some
Windows users may also access the files. The IT manager wants you to
understand root squash and the permission behavior.
Given in the lab exercise is the information needed to configure NFS exports. If
needed, use the WebUI or the CLI step-by-step walk through portion of the
exercise for guidance.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 147
NFS Exports
NFS Exports
1. Create a general purpose NFS export for the marketing members using Linux
clients.
b. Enable NFS service. You can use the UNIX sharing (NFS) page of
the WebUI or the isi services command.
Instructor Note: Any client not restricted by rules can access the
export.
d. Create an NFS alias for the export in the marketing access zone.
You can use the NFS aliases tab of the UNIX sharing (NFS) page in
the WebUI or the isi nfs aliases command.
2. Mount the NFS directory exported from the cluster to the Centos machine.
This step can be performed only with the CLI.
Hint: If the DNS and alias are not configured, use the IP
address for the marketing access zone and the directory path
for the export.
c. Create a empty file named test using the touch command. Verify the
owner and group for the test file using the ls command.
Question: Why is the user and group not root on the file test that
user root on the Centos client created?
d. On the cluster, revoke the Write permission for Others in the mktg-
export directory and note that the owner for the directory is user root.
You can use the File system explorer page of the WebUI or the
chmod and ls commands.
e. In the Centos client, try to create another empty file named test1.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 149
NFS Exports
Instructor Note: Because root on the cluster and root on the client
are not the same user.
f. Add the Centos root (192.168.3.3) to the Root clients rule of the NFS
export for the mktg-export directory. You can use the UNIX sharing
(NFS) page of the WebUI or the isi nfs exports command.
g. Try to create file test1 again on the Centos client. Verify the owner
and group for file test1.
a. Navigate to the File System, File system explorer page. Set the
Current access zone field to marketing. Click on Create directory.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 151
NFS Exports
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 153
NFS Exports
5. Create an NFS alias for the export so the user will not have access to the full
directory path of their export. An alias creates a shortcut for the directory path
name.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 155
NFS Exports
2. Create the mktg-export directory. At the boston-1 prompt, run the following
commands:
4. Create the export for the mktg-export directory. At the boston-1 prompt, run
the following commands:
Replace the <ID> field in the command with the export ID.
5. Create an NFS alias for the export so the user will not have access to the full
directory path of their export. An alias creates a shortcut for the directory path
name. At the boston-1 prompt, run the following commands:
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 157
NFS Exports
Note: If DNS were not configured, you can use an IP address that is
in the marketing IP address pool configured on the cluster. Without
using DNS and without using an NFS alias, the command may look
like mount 192.168.3.33:/ifs/marketing/mktg-export
/mnt/gen-purpose
3. On the Centos SSH connection, run the whoami command to verify you are
user root.
4. Revoke the Write permission for Others. You may perform this step either
using the WebUI or CLI.
ii. Select Edit properties. Note that the user is root. Uncheck the
write permission for Others. Click on Save changes and
Close.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 159
NFS Exports
6. The root user of the Centos client is a trusted root for the cluster. Change the
root squash for the client. You may perform this step either using the WebUI or
CLI.
ii. Click Edit export. In the Edit NFS export details window,
enter 192.168.3.3 in the Root clients field. This is the Centos
client where user root is trusted. Click Save changes and then
Close.
ii. Add root client: Run the isi nfs exports modify <ID>
--zone marketing --root-clients 192.168.3.3
command. Replace the <ID> field with the export ID.
iii. Verify: Run the isi nfs exports view <ID> --zone
marketing command. Replace the <ID> field with the export
ID. Notice that 192.168.3.3 has been added as a root client.
This is the Centos client where user root is trusted.
Note: The 192.168.3.3 client is configured under the Root clients rule as well as
the Always read/write clients rule. The Root clients rule takes precedence.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 161
NFS Exports
the test1 file. Alternatively, you may navigate to the File system, File
system explore page of the WebUI, marketing access zone, mktg-
export directory, and view the details on the test1 file.
S3
Scenario:
The IT manager wants the Marketing users to access data from an S3 application.
Demonstrate the S3 configuration steps to create a bucket and access the bucket
using other protocols.
Given in the lab exercise is the information needed to configure S3. If needed, use
the WebUI or the CLI step-by-step walk through portion of the exercise for
guidance.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 163
S3
S3 Buckets
1. Configure S3 for the marketing access zone and verify multi-protocol access.
a. Enable the S3 service. Use the Object storage (S3) page of the
WebUI or the isi services command.
b. Set the base domain for the marketing access zone settings. Use the
Object storage (S3) page of the WebUI or the isi s3 settings
command.
d. Create a secret key for the Active Directory user John. Note the
Access id and the Secret Key. Use the Object storage (S3) > Key
management page of the WebUI or the isi s3 keys command.
e. Create a new account in the S3 Browser. Open the browser from the
desktop or taskbar. Use the Access id and the Secret Key obtained
from the previous task.
g. Create a new SMB share in the marketing access zone. Use the
Windows Sharing (SMB) page of the WebUI or the isi smb
shares command.
i. Name: mktg-bucket
h. Map the mktg-bucket share using Windows Explorer. Verify that you
are able to access the Animation folder.
i. Folder: \\marketing\mktg-bucket
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 165
S3
b. Configure the zone settings. Select the Zone settings tab and
change the current access zone to marketing. Enter the base
domain and as shown and click Save changes.
c. To create a bucket, select the Buckets tab and set the current
access zone to marketing. Click Create Bucket and enter the fields
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 167
S3
d. Once done, you will find the bucket listed in the Buckets table.
a. Create a secret key for the Active Directory user John. On the Object
storage (S3) page, select the Key management tab. Set the current
access zone to marketing. Select user John and click Create new
key. A new key secret key will be listed in the Secret key Details
table. Note the Access id and the Secret key for user John.
b. Open the S3 Browser using the taskbar or the desktop. Enter the
details as shown in the graphic to create a new account. You need to
copy the Access id and the Secret Key from the WebUI and paste
them in the new account form. Once done, Add new account.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 169
S3
c. The mktg-bucket will be listed on the left. To add contents, select the
bucket and then click Upload > Upload Folder. Choose the
Animation directory from the File-Repository directory and click
Select Folder. Once uploaded, the folder will appear on the right.
d. Now, try to access the bucket using SMB. Create a SMB share in the
marketing access zone as shown. Refer to the SMB Shares
exercise of the lab guide for the detailed procedure.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 171
S3
e. Map the mktg-bucket share on the Jumpstation. Verify that you are
able to access the Animation folder.
c. Configure the zone settings. At the boston-1 prompt, run the isi s3
settings zone modify --base-domain
marketing.dees.lab --zone marketing command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 173
S3
a. Create a secret key for the Active Directory user John. At the
boston-1 prompt, run the isi s3 keys create DEES\\john --
zone marketing command. Note the Access id and the Secret
key for user John.
b. Open the S3 Browser using the taskbar or the desktop. Enter the
details as shown in the graphic to create a new account. You need to
copy the Access ID and the Secret Key from the CLI and paste them
in the new account form. Once done, Add new account.
c. The mktg-bucket will be listed on the left. To add contents, select the
bucket and then click Upload > Upload Folder. Choose the
Animation directory from the File-Repository directory and click
Select Folder. Once uploaded, the folder will appear on the right.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 175
S3
d. Now, try to access the bucket using SMB. At the boston-1 prompt,
run the isi smb shares create mktg-bucket --
zone=marketing --path=/ifs/marketing/mktg-bucket -
-i=yes command. Change the permissions by running the isi smb
shares permission modify mktg-bucket --
zone=marketing --wellknown Everyone --permission-
type=allow --permission=full command.
f. Map the mktg-bucket share on the Jumpstation. Verify that you are
able to access the Animation folder.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 177
HTTP - Optional
HTTP - Optional
Scenario:
The Marketing business unit has requested the use of HTTP for their clients. The IT
manager has asked you to configure HTTP on the cluster for the Marketing
business unit and test the connection.
Given in the lab exercise is the information needed to configure HTTP. If needed,
use the WebUI or the CLI step-by-step walk through portion of the exercise for
guidance. This is an optional lab exercise for the course. You may choose to either
perform or skip this exercise.
HTTP
1. Use the following information to configure HTTP for the cluster.
a. Create the HTTP root directory in the marketing access zone. You
can use the File system explorer page of the WebUI or the mkdir
command.
b. Configure HTTP for the cluster. You can use the HTTP settings page
of the WebUI or the isi http settings command.
iii. Give full permission to the copied log file in the HTTP base
directory. Use the chmod command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 179
HTTP - Optional
a. Navigate to the File System, File system explorer page. Set the
Current access zone field to marketing. Click on Create directory.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 181
HTTP - Optional
2. Create HTTP root directory. At the boston-1 prompt, run the following
commands:
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 183
HTTP - Optional
2. Add contents to the HTTP root directory. At the boston-1 prompt, perform the
following steps:
c. Give full permissions to the log file by running the chmod 777
/ifs/marketing/streaming/main_config_changes.log
command. You will get an access denied message if the permissions
are not set.
c. To verify access, on the Index page, click on the log file to ensure the
file opens.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 185
HTTP - Optional
Scenario:
As part of the documentation process, the IT manager has asked you to analyze
how OneFS uses data protection levels to protect data, and then lead a small
workshop for the team to ensure everyone understands OneFS data protection.
The IT manager has created exercises to identify the protection levels of the file
layouts for various protection strips, as well as determine the overhead for cluster
sizing. You will review and make changes to the data protection settings for
directory and file protection levels that can affect node pool, folder, and file.
Given in the lab exercise is the information needed to configure Data Protection. If
needed, use the WebUI or the CLI step-by-step walk through portion of the
exercise for guidance.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 187
Data Protection and File Layout
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 189
Data Protection and File Layout
Hint: Ensure to get the node pool ID if you are performing this step
using the CLI.
Question: Note the Storage Pools > Summary tab has an Info
message stating vmware_100gb_6gb requested protection is
different than the suggested protection. What does this mean?
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 191
Data Protection and File Layout
b. Change the protection settings for the mktg-DP directory. Use the
File system explorer page of the WebUI or the isi set command.
c. Create an SMB share for the mktg-DP directory. Use the Windows
sharing (SMB) page of the WebUI or the isi smb shares
command.
Note: Although the file policy is set at +2d:1n protection, the explicit
directory permission on the mktg-DP directory takes precedence.
Instructor Note: 5 data stripe units and 1 FEC stripe unit spread
across 1 drive per node, which means N+1n protection.
Note: Although the file policy is set at +2d:1n protection, and the
explicit directory permission on the mktg-DP directory is set at +1n,
the explicit protection policy at the file level takes precedence.
Instructor Note: 15 data stripe units and 3 FEC stripe units spread
across 3 drive per node.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 193
Data Protection and File Layout
2. Navigate to the File System, Storage pools page, Summary tab. The
Summary tab displays storage pool status and local storage usage statistics.
The SmartPools info message indicates that the requested protection is
different than the protection suggested by OneFS. The Info message is not an
error.
3. Configure the +1n protection level for the vmware_100gb_6gb node pool.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 195
Data Protection and File Layout
4. Change the protection level for the default file pool policy. Navigate to the File
Pool Policies tab and click View details for Default policy. Click on Edit
policy. Change the Requested protection to +2d:1n and click Save
changes.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 197
Data Protection and File Layout
2. Configure the +1n protection level for the vmware_100gb_6gb node pool.
b. To change the Protection Policy, you need the ID of the node pool
that you want to change. Run the isi storagepool nodepools
modify <node pool ID> -p +1n command. The -p option tells
you which protection level you are going to use. Replace the <node
pool ID> in the command with the ID for the vmware_100gb_6gb
node pool.
3. Change the protection level for the default file pool policy to +2d:1n.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 199
Data Protection and File Layout
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 201
Data Protection and File Layout
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 203
Data Protection and File Layout
b. Map the mktg-DP share and then copy the Animation folder from the
File-Repository directory to the mapped share.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 205
Data Protection and File Layout
b. View the file properties once again to verify the changes and close
the window.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 207
Data Protection and File Layout
c. Although the file policy is set at +2d:1n protection, and the explicit
directory permission on the mktg-DP directory is set at +1n, the
explicit protection policy at the file level takes precedence.
2. Configure the +1n protection level for the mktg-DP directory. At the boston-1
prompt, run the following commands:
Note: Replace the <node pool ID> in the command with the ID for
the vmware_100gb_6gb node pool. Run the isi storagepool
nodepools list command to get the ID.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 209
Data Protection and File Layout
b. Map the mktg-DP share and then copy the Animation folder from the
File-Repository directory to the mapped share.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 211
Data Protection and File Layout
c. Although the file policy is set at +2d:1n protection, and the explicit
directory permission on the mktg-DP directory is set at +1n, the
explicit protection policy at the file level takes precedence.
SmartPools
Scenario:
The IT manager has asked you to tier cluster data on two different types of storage.
Use SmartPools to create node pools and tiers. Use file pools policies to specify
target files to move and store on a different tier with different protection settings.
The protection levels you choose are unique to your workflows, your clusters, and
your organization requirements. What is an acceptable failure rate for one
organization may be unacceptable to another.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 213
SmartPools
SmartPools
1. Configure SmartPools.
a. Activate SmartPools license. For the lab environment, you will use
the trial license functionality. Use the Licensing page of the WebUI
or the isi license command.
i. Name: media_pool
d. Build a tier to encompass the newly created node pool. Use the
Storage pools page of the WebUI or the isi storagepool
tiers command.
i. Name: media_tier
2. Navigate to the File System, Storage pools page, Summary tab. The
Summary tab displays storage pool status and local storage usage statistics.
Note that in the Status section, under the Module column, you will notice that
Policies indicates a Status of Caution because SmartPools is not yet
licensed. The SmartPools and SmartPool settings module should both show
a status of Info or Good. The CloudPools module should show a status of
Caution because CloudPools is also not yet licensed.
3. Activate SmartPools license. For the lab environment, you will use the trial
license functionality.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 215
SmartPools
b. Check SmartPools and click on Start trial. Once done, close the
window.
4. Create a manual node pool named media_pool by splitting the existing six-
node, node pool into two three-node, node pools. Note that you can only
perform this operation using the OneFS CLI.
d. The blue "X" in the actions column is an indication the newly created
media_pool is manually created and can be deleted. The gray "X" on
the vmware_100gb_6gb indicates that it is automatically created by
OneFS.
a. In the SmartPools tab, click the Pool: media_pool link. Set the
requested protection for the node pool to +3d:1n, click Save
changes.
6. Create a tier to encompass the newly created node pool. You will configure a
file pool policy shortly to use the tier.
a. Ensure you are still on the Storage pools page, SmartPools tab. To
create a tier, click the Create a tier button and name the tier as
media_tier. Add the media_pool to the tier. Click Create tier.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 217
SmartPools
2. Activate SmartPools license. For the lab environment, you will use the trial
license functionality.
3. Create a manual node pool named media_pool by splitting the existing six-
node, node pool into two three-node, node pools. Note that this operation can
only be perform using the OneFS CLI.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 219
SmartPools
5. Create a tier to encompass the newly created node pool. You will configure a
file pool policy shortly to use the tier.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 221
File Pool Policies
Scenario:
The IT manager has asked you to use file pools policies to specify targets files to
move and store on a different tier with different protection settings.
Given in the lab exercise is the information needed to configure File Pool Policies.
If needed, use the WebUI or the CLI step-by-step walk through portion of the
exercise for guidance.
a. Review the cluster default file pool policy. Use the Storage pools
page of the WebUI or the isi filepool command.
Instructor Note: The default policy will be applied to all the data on
the cluster. The only exception is data for which there is some manual
configuration specified.
b. For this scenario, ensure that files only go to the Default node pool
(vmware_100gb_6gb). Change the default file policy in the Storage
pools page of the WebUI or use the isi filepool default-
policy command.
2. Create a file pool policy. A file pool policy consists of two components, Filters
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 223
File Pool Policies
and Actions. The filters are used to select the files and the actions modify the
setting of the files selected by the filter.
Instructor Note: Initially when data is written, the files may reside in
other node pools as well. In time, the data will eventually be moved to
the target node pool configured in the default policy. This operation
can also be manually triggered by running the SmartPoolsTree job.
d. Test the effect the media_policy has on the files. Run the
SmartPools Tree job. Note that the job may take a few minutes to
complete. Use the Job operations page of the WebUI or the isi
job jobs command.
i. Path: /ifs
e. Verify that the policy and disk pool is media_pool for all the .mp4
files. Use the isi get command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 225
File Pool Policies
a. Navigate to the File system, Storage pools page. Click the File pool
policies tab. Only the Default policy is displayed. At this stage the
Default policy applies to all data written to the cluster for which some
manual configuration has not been specified.
Note: You will not see the Create a file pool policy button if
SmartPools is not licensed. Without the license, you cannot create
file pool policies.
b. In the File pool policies table, in the Action column, click View
Details for the Default policy. In the View default policy details
window, review the different settings and close the window.
3. For this scenario, ensure that files only go to the default node pool
(vmware_100gb_6gb). With the Storage target and Snapshot storage
target set to anywhere, files will be placed across all node pools.
a. Click the View/Edit button for the Default policy. Click Edit policy.
Make the changes as shown and then click Save changes and
Close.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 227
File Pool Policies
a. At this stage, only the default policy file pool policy is configured. The
default policy applies to all data written to the cluster for which some
manual configuration has not been specified.
Note: You will not be able to create a new file pool policy if
SmartPools is not licensed.
3. For this scenario, ensure that files only go to the default node pool
(vmware_100gb_6gb). With the Storage target and Snapshot storage
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 229
File Pool Policies
target set to anywhere, files will be placed across all node pools.
2. Open the Windows File Explorer on the Jumpstation and map the mktg-
share, if it is not mapped. Copy the mydocuments folder from the File-
Repository directory into the mapped share mktg-share. If needed, you can
review the SMB Shares lab for the steps to map a share.
3. Review the location and protection level of data residing in the mydocuments
folder.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 231
File Pool Policies
d. If you do not change the target node pools for the Default policy, the
files can be in either of the two node pools because the Default
policy writes to anywhere, and will randomly balance which node
pool a file is written. Running the isi get -D
/ifs/marketing/mktg-share/mydocuments/* | grep
"Disk pools" command shows the distribution of the files across
the two pools.
4. Create a file pool policy to change the protection and location of .mp4 files
present in the mydocuments directory. You may perform this step either
using the WebUI or the CLI.
ii. Verify that the media_policy is listed in the File pool policies
tab.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 233
File Pool Policies
5. Test the effect the media_policy has on the files. Run the SmartPools Tree
job. Note that the job may take a few minutes to complete. You may perform
this step either using the WebUI or the CLI.
ii. On the Job summary tab, monitor the job. The indication that
the job is completed is when the job is no longer present in the
Active jobs table.
ii. Run the isi job jobs list command. This displays a list
of all the active jobs. The indication that the job is completed is
when the job is no longer present in the list. Run the command
once every minute to find out if it has been completed.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 235
File Pool Policies
6. When the job finishes, go to the SSH session and run the isi get -D
/ifs/marketing/mktg-share/mydocuments/*.mp4 | grep "Disk
pools" command. Verify that the policy and disk pool is media_pool for all
the .mp4 files.
File Filtering
Scenario:
A problem with the last storage system was the large amount of capacity
consumption due to individuals storing personal pictures on the shares. The IT
manager wants to ensure that users cannot use storage for specific files types
within specific shares. Use the file filtering feature to prevent certain files from
consuming storage in a share.
Given in the lab exercise is the information needed to configure file filtering. If
needed, use the WebUI or the CLI step-by-step walk through portion of the
exercise for guidance.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 237
File Filtering
File Filtering
1. Use the information given below to add and verify a file filter on a SMB share.
a. Add a file filter to prevent users from writing JPEG image files on the
marketing general purpose share (mktg-share). Edit the share using
the Window sharing (SMB) page of the WebUI or the isi smb
shares command.
Instructor Note: As some files in the the M & E folder are .jpg files,
the filter added on the share prevents them from being copied.
ii. Compare the File-Repository > M & E folder with the mktg-
share > M & E folder.
2. Add a file filter to prevent users from writing JPEG image files on the
marketing general purpose share.
b. In the Edit SMB share details window, go to the File filter section
and check Enable file filters. Add the file extension .jpg as shown
and then save the changes.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 239
File Filtering
c. Compare the File-Repository > M & E folder with the mktg-share >
M & E folder. Notice that the JPEG files did not copy over from the
Sat Images folders, but other files such as the Thumbs.db are
copied successfully.
4. Similarly, you can add file filters to an access zone in the Access, File filter
page of the WebUI. The file filter on an SMB shares takes precedence over
the filter filter on the access zone where the share resides.
2. Add a file filter to prevent users from writing JPEG image files on the
marketing general purpose share.
a. At the boston-1 prompt, run the isi smb shares modify mktg-
share --zone=marketing --file-filtering-enabled=yes
--file-filter-extensions=.jpg --file-filter-
type=deny command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 241
File Filtering
c. Compare the File-Repository > M & E folder with the mktg-share >
M & E folder. Notice that the JPEG files did not copy over from the
Sat Images folders, but other files such as the Thumbs.db are
copied successfully.
4. Similarly, you can add file filters to an access zone using the isi file-
filter settings modify command. The file filter on an SMB shares
takes precedence over the filter filter on the access zone where the share
resides.
SmartQuotas
Scenario:
A portion of the storage capacity of the sales general purpose directory will be
allocated to each user of the directory. The IT manager has asked you to use
quotas to enforce storage limits. You will start by setting up report generation and
event notifications. Next, create a directory type quota and apply user quotas to the
general purpose directory. Ensure to test the settings.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 243
SmartQuotas
SmartQuotas
1. Use the following information to change the global quota reporting frequency
and notification rules.
a. Activate SmartQuotas license. For the lab environment, you will use
the trial license functionality.
c. Add the global notification rules for quotas. Use the SmartQuotas
page of the WebUI or the isi quota settings notifications
command.
2. Use the following information to configure directory type quotas and default-
user quotas on the general purpose sales directory.
a. Create a general purpose share for the sales access zone. Use the
Windows sharing (SMB) page of the WebUI or the isi smb
shares command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 245
SmartQuotas
Instructor Note: No, because the folder size is larger than the hard
limit enforced on the user.
Instructor Note: The quotas are configured with the global reporting
settings which schedule reports every 2 minutes of everyday.
iii. View the list of all quotas. Use the SmartQuotas page of the
WebUI or the isi quota quotas command.
i. Change the hard limit enforced on user John to 200 MB. Use
the SmartQuotas page of the WebUI or the isi quota
quotas command.
ii. Using Windows Explorer, copy the File-Repository > M & E >
Sat Images folder to the mapped sales-gen share.
Instructor Note: Yes, because the folder size is smaller than the hard
limit enforced on user John.
Question: Is the hard limit updated for user John? Why or Why
not?
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 247
SmartQuotas
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 249
SmartQuotas
c. Leave the Manual reporting section and the Email mapping section
with its default settings.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 251
SmartQuotas
5. Once done, scroll to the bottom of the page and click Save changes.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 253
SmartQuotas
3. Change the frequency of the quota reports. At the boston-1 prompt, run the
following commands:
4. Add the global notification rules for quotas. At the boston-1 prompt, run the
following commands:
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 255
SmartQuotas
i. Enter the fields as shown. This is the share you will use for
auditing events. Once all the fields are entered click Create
Share.
b. Select the quota type as Directory quota and fill the other fields as
shown. Once done, click Create quota. A success message is
displayed on the Quota and usage tab.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 257
SmartQuotas
c. In the Quotas and usage table, click View/Edit for the directory and
review all of your settings. If you require a custom notification for this
quota, you can change the configuration in the Quotas & Usage
table.
b. Select the quota type as User quota and fill the other fields as
shown. Once done, click Create quota. A success message is
displayed on the Quota and usage tab.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 259
SmartQuotas
c. Verify that the default-user has been added in the Quotas and usage
table.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 261
SmartQuotas
c. Navigate to the Quotas and Usage tab and note all the quotas listed.
Observe that a quota for user john is listed. The default-user quota
automatically creates a quota for user john.
a. Edit the quota for user john. Navigate to the File System,
SmartQuotas, Quotas and usage tab. Find the user: Dees\john
linked entry. In order to edit this quota, you need to first remove its
link with the default-user quota. For the entry, click the Unlink
button.
b. Click the View/Edit button and then click Edit. Change the Hard limit
to 200 MB and click Save changes and close the window. You
should see a success message.
d. Next, update the quota for all users to 250 MB. On the Quotas and
usage tab, click on View/Edit for the default-user quota entry.
Update the hard limit to 250 MB and save the changes.
e. Notice the quotas on the Quotas and usage tab. The hard limit for
user john has not been updated because user john is not linked to the
default-user quota and has a specific quota setting. Any specific user
quotas that you configure take precedence over a default user quota.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 263
SmartQuotas
a. Create Share: At the boston-1 prompt, run the isi smb shares
create sales-gen --zone=sales --
path=/ifs/sales/sales-gen --i=yes --create-path
command.
3. Go to Windows Explorer and map the \\sales\sales-gen share and then click
Finish. Refer to the SMB Shares lab if you need the detailed steps to map a
Network Drive.
a. To create the quota, at the boston-1 prompt, run the isi quota
quotas create /ifs/sales/sales-gen directory --
hard-threshold 400M --soft-threshold 300M --soft-
grace 30m --container yes command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 265
SmartQuotas
a. To create the quota, at the boston-1 prompt, run the isi quota
quotas create /ifs/sales/sales-gen default-user --
hard-threshold 10M command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 267
SmartQuotas
c. View all the quotas enforced by running the isi quota quotas
list command. Observe that a quota for user john is listed. The
default-user quota automatically creates a quota for user john.
a. Edit the quota for user john. In order to edit this quota, you need to
first remove its link with the default-user quota. At the boston-1
prompt, run the isi quota quotas modify
/ifs/sales/sales-gen user --user "DEES\john" --
linked no command.
b. Change the hard limit to 200 MB by running the isi quota quotas
modify /ifs/sales/sales-gen user --user "DEES\john"
--hard-threshold 200M command.
e. Next, update the quota for all users to 250 MB. Run the isi quota
quotas modify /ifs/sales/sales-gen default-user --
hard-threshold 250M command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 269
SmartDedupe
SmartDedupe
Scenario:
Another problem with the last storage system was the large amount of storage
consumed by duplicate data. The IT manager wants you to run a deduplication
assessment to verify you can assess space savings, specify deduplication settings,
and view space savings. Also, view the deduplication report.
SmartDedupe
1. Use the following information to configure SmartDedupe and deduplicate the
marketing general purpose share.
a. Activate the SmartDedupe license. For the lab environment, you will
use the trial license functionality.
ii. View the report and note the space savings. Use the
Deduplication page of the WebUI or the isi dedupe
reports command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 271
SmartDedupe
i. Start the Dedupe job with its default values. Use the Job
operations page of the WebUI or the isi job jobs
command.
ii. Check the deduplication stats and note the amount of disk
space saved. Use the Deduplication page of the WebUI or the
isi dedupe command.
iii. View the report and note the space savings. Use the
Deduplication page of the WebUI or the isi dedupe
reports command.
Instructor Note: Identical files are deduplicated only when the files or
identical 8K blocks are stored in the same node pool and have the
same protection policy.
a. Map the mktg-share directory, if not already mapped. Ensure that the
Animation, M & E, and mydocuments directories are present as
shown. If not, copy the directories from the File-Repository directory
using Windows Explorer.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 273
SmartDedupe
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 275
SmartDedupe
b. On the Start a job dialog box, click Start job. A Job Start
Succeeded message displays briefly.
b. On the Start a job dialog box, click Start job. A Job Start
Succeeded message displays briefly.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 277
SmartDedupe
c. Check the status of the job in the Summary tab. Once the job
completes, it will be removed from the Active jobs table. This may
take a few minutes to complete.
Note: Deduplication is not efficient for all file types. For example,
media, PDFs, and zip files are not good candidates for deduplication
unless the files are heavily replicated in the same node pool. Typical
user files such as Microsoft Office files that are shared, modified, and
copied among users are good candidates for deduplication. A good
use case for deduplication is user home directories where hundreds
of users may typically keep multiple versions of the same files.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 279
SmartDedupe
a. Map the mktg-share directory, if not already mapped. Ensure that the
Animation, M & E, and mydocuments directories are present as
shown. If not, copy the directories from the File-Repository directory
using Windows Explorer.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 281
SmartDedupe
a. To run the assessment, at the boston-1 prompt, run the isi job
jobs start DedupeAssessment command.
b. To check the job status, run the isi job jobs list command to
display the active jobs list. Once the job completes, it will no longer be
present in the list. Run the command every few seconds to check if
the job is finished. This may take a few minutes to complete.
c. Once the job finishes, view the report generated. We need the job ID
to view the report. Run the isi dedupe reports list command
to get the job ID. Replace the ID in the isi dedupe reports
view <ID> command to view the report. Note that it may take as
much as five minutes for the report to display in the list.
b. Check the job status by running the isi job jobs list
command. This may take a few minutes to complete.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 283
SmartDedupe
c. Once the job completes, view the amount of disk space saved by
running the isi dedupe stats command.
d. View the report generated. Run the isi dedupe reports list
command to get the job ID. Replace the ID in the isi dedupe
reports view <ID> command to view the report.
Note: Deduplication is not efficient for all file types. For example,
media, PDFs, and zip files are not good candidates for deduplication
unless the files are heavily replicated in the same node pool. Typical
user files such as Microsoft Office files that are shared, modified, and
copied among users are good candidates for deduplication. A good
use case for deduplication is user home directories where hundreds
of users may typically keep multiple versions of the same files.
SnapshotIQ
Scenario:
The IT manager wants to ensure you can recover files from snapshots. Use
SnapshotIQ to access specific point in time copies of files and restore the data from
the snapshots.
Many of the steps need to be done in quick succession in order to see the
snapshot behavior. For this reason, it is suggested that you read through all the
steps for tasks 1 and 2 before you perform the steps.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 285
SnapshotIQ
SnapshotIQ
1. Create Snapshot Schedule.
d. View the snapshots created. Use the SnapshotIQ page of the WebUI
or the isi snapshot snapshots command. In accordance with
the configured schedule, it could take up to 2-3 minutes for snapshots
to appear.
2. Create Snapshots. Perform this step using the OneFS CLI and Windows
Explorer.
a. Create multiple point-in-time versions for a file using CLI. Use the
date > filename command.
iv. After 2-3 minues, create snapfile once again with the current
timestamp.
b. View the amount of space that the snapshot uses after data is
modified. Use the SnapshotIQ page of the WebUI or the isi
snapshot snapshots command.
3. Restore files from Snapshots. This step can be performed only using the
OneFS CLI and Windows Explorer.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 287
SnapshotIQ
iii. Use the Open, Copy and Restore operations to restore earlier
versions of snapfile.
a. Map the sales-gen share, if not already mapped. Open the mapped
sales-gen share. Delete all contents in the directory, if any.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 289
SnapshotIQ
Tip: When creating a snapshot policy, keep in mind the length of time that you want
to set for expiration of the snapshots. If you are taking a snapshot every two
minutes as you do in the lab, you may not want to keep each snapshot for more
than two hours. If you use a setting that takes a snapshot every few days, then you
may want to set the expiration time so that you have at least one saved snapshot
when the next snapshot is taken. If you manually delete the policy, the snapshots
are not deleted because a service running in the background monitors the
snapshots and does not allow them to be deleted or removed until the expiration
time that you set in the policy.
5. Go to the Snapshots tab. You should begin to see snapshots. The snapshots
do not consume any space until data has been changed or deleted. In
accordance with the configured schedule, it could take up to 2-3 minutes for
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 291
SnapshotIQ
snapshots to appear.
Tip: Click on another tab and then back to the Snapshots tab to refresh the
Snapshots table.
2. Activate the trail SnapshotIQ license. At the boston-1 prompt run the following
commands:
a. Map the sales-gen share, if not already mapped. Open the mapped
sales-gen share. Delete all contents in the directory, if any.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 293
SnapshotIQ
Tip: When creating a snapshot policy, keep in mind the length of time that you want
to set for expiration of the snapshots. If you are taking a snapshot every two
minutes as you do in the lab, you may not want to keep each snapshot for more
than two hours. If you use a setting that takes a snapshot every few days, then you
may want to set the expiration time so that you have at least one saved snapshot
when the next snapshot is taken. If you manually delete the policy, the snapshots
are not deleted because a service running in the background monitors the
snapshots and does not allow them to be deleted or removed until the expiration
time that you set in the policy.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 295
SnapshotIQ
b. Create a file named snapfile which contains the current time. At the
boston-1 prompt, run the date > snapfile command. This
command writes the current timestamp to the file.
c. Navigate to the share folder that was mapped and verify that snapfile
is created. Open the file with Notepad and view its contents.
d. Again, in the CLI, wait for 2 minutes and run the date >>
snapfile command to append the latest timestamp to the end of the
existing file.
e. In the share, open the file and view its contents after the edits. For
demonstration purposes, append the file 4-5 times. Note the time
each time you append to the file.
g. After 2-3 minutes, run the date > snapfile to recreate the file.
The contents will only include the current time at which you created
the file.
3. View the amount of space that the snapshot uses after data is modified. You
may perform this step either using the WebUI or the CLI.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 297
SnapshotIQ
b. At the boston-1 prompt, run the ls -al command. Notice that the
.snapshot directory is not seen. It is a hidden directory but can be
accessed.
d. Run the pwd command and notice that the present working directory
is .snapshot directory.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 299
SnapshotIQ
f. You should see a subdirectory here for each snapshot which has
been taken (one every two minutes). In accordance with the snapshot
schedule you created earlier in this lab, the directory names should
be MySales_<timestamp>. Take note of the timestamps in each
directory name. The contents of these snapshot directories are in a
read only state. Earlier versions of files can be copied from the
snapshot directories, but files in the directories are in a read only
state. They cannot be edited or modified.
h. Copy the earlier version of the file from the snapshot into the present
directory. Rename the file to old-snapfile by running the cp
snapfile /ifs/sales/sales-gen/MySales/old-snapfile
command.
i. Go back to the share and verify if the file is recovered. Open the file
and review the contents of the file.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 301
SnapshotIQ
c. Select one of the earlier versions. For the demonstration, the snapfile
from time 07:06 is selected. Click on Open. The earlier version of the
file will be displayed. Review the date information placed into this
point in time view of the file and close the file.
d. To copy the file, select the file and click Copy. For the
demonstration, the snapfile from time 07:06 is selected. Select your
desktop as the save location.
e. Navigate to the desktop and verify that the earlier version of the file is
present.
f. To restore the original file, open the Previous Versions tab in the
Properties window. Select the version of the file to restore. For the
demonstration, the snapfile from time 07:08 is selected. Click
Restore. Click Restore again to restore the file. Click OK in the
success window. Close the Properties window.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 303
SnapshotIQ
g. In the share, open snapfile and verify that it has been restored to the
previous version. Close the file.
SyncIQ
Scenario:
The Engineering business unit has requested a general use data set to test against
their application development effort. The IT manager has asked you to configure
SyncIQ to copy "real-type" data to the engineering group. Use this document as
part of your presentation to the team who are not familiar with the way SyncIQ
works.
The task covers how to create a SyncIQ policy, run the SyncIQ policy, and edit
SyncIQ policy. Also, re-run the policy and view the updated results to achieve the
SyncIQ test.
Given in the lab exercise is the information needed to configure SyncIQ. If needed,
use the WebUI or the CLI step-by-step walk through portion of the exercise for
guidance.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 305
SyncIQ
SyncIQ
1. Use the following information to create and run a SyncIQ policy to copy data
from the marketing general purpose to a target directory in the engineering
access zone.
a. Activate SyncIQ license. For the lab environment, you will use the
trial license functionality.
d. Create the SyncIQ policy. Use the SyncIQ page of the WebUI or the
isi sync policies command.
e. Run a policy assessment for the MktgToEng policy. Use the SyncIQ
page of the WebUI or the isi sync jobs command.
f. View the report for the policy assessment. Use the SyncIQ page of
the WebUI or the isi sync reports command.
g. Run the MktgToEng policy. Use the SyncIQ page of the WebUI or
the isi sync jobs command.
2. Use the following information to modify and test the SyncIQ policy created in
the previous step.
ii. Verify that the policy runs automatically after the copy
operation and view the report. Use the SyncIQ page of the
WebUI or the isi sync reports command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 307
SyncIQ
3. Create target directory targetCopy for the replication. In this exercise, we will
replicate the mktg-share directory to the targetCopy directory in the
engineering access zone.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 309
SyncIQ
a. Map the mktg-share directory, if not already mapped. Ensure that the
directory is not empty. If empty, copy the M & E folder from the File-
Repository directory using Windows Explorer.
b. Enter the information as shown in the graphic. Keep all other fields
with the default setting.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 311
SyncIQ
c. Navigate to the Settings tab and review the default settings. Do not
modify these settings without assistance from PowerScale Technical
Support.
a. On the Policies tab, click More in the Actions column for the
MktgToEng policy, and then select Assess sync.
b. The Assess sync should take less than a minute. Navigate to the
Summary tab and notice the Running status in the Active jobs list.
Once the job completes, it will no longer be listed in the Active jobs
list.
c. Navigate to the Reports tab and view the report, including the
duration, total files, and total data transferred. The policy assessment
did not actually transfer the files, but it provides an estimate of
duration and data to be copied before the first run of the policy.
a. On the Policies tab, click More in the Actions column for the
MktgToEng policy, and then select Start job. The SyncIQ job runs in
the background.
b. You may view the job status in the Active jobs table in the Summary
tab. Once the job completes, navigate to the Reports tab and view
the new report.
c. Navigate to File system, File system explorer. Set the access zone
to engineering. Expand the targetCopy directory. Compare the
contents with the /ifs/marketing/mktg-share directory and verify that
the copy was successful.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 313
SyncIQ
3. Create target directory targetCopy for the replication. In this exercise, we will
replicate the mktg-share directory to the targetCopy directory in the
engineering access zone.
a. Map the mktg-share directory, if not already mapped. Ensure that the
directory is not empty. If empty, copy the M & E folder from the File-
Repository directory using Windows Explorer.
b. The job should take less than a minute to run. Run the isi sync
jobs view MktgToEng command to view the job status. Run the
command every 20 seconds until the job is no more listed. This
indicates that the job has completed.
c. To view the report, we first need the job ID. Run the isi sync
reports list command. Note the Job ID.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 315
SyncIQ
a. To run the policy, at the boston-1 prompt, run the isi sync jobs
start MktgToEng command.
b. Run the isi sync jobs view MktgToEng command to view the
job status. Once done, check the report.
c. Get the job ID by running the isi sync reports list command.
Replace the Job ID in the isi sync reports view MktgToEng
<Job ID> command.
b. Change the policy to run whenever the source is modified. Leave the
rest of the fields unchanged. Click Save changes.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 317
SyncIQ
b. On the WebUI, select the Summary tab on the SyncIQ page. You
may either see the job running in the Active jobs section or a report
generated in the SyncIQ recent reports section. View the report
once the job completes. The MktgToEng policy runs every time the
source directory is modified.
c. Navigate to the File system, File system explorer page. Set the
access zone to engineering and expand the targetCopy directory.
Verify that the legal discovery directory is present.
4. On the File system explorer page, click View/Edit for the targetCopy
directory. Click Edit and remove the Write permission for Group. Save the
changes. You will see an error indicating that you can't make changes to the
file system. This is because the replicated directory is read-only and only the
SyncIQ process can modify it.
b. Run the isi sync jobs view MktgToEng command to view the
job status. Once done, check the report using the isi sync
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 319
SyncIQ
SmartLock
Scenario:
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 321
SmartLock
SmartLock
1. Use the following information to create a SmartLock domain on directory in the
marketing general purpose share.
a. Activate the SmartLock license. For the lab environment, you will use
the trial license functionality.
b. Create a empty target directory for SmartLock. Use the File system
explorer page of the WebUI or the mkdir command.
ii. Enter text into the file. Save the file and close. Reopen the file
and note the contents. Once done, try to delete the file.
Instructor Note: The file can be deleted during its retention period if
the privileged delete feature is enabled. To access the privilege delete
functionality, you must be assigned the
ISI_PRIV_IFS_WORM_DELETE privilege and own the file you are
deleting. You can also access the privilege delete functionality for any
file if you are logged in through the root or compadmin user account.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 323
SmartLock
2. Activate the SmartLock license. Add the trial version for the purpose of this
lab exercise.
a. Navigate to the File system, File system explorer page. Set the
access zone to marketing. Open the mktg-share directory. Click
Create directory and enter the fields as shown.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 325
SmartLock
b. Verify that the domain is listed in the WORM domains table. Click
View/Edit for the created domain to verify the settings.
b. Open the file using Notepad. Enter some text. Save the file and
close. Reopen the file and notice that the text you entered is not
saved.
c. Now, try to delete the file. You will get an error message saying that
the file cannot be deleted. This is because the file is committed to a
WORM state and will only be released after the retention period is
complete. During the retention period, the file is read-only.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 327
SmartLock
d. Try to delete the file, 5 minutes after it was created. The file will
successfully be deleted as the retention period since the last
modification has expired.
2. Activate the SmartLock license. Add the trial version for the purpose of this
lab exercise.
3. Create a empty target directory for SmartLock. At the boston-1 prompt, run
the following commands:
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 329
SmartLock
b. Open the file using Notepad. Enter some text. Save the file and
close. Reopen the file and notice that the text you entered is not
saved.
c. Now, try to delete the file. You will get an error message saying that
the file cannot be deleted. This is because the file is committed to a
WORM state and will only be released after the retention period is
complete. During the retention period, the file is read-only.
d. Try to delete the file, 5 minutes after it was created. The file will
successfully be deleted as the retention period since the last
modification has expired.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 331
HealthCheck
HealthCheck
Scenario:
HealthCheck
1. Use the following information to perform health checks on different cluster
items.
e. Wait for 2 minutes and check the evaluation details for the scheduled
checklists. Use the HealthCheck page of the WebUI or the isi
healthcheck command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 333
HealthCheck
b. Identify the cluster_capacity checklist from the table and click View
details. Note the list of items that will be evaluated in the checklist.
Once done, scroll to the bottom and click Close.
c. For the cluster_capacity checklist, click More and select Run. Once
done, you will notice that the checklist has been queued for
evaluation. The evaluation may take a minute to complete.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 335
HealthCheck
d. Once complete, the Last run start time column for the checklist will
change to PASS or FAIL with a timestamp.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 337
HealthCheck
c. Wait for 2 minutes and check the Evaluations tab. Notice that the
checklists from the newly created schedule are either queued or
evaluated. Click View details to view the evaluation details for any of
the checklists.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 339
HealthCheck
d. Wait for 2 minutes and run the isi healthcheck list command.
Notice that the checklists from the newly created schedule are either
queued or evaluated. Replace the ID in the isi healthcheck
view <ID> command to view the evaluation details.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 341
Monitoring
Monitoring
Scenario:
The IT manager wants to ensure you understand the monitoring tools. Both
InsightIQ and DataIQ are installed on stand-alone servers. Both applications are a
single point of monitoring for multiple PowerScale clusters in the data center.
DataIQ can monitor non-PowerScale volumes, InsightIQ can only monitor
PowerScale volumes.
Given in the lab exercise is the information needed to monitor the cluster using
DataIQ, InsightIQ and isi statistics. If needed, use the step-by-step walk
through portion of the exercise for guidance.
Monitoring
1. Use the following information to perform the initial configuration for InsightIQ.
a. Enable the insightiq user account. Use the Membership and roles
page of the WebUI or the isi auth users command.
i. URL: http://192.168.3.30
v. Username: insightiq
c. Enable and configure the File System Analytics (FSA). Use the
SETTINGS page of the InsightIQ interface.
2. Use to following information to view and create baseline reports on the boston
cluster.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 343
Monitoring
iii. Add a filter to report the Node Performance only for node 1.
d. Create a data usage report showing the largest physical files. Use the
File System Analytics page of the InsightIQ interface.
iii. View the data properties for any available FSA report.
3. Use the following information to add the boston cluster to the DataIQ
application.
a. Create a NFS export to mount /ifs and add the DataIQ server as a
root client. Use the UNIX sharing (NFS) page of the WebUI or the
isi nfs exports command.
c. Add the boston cluster to the DataIQ application. Use the page
Settings > Data management configuration page.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 345
Monitoring
d. View the directory details for /ifs. Use the Data Management page.
4. View the following cluster information using the isi statistics command.
e. View the performance per drive for the node you are logged into.
h. View the performance for the top active results per node.
2. Enable the insightiq user account. The insightiq user is one of the accounts
automatically created by OneFS. By default the account is disabled. You may
either use the WebUI or the CLI to perform this step.
ii. Locate the insightiq user and click View/Edit. Click Edit user.
Change the password to Dees12345! and enable the account
as shown. Once done, save the changes and close.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 347
Monitoring
a. Open the Google Chrome web browser or open a new tab in the
current Chrome browser.
d. The one-time InsightIQ EULA page appears. Review the EULA, click I
have read and agree to..., and then click Submit.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 349
Monitoring
e. The first time you log in, you are prompted to configure the datastore.
For this demonstration, we use a local datastore (as opposed to a
remote datastore) in this implementation of InsightIQ. The local
datastore option is included by default with the InsightIQ virtual
appliance. If you want to place a datastore on a PowerScale storage
cluster or on a NFS-mounted server, you can specify the cluster or
server in this task.
f. Specify the path for the Local Datastore as /datastore and click
Submit.
g. Next, the Add Cluster dialog box appears. Set InsightIQ to monitor
the boston cluster. While in this lab, you will identify the cluster by an
IP address or a host name, alternatively you could specify the name
of a SmartConnect zone.
h. Enter the fields as listed and click OK. It will take a few minutes to
complete adding the cluster. Once done, the monitored cluster is
displayed in the SETTINGS page.
• Username: insightiq
• Password: Dees12345!
b. On the Configuration for boston page, select the Enable FSA tab.
Check the Generate FSA reports and View FSA reports in
InsightIQ features in order to view chart information in the Data
Usage and Data Properties views. Click Submit.
c. Next, Select the FSA Configuration tab. Under the Job Settings
fields, review the Result Set Options settings:
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 351
Monitoring
• Maximum result set age in days: This number represents, in days, how
long to retain each data set before automatically deleting it. To retain data
sets indefinitely, type 0. You can manually delete a data set at any time.
• Maximum result set count: This number represents the maximum number
of data sets to retain. To retain an unlimited number of data sets, type 0.
f. Verify that the fields are set as shown and click Submit.
c. Select the date range from the course start date to the current day
(For example, if day 1 of the course was Monday 4/20/2020 then the
Date Range is Mon, Apr 20, 2020, 8:00 am through the current date
and time). Click View Report.
d. Scroll down for the chart view of reports on different metrics such as
External Network Throughput Rate, Protocol Operations Rate,
CPU, Disk Throughput, etc. Any of the report views created here
can be scheduled, generated as a PDF, and emailed. The report can
be customized allowing administrators to choose exactly which charts
to include in the report.
e. You may also change the zoom level, move forward and move
backward in time. Set the Zoom Level to 6hrs and scroll down to a
chart. Note the start time. Mouse-over the chart data to note the fine
granularity of information.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 353
Monitoring
f. Next, let us add filters to the reports. Scroll to the top of the page and
click Create/manage data filters. Note that the rules can be used to
create filters that allow administrators to quickly pinpoint and view
data.
g. Click Add Rule and then click the drop-down arrow that appears.
Select Node from the list. In the Match column, select 1 to view Node
1. Click Apply.
h. Scroll through the reports and notice that the reports only show
information pertaining to Node 1 as per our filter selection in the
previous step. This filter allows you to view or monitor a specific item.
If Node 1 were having performance issues, this live report would
allow you to filter through the metrics and just present the results from
Node 1, thereby making it easier and faster to perform
troubleshooting or performance analytics.
j. To share a view with other administrators, scroll back to the top of the
page, and in the top right corner of the page, click PERMALINK. The
URL can be sent to an individual. When that person clicks on it, they
can login to the InsightIQ server and see the exact same view.
b. To start with blank report, click Create from Blank Report. There are
many standard templates in the lower half of this page that you can
use as starting point for a custom report.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 355
Monitoring
c. Name the report Lab Performance Report Test. Check the Live
Performance Reporting box. This makes the report show up as an
option in Live Performance Reporting.
e. In the Select a Module for this Position field, select Active Clients.
Note the other options in the field. Select the Protocol radio button.
Click Add another performance module, then select Deduplication
Summary (Physical). Once done, click Finish.
f. This generates reports that show active clients broken down by the
protocols they are using. Confirmation of successfully saving the
report is displayed.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 357
Monitoring
b. Scroll down to the Capacity Forecast section. Under Plot data, keep
the default option of Total Usage checked.
d. Select different options under the Plot Data section, and note how
the presentation of the chart changes.
e. To view Forecast data, hold down the left mouse button and drag it
across the chart for a particular range. This will highlight that area of
the chart. Select the Click to view button to view the forecast data.
b. In the Report Type drop-down list, select Data Usage. On the FSA
Report menu, select one of the available days. Click View Report.
c. Mouse over the pie chart. Note that any of the slices can be selected
and the detail of the directory is shown on the right. All the columns in
the right section can be selected to change how the data is sorted.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 359
Monitoring
d. To view the largest cluster file, scroll down the page to the Top 1000
Files by table and select Largest Physical Size from the drop-down.
e. Scroll to the top of the page. To explore data properties, from the
Report Type menu, select Data Properties. Then from the FSA
Report menu, select an available day. Click View Report.
f. View the file type that contains the most files. Scroll down to the File
Count by Physical Size chart. Next to Breakout by: select the File
Extension option and note the files by extension are listed.
g. With charts like this, you can see files that haven't been used in a
long time and what sizes they are. Continue to explore the graphs
and data views of your cluster as time allows.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 361
Monitoring
2. Create a NFS export for to mount IFS and add the DataIQ server as a root
client. You may either use the WebUI or CLI to perform this step.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 363
Monitoring
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 365
Monitoring
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 367
Monitoring
b. Identify the options for client, drive, heat, list, protocol, pstat,
query, system, and workload. Run the isi statistics with each
option to view the output and type of data. For example, run the isi
statistics client command.
d. View the performance per drive for the node you are logged into by
your SSH session by running the isi statistics drive --
type sas command.
4. Use the isi commands to manipulate additional information about the cluster.
b. Show the read and writes by class using the isi statistics
heat --classes read,write command.
c. Show performance for the top active results per node by running the
isi_for_array "isi statistics drive | head -5"
command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 369
Monitoring
g. Get the detailed broken down cache reads and prefetches by running
the isi_cache_stats -v command.
© Copyright
Internal Use - Confidential 2020 Dell Inc. Page 371
POWERSCALE
HARDWARE CONCEPTS
PARTICIPANT GUIDE
PARTICIPANT GUIDE
Table of Contents
Installation Engagement............................................................................................ 7
Module Objectives ............................................................................................................... 8
Customer Engagement Responsibility ................................................................................. 9
Physical Tools Required .................................................................................................... 10
Installation and Implementation Phases ............................................................................. 12
SolVe ................................................................................................................................. 13
Safety Precautions and Considerations ............................................................................. 16
Onsite Do's and Don'ts....................................................................................................... 18
Appendix ................................................................................................. 79
Glossary .................................................................................................. 89
The graphic shows the PowerScale Solutions Expert certification track. You can
leverage the Dell Technologies Proven Professional program to realize your full
potential. A combination of technology-focused and role-based training and exams
to cover concepts and principles as well as the full range of Dell Technologies'
hardware, software, and solutions. You can accelerate your career and your
organization’s capabilities.
PowerScale Solutions
(C) - Classroom
Prerequisite Skills
To understand the content and successfully complete this course, a student must
have a suitable knowledge base or skill set. The student must have an
understanding of:
• Current PowerScale hardware portfolio and the OneFS operating system
• PowerScale Concepts
• Isilon InfiniBand to Ethernet Backend Conversion
Course Objectives
Installation Engagement
Module Objectives
There are five steps or phases for acquiring a PowerScale cluster. Each phase has
a separate team that engages with the customer. In the design phase a Solution
Architect (SA) works with the customer, determine their specific needs, and
documents what the solution looks like. After the product purchase, shipment, and
delivery to the customer site the install and implementation phase of a PowerScale
cluster begins. The result of the SA engagement is PowerScale Pre-Engagement
Questionnaire (PEQ) that the Customer Engineers (CE) and Implementation
Specialist (IS) uses to install and configure the cluster. Before the install phase, all
design decisions have been made.
3
2 4
1 5
1: The cables that are required are a single CAT5/CAT6 network patch cord, to
directly connect your laptop to the node. USB-to-serial adapter, preferably one that
uses the Prolific 2303 Chipset.
4: Basic hand tools: screwdrivers (flat-head and Phillips), wire cutters, anti-static
wrist strap.
There are three distinct steps in the install and implementation phase: Install, Build,
and Implement.
1: During the install, the components are unpacked and racked, and switches are
rack that is mounted. Nodes are connected to the back-end switches, power is
added, and front-end network cables are connected between the cluster and
customer network. The Customer Engineer or CE performs these tasks.
2: Depending on the role, the CE may perform the cluster build also. The cluster
build is achieved when the system is powered on, the PowerScale Configuration
Wizard has been launched and the information added.
3: In some regions, running the Configuration Wizard may be the sole responsibility
of the IS. After the cluster is built, the IS configures the features of OneFS as
written in the PEQ.
SolVe
Before you arrive at a client site, remember to read the call notes and follow the
processes that are detailed in them. Check if there are any special instructions from
PowerScale Technical Support that you must follow.
1: Download SolVe Desktop application on the system. Go to the Tools and Sites
section, choose SolVe. And select SolVe Desktop Executable. Depending on the
browser used, you may be presented with security dialogue boxes. Take the
needed actions to launch the executable.
2:
Click through the Setup wizard and then select Install. Clicking Finish launches
the SolVe Desktop. SolVe must be authorized for use. Select OK. A few general
items1.
1 Notice the dialog in the lower left showing the version. This area also shows the
progress when upgrading and downloading content. Also notice in the lower right
the service topics. Once connected, many of articles that are shown may not be
3: From the menu, select Authorize and download the list of available products.
Adhere to the instructions shown, that is to leave SolVe open, enter credentials,
this is using SSO, and open the keychain file. Select OK. And then go to
downloads and open the keychain file.
4: Next are the Release Notes. Review and then close this window. Bring back the
SolVe. Notice the dialog2 in the lower left indicating the keychain is loaded, that
means you are authorized, and content is updated. Now, scroll down, and click
PowerScale to gather the PowrScale content.
5: Click OK. Again, note the progress in the lower left. Once the download is
complete, you see that the PowerScale image has changed. Tools that are
downloaded appear in the upper left corner of the screen without the green arrow
present.
6: Now you can click PowerScale and view the available procedures. If updates are
available for download, you see an information icon, click the icon, and approve the
updated content download.
2The icons with a green arrow indicate that the user must click the icon in order to
download the tool.
3The view is dependent upon Partner Type. A service partner sees what an
employee sees, a direct sales partner sees what a customer sees, and an
ASP/ASN partner sees products depending upon credentials.
When working with PowerScale equipment, it is critical to ensure you adhere to the
following precautions.
6
1
2
5
3
1: The AC supply circuit for PowerScale nodes must supply the total current that is
specified on the label of the node. All AC power supply connections must be
properly grounded. Connections that are not directly connected to the branch
circuit, such as nodes that are connected to a power strip, must also be properly
grounded. Do not overload the branch circuit of the AC supply that provides power
to the rack holding PowerScale nodes. The total rack load should not exceed 80%
of the branch circuit rating. For high availability, the left and right sides of any rack
must receive power from separate branch feed circuits. To help protect the system
from sudden increases or decreases in electrical power, use a surge suppressor,
line conditioner, or uninterruptible power supply or UPS.
2: To avoid personal injury or damage to the hardware, always use two people to
lift or move a node or chassis. A Gen 6 chassis can weigh more than 200 lbs. It is
recommended to use a lift to install the components into the rack. If a lift is not
available, you must remove all drive sleds and compute modules from the chassis
before lifting. Even when lifting an empty chassis, never attempt to lift and install
with fewer than two people.
3:
Electrostatic Discharge
4: If you install PowerScale nodes in a rack that is not bolted to the floor, use both
front and side stabilizers. Installing PowerScale nodes in an unbolted rack without
these stabilizers could cause the rack to tip over, potentially resulting in bodily
injury. Use only approved replacement parts and equipment.
6: You can install racks in raised or nonraised floor data centers capable of
supporting that system. It is your responsibility to ensure that data center floor can
support the weight of the system. A fully populated rack with A2000 chassis’
weighs about 3,500 lbs (1,590 kg). If the floor is rated at less than 3,500 lbs, then
additional care and planning must be taken. Some data center floors have different
static load vs. dynamic (rolling) load specifications, and sectional weight and load
point limits. This becomes important while moving preracked solutions around the
data center.
4Failure to heed these warnings may also void the product warranty. Only trained
and qualified personnel should install or replace equipment. Select the button
options for specific information. Always refer to the current Site Preparation and
Planning Guide for proper procedures and environmental information.
When onsite, remember to represent Dell EMC and yourself in the best possible
light. Do not change the PEQ without the approval of the design team. Any
approved changes should be meticulously tracked and any appropriate change
control processes should be followed. Remember to bring your documentation and
copies to provide to the customer.
Tip: To make an Install Base entry, use the IB Status Change page
link.
Module Objectives
5The Gen 6 platform reduces the data center rack footprints with support for four
nodes in a single 4U chassis. It enables enterprise to take on new and more
demanding unstructured data applications. The Gen 6 can store, manage, and
protect massively large datasets with ease. With the Gen 6, enterprises can gain
new levels of efficiency and achieve faster business outcomes.
6 The ideal use cases for Gen 6.5 (F200 and F600) is remote office/back office,
factory floors, IoT, and retail. Gen 6.5 also targets smaller companies in the core
verticals, and partner solutions, including OEM. The key advantages are low entry
price points and the flexibility to add nodes individually, as opposed to a chassis/2
node minimum for Gen 6.
The design goal for the PowerScale nodes is to keep the simple ideology of NAS,
provide the agility of the cloud, and the cost of commodity.
The Gen 6x family has different offerings that are based on the need for
performance and capacity. As Gen 6 is a modular architecture, you can scale out
compute and capacity separately. OneFS powers all the nodes.
Gen 6
• F-Series
• H-Series
• A-series
Gen 6.5
Gen 6.5 requires a minimum of three nodes to form a cluster. You can add single
nodes to the cluster. The F600 and F200 are a 1U form factor and based on the
R640 architecture.
• F6007
• F2008
7Mid-level All Flash Array 1U PE server with 10 (8 usable) x 2.5” drive bays,
enterprise NVMe SSDs (RI, 1DWPD), data reduction standard. Front End
networking options for 10/25 GbE or 40/100 GbE and 100 GbE Back End. Also
called as Cobalt Nodes.
8 Entry-level All Flash Array 1U PE server with 4 x 3.5” drive bays (w/ 2.5” drive
trays), enterprise SAS SSDs (RI, 1DWPD), data reduction standard. 10/25 GbE
Front/Back End networking. Also called as Sonic Nodes.
Gen 6 requires a minimum of four nodes to form a cluster. You must add nodes to
the cluster in pairs.
The chassis holds four compute nodes and 20 drive sled slots.
Both compute modules in a node pair power-on immediately when one of the
nodes is connected to a power source.
Gen 6 chassis
1 10 9
2 8
4
6
3
5 7
1: The compute module bay of the two nodes make up one node pair. Scaling out a
cluster with Gen 6 nodes is done by adding more node pairs.
2: Each Gen 6 node provides two ports for front-end connectivity. The connectivity
options for clients and applications are 10 GbE, 25 GbE, and 40 GbE.
3: Each node can have 1 or 2 SSDs that are used as L3 cache, global namespace
acceleration (GNA), or other SSD strategies.
4: Each Gen 6 node provides two ports for back-end connectivity. A Gen 6 node
supports 10 GbE, 40 GbE, and InfiniBand.
5: Power supply unit - Peer node redundancy: When a compute module power
supply failure takes place, the power supply from the peer node temporarily
provides power to both nodes.
6: Each node has five drive sleds. Depending on the length of the chassis and type
of the drive, each node can handle up to 30 drives or as few as 15.
8: The sled can be either a short sled or a long sled. The types are:
9: The chassis comes in two different depths, the normal depth is about 37 inches
and the deep chassis is about 40 inches.
10: Large journals offer flexibility in determining when data should be moved to the
disk. Each node has a dedicated M.2 vault drive for the journal. A node mirrors
their journal to its peer node. The node writes the journal contents to the vault when
a power loss occurs. A backup battery helps maintain power while data is stored in
the vault.
Gen 6.5 requires a minimum of three nodes to form a cluster. You can add single
nodes to the cluster. The F600 and F200 are a 1U form factor and based on the
R640 architecture.
1
5
8 2
7 4
1: Scaling out an F200 or an F600 node pool only requires adding one node.
3: Each Gen F200 and F600 node provides two ports for backend connectivity. The
PCIe slot 1 is used.
4: Redundant power supply units - When a power supply fails, the secondary
power supply in the node provides power. Power is supplied to the system equally
from both PSUs when the Hot Spare feature is disabled. Hot Spare is configured
using the iDRAC settings.
5: Disks in a node are all the same type. Each F200 node has four SAS SSDs.
6: The nodes come in two different 1U models, the F200 and F600. You need
nodes of the same type to form a cluster.
7: The F200 front-end connectivity uses the rack network daughter card (rNDC).
Important: The F600 nodes have a 4-port 1 GB NIC in the rNDC slot.
OneFS does not support this NIC on the F600.
Gen 6 Chassis
All Gen 6 chassis come with the front panel and the front panel display module.
The front panel covers the drive sleds while allowing access to the display.
Movie:
The web version of this content contains a movie.
Script: This demonstration takes a tour of the Gen 6 front panel display, drive
sleds, and an outside look at the node’s compute modules. We’ll focus on
identifying components and indicator function.
We’ll start the tour on the front panel display. This allows various administrative
tasks and provides alerts. There are 5 navigation buttons that let the administrator
select each node to administer. There are 4 node status indicators. If a node’s
status light indicator is yellow, it indicates a fault with the corresponding node. The
product badges indicate the types of nodes installed in the chassis. Only two
badges are necessary because nodes can only be installed in matched adjacent
node pairs. The front panel display is hinged to allow access to the drive sleds it
covers and contains LEDs to help the administrator see the status of each node.
Sleds
Now, taking the front bezel off the chassis and you will see the drive sleds for the
nodes. The Gen 6 chassis has 20 total drive sled slots that can be individually
serviced, but only one sled per node can be safely removed at a time. The graphic
shows that each node is paired with 5 drive sleds. The status lights on the face of
the sled indicate whether the sled is currently in service, and whether the sled
contains a failing drive. The service request button informs the node that the sled
needs to be removed, allowing the node to prepare it for removal by moving key
boot information away from drives in that sled. This temporarily suspends the
drives in the sled from the cluster file system, and then spins them down. This is
done to maximize survivability in the event of further failures and protect the cluster
file system from the effect of having several drives temporarily go missing. The do-
not-remove light blinks while the sled is being prepared for removal, and then turns
off when it is ready. We’ll see this here. The sleds come in different types. First,
when configured for nodes that support 3.5" drives, there are 3 drives per sled, as
shown here, equaling 15 drives per node, making 60 drives per chassis. The
second type is a longer sled that holds four 3.5” drives. This is used in the deep
archive, deep rack chassis for A2000 nodes. The long sleds have 20 drives per
node, for up to 80 3.5" drives per chassis. In the 3.5" drive sleds, the yellow LED
drive fault lights are on the paddle cards attached to the drives, and they are also
visible through the cover of the drive sled as indicated here. The long sled has 4
LED viewing locations. The third type of sled applies to nodes supporting 2.5"
drives. The 2.5” drive sleds can have 3 or 6 drives per sled (as shown), 15 or 30
drives per node, making 60 or 120 drives per fully populated chassis. Internally to
the 2.5" sled, there are individual fault lights for each drive. The yellow LED
associated with each drive is visible through holes in the top cover of the sled so
that you can see which drive needs replacement. The LED will stay on for about 10
minutes while the sled is out of the chassis.
Compute
When we look at the back, we see the four nodes’ compute modules in the chassis’
compute bays. We also see the terra cotta colored release lever on each compute
module, secured by a thumb screw. As shown, compute module bay 1 and 2 make
up one node pair and bay 3 and 4 make up the other node pair. In the event of a
compute module power supply failure, the power supply from the peer compute
module in the node pair will temporarily provide power to both nodes. Let’s move
the upper right of a compute module. The top light is a blue, power LED and below
that is an amber, fault LED. Each compute module has a ‘DO NOT REMOVE’
indicator light which is shaped like a raised hand with a line through it. To service
the compute module in question, shut down the affected node and wait until the
‘DO NOT REMOVE’ light goes out. Then it is safe to remove and service the unit in
question. The uHDMI port is used for factory debugging. The PCIE card on the
right is for external network connectivity and the left PCIE card is for internal
network connectivity. The compute module has a 1GbE management port, and the
DB9 serial console port. Each compute module has either a 1100W dual-voltage
(low and medium compute) or a 1450W high-line (240V) only (high and ultra-
compute) power supply unit. If high-line only nodes are being installed in a low-line
(120V) only environment, two 1U rack-mountable step-up transformers are required
for each Gen 6 chassis. Always keep in mind that Gen 6 nodes do not have power
buttons - both compute modules in a node pair will power on immediately when one
is connected to a live power source. There are also status indicator lights such as
the PSU fault light. All nodes have an ASHRAE (American Society of Heating,
This hardware tour will take a deeper look inside the node’s compute module.
Movie:
The web version of this content contains a movie.
Script: This demonstration takes a tour of the inside of the Gen 6 compute module.
First, let’s take at the back of the chassis. The chassis can have two or four
compute modules. Remember that a node is a ¼ of the chassis and consists of a
compute module and five drive sleds. Each node pairs with a peer node to form a
node pair. Shown here, nodes three and four form a node pair. Let’s start by
removing the node’s compute module to get a look inside. This demonstration does
not use a powered system. This tour does not highlight the steps for removing
components. Remember to always follow the proper removal and install
procedures from the SolVe Desktop.
WARNING: Only qualified Dell EMC personnel are allowed to open compute
nodes.
Let’s remove the node’s lid. This can be a bit tricky on the first time. Pull the blue
release handle without pressing down on the lid. Pressing down on the lid while
trying to open will keep the node lid from popping up. The lid portion of the compute
module holds the motherboard, CPU and RAM. There are two different
motherboard designs to accommodate different CPU types; the performance-based
Broadwell-EP or the cost optimized Broadwell-DE. Shown here is the Broadwell-DE
based board that the H400, A200, and A2000 use. Note the position of the four
DIMMs and their slot numbering. Here is the Broadwell-EP based board that the
F800, H600 and H500 use. Note the position of the four DIMMs and their slot
numbering. The DIMMs are field replaceable units. The CPU is not. Due to the
density and positioning of motherboard components around the DIMM slots,
damage to the motherboard is possible if care is not taken while removing and
installing DIMM modules.
Let’s turn to the lower portion of the compute module. First, we see the fan module.
This is a replaceable unit. Shown is the release lever for the fans.
The riser card, on the right side of the compute module, contains the PCIE card
slots, the NVRAM vault battery, and the M.2 card containing the NVRAM vault.
Let’s remove this to get a closer look. Removing the riser card can be tricky the first
time. Note the two blue tabs for removing the HBA riser, a sliding tab at the back
and a fixed tab at the front. At the same time, push the sliding tab in the direction of
the arrow on the tab and free the front end by pulling the riser away from the
locking pin on the side of the chassis with the fixed tab. Lift the tabs to unseat the
riser and pull it straight up. Try this at least once before going onsite to replace a
component. Here are the two PCIe slots and the ‘Pelican’ slot. They are x4 or x8
depending on the performance level of the node. The internal NIC for
communication between nodes is the PCI card shown on the left, the external PCI
card is on the right. The external NIC is used for client and application access.
Depending on the performance level of the node, the external NIC may either be a
full-size PCIe card facing left, or a ‘Pelican’ card connected to the smaller
proprietary slot between the two PCIe slots and facing right.
Next is the battery. The backup battery maintains power to the compute node while
journal data is being stored in the M.2 vault during an unexpected power loss
event. Note that because the riser card and the battery are paired, if the battery
needs to be replaced, it is replaced together with the riser card. Lastly, as seen
here, the M.2 vault disk is located under the battery. The M.2 vault disk is also a
field replaceable unit. This concludes the inside tour. Remember to review the
documentation on the SolVe Desktop for proper removal and replacement of the
node’s compute module components.
Generation 6 4U Node
New PowerScale F600 nodes with full NVMe support deliver massive performance
in a compact form factor. OneFS delivers up to 80% storage utilization for
maximum storage efficiency. Data deduplication can further reduce storage
requirements by up to 30% and inline data compression on the F200, F600, F810
all-flash platforms, and the H5600 hybrid platform can reduce the space that is
consumed.
Generation 6 Terminologies
Pre-Engagement Questionnaire
Module Objectives
Job Roles
There are four job roles that are associated with PowerScale hardware installation
and implementation process.
Pre-Engagement Questionnaire
The PowerScale PEQ is the replacement for the Configuration Guide. The stated
purpose of the PEQ is to document the Professional Services project installation
parameters and to facilitate the communication between the responsible resources.
The PEQ incorporates the process workflow and eases hand-off from Pre-Sales to
Delivery. It is a delivery document, which benefits other roles, helps define roles
and responsibilities and is not the same as the Qualifier.
PEQ Tour
Cover
To start the application, open the PEQ spreadsheet tool. The first tab that is
displayed is the Cover tab. The Cover tab contains the creation date and the
customer name.
Begin filling out the document from upper left to bottom right. SE shares the
Customer contact information and describes at a high level what the project team is
expected to do at each site, using the provided drop-down menus. The SE also
provides general customer environment information, such as Operating Systems in
use, backup apps and protocols, and any specialty licenses sold. Accurate and
On the Solution Diagram tab, the SE provides the solution diagrams or topologies
that are used during the presales cycle.
Checklist (PM)
Project Manager begins with the Engagement Checklist tab to help them plan
project tasks with a great deal of granularity.
It is also the responsibility of the Project Manager to maintain the Data Center
readiness information about the Project Details tab. Here the PM focuses on
verifying that each site has met the power, cooling, networking, and other
prerequisites before scheduling resources. The PM should also complete the
Administrative Details section with team member information, project Id details, and
an optional timeline.
Hardware
The hardware tab shows the physical connections parameters and some basic
logical parameters necessary to “stand up” the cluster. When multiple node types
are selected and defined on the Engagement Details tab, the Cluster Details
section includes a complete listing of the extended Node Details and Front-End
Switch details.
Cluster
The Cluster tab represents a single cluster and its logical configuration. Each
section on the Cluster Tab has a designated number (Yellow Chevron). The
numbers represent the listed priority of that section and should be completed in
order starting with number one. This tab is split into sections that describe different
features. These tabs are enabled through the questions in the Licensing \ Features
section.
Reference
Note: The Solution Architect (SA) typically fills out the PEQ.
Module Objectives
Network: There are two types of networks that are associated with a cluster:
internal and external.
Ethernet
Clients connect to the cluster using Ethernet connections10 that are available on all
nodes.
9 In general, keeping the network configuration simple provides the best results with
the lowest amount of administrative overhead. OneFS offers network provisioning
rules to automate the configuration of additional nodes as clusters grow.
10Because each node provides its own Ethernet ports, the amount of network
bandwidth available to the cluster scales linearly.
OneFS supports a single cluster11 on the internal network. This back-end network,
which is configured with redundant switches for high availability, acts as the
backplane for the cluster.12
12 This enables each node to act as a contributor in the cluster and isolating node-
to-node communication to a private, high-speed, low-latency network. This back-
end network utilizes Internet Protocol (IP) for node-to-node communication.
The Gen 6x back-end topology in OneFS 8.2 and later supports scaling a
PowerScale cluster to 252 nodes. See the participant guide for more details.
27 uplinks per
spine switch
Leaf-Spine is a two-level hierarchy where nodes connect to leaf switches, and leaf
switches connects to spine switches. Leaf switches do not connect to one another,
and spine switches do not connect to one another. Each leaf switch connects with
each spine switch and all leaf switches have the same number of uplinks to the
spine switches.
The new topology uses the maximum internal bandwidth and 32-port count of Dell
Z9100 switches. When planning for growth, F800 and H600 nodes should connect
over 40 GbE ports whereas A200 nodes may connect using 4x1 breakout cables.
Scale planning enables for nondisruptive upgrades, meaning as nodes are added,
no recabling of the backend network is required. Ideally, plan for three years of
growth. The table shows the switch requirements as the cluster scales. In the table,
Max Nodes indicate that each node is connected to a leaf switch using a 40 GbE
port.
5. Create a cluster by using any four nodes on the first Leaf switch.
6. Confirm that OneFS 8.2 or later is installed on the cluster.
7. Add the remaining nodes to the cluster that was created in step 5.
8. Confirm the cluster installation by checking the CELOG events.
Legacy Connectivity
Three types of InfiniBand cable are used with currently deployed clusters. Older
nodes and switches, which run at DDR or SDR speeds use the legacy CX4
connector. In mixed environments (QDR nodes and DDR switch, or conversely) a
hybrid IB cable is used. This cable has a CX4 connector on one end and a QSFP
connector on the other. However, QDR nodes are incompatible with SDR switches.
On each cable, the connector types identify the cables. The graphic shows, the
combination of the type of node and the type of InfiniBand switch port determines
the correct cable type.
Node Interconnectivity
1: Backend ports int-a and int-b. The int-b port is the upper port. Gen 6 backend
ports are identical for InfiniBand and Ethernet and cannot be identified by looking at
the node. If Gen 6 nodes are integrated in a Gen 5 or earlier cluster, the backend
will use InfiniBand. Note that there is a procedure to convert an InfiniBand backend
to Ethernet if the cluster no longer has pre-Gen 6 nodes.
2: PowerScale nodes with different backend speeds can connect to the same
backend switch and not see any performance issues. For example, an environment
has a mixed cluster where A200 nodes have 10 GbE backend ports and H600
nodes have 40 GbE backend ports. Both node types can connect to a 40 GbE
switch without effecting the performance of other nodes on the switch. The 40 GbE
switch provides 40 GbE to the H600 nodes and 10 GbE to the A200 nodes.
3: There are two speeds for the backend Ethernet switches, 10 GbE and 40 GbE.
Some nodes, such as archival nodes, might not need to use all of a 10 GbE port
bandwidth while other workflows might need the full utilization of the 40 GbE port
bandwidth. The Ethernet performance is comparable to InfiniBand so there should
be no performance bottlenecks with mixed performance nodes in a single cluster.
Administrators should not see any performance differences if moving from
InfiniBand to Ethernet.
Gen 6 nodes can use either an InfiniBand or Ethernet switch on the backend.
InfiniBand was designed as a high-speed interconnect for high-performance
computing, and Ethernet provides the flexibility and high speeds that sufficiently
support the PowerScale internal communications.
Gen 6.5 only supports Ethernet. All new, PowerScale clusters support Ethernet
only.
The graphic shows a closer look at the external and internal connectivity. Slot 1 is
used for backend communication on both the F200 and F600. Slot 3 is used for the
F600 2x 25 GbE or 2x 100 GbE front-end network connections. The rack network
daughter card (rNDC) is used for the F200 2x 25 GbE front-end network
connections.
Note: The graphic shows the R640 and does not represent the F200 and F600 PCIe and rNDC
configuration.
The external network provides connectivity for clients over standard file-based
protocols. It supports link aggregation, and network scalability is provided through
software in OneFS. A Gen 6 node has to 2 front-end ports - 10 GigE, 25 GigE, or
40 GigE, and one 1 GigE port for management. Gen 6.5 nodes have 2 front-end
ports - 10 GigE, 25 GigE, or 100 GigE.
Breakout Cables
The 40 GbE and 100 GbE connections are 4 individual lines of 10 GbE and 25
GbE. Most switches support breaking out a QSFP port into four SFP ports using a
1:4 breakout cable. The backend is done automatically when the switch detects the
cable type as a breakout cable. The front end is often configured manually on a per
port basis.
Cabling Considerations
Module Objectives
• Serial Console
• Web Administration Interface (WebUI)
• Command Line Interface (CLI).
• OneFS Application Programming Interface (API)
• Front Panel Display
Movie:
Link:
https://edutube.emc.com/Player.aspx?vno=KjBgi9m8LmZLw58klDHmOA==&autopl
ay=true
Script: Four options are available for managing the cluster. The web
administration interface (WebUI), the command-line interface (CLI), the serial
console, or the platform application programming interface (PAPI), also called the
OneFS API. The first management interface that you may use is a serial console to
node 1. A serial connection using a terminal emulator, such as PuTTY, is used to
initially configure the cluster. The serial console gives you serial access when you
cannot or do not want to use the network. Other reasons for accessing using a
serial connection may be for troubleshooting, site rules, a network outage, and so
on. Shown are the terminal emulator settings.
Configuration Manager
For initial configuration, access the CLI by establishing a serial connection to the
node designated as node 1. The serial console gives you serial access when you
cannot or do not want to use the network. Other reasons for accessing using a
serial connection may be for troubleshooting, site rules, a network outage, so on.
Serial Port14
14 The serial port is usually a male DB9 connector. This port is called the service
port. Connect a serial null modem cable between a serial port of a local client, such
as a laptop, and the node service port. Connect to the node designated as node 1.
As most laptops today no longer have serial ports, you might need to use a USB-
to-serial converter. On the local client, launch a serial terminal emulator.
• Data bits = 8
• Parity = none
• Stop bits = 1
• Flow control = hardware
isi config
Changes
prompt to
>>>
The isi config command, pronounced "izzy config," opens the configuration
console. The console contains configured settings from the time the Wizard started
running.
Use the console to change initial configuration settings. When in the isi config
console, other configuration commands are unavailable. The exit command is
used to go back to the default CLI.
OneFS
version
User must have logon privileges
Connect to
any node in
cluster over
HTTPS on
port 8080
The WebUI requires at least one IP address that is configured16 on one of the
external Ethernet ports presents in one of the nodes.
choose option 1, the Configuration Wizard steps you through the process of
creating a cluster. If you choose option 2, the Configuration Wizard ends after the
node finishes joining the cluster. You can then configure the cluster using the
WebUI or the CLI.
• Out-of-band17
• In-band18
Both methods are done using any SSH client such as OpenSSH or PuTTY. Access
to the interface changes based on the assigned privileges.
OneFS commands are code that is built on top of the UNIX environment and are
specific to OneFS management. You can use commands together in compound
command structures combining UNIX commands with customer facing and internal
commands.
4
1
5
3 6
17Accessed using a serial cable that is connected to the serial port on the back of
each node. As many laptops no longer have a serial port, a USB-serial port adapter
may be needed.
5: The CLI command use includes the capability to customize the base command
with the use of options, also known as switches and flags. A single command with
multiple options result in many different permutations, and each combination
results in different actions performed.
6: The CLI is a scriptable interface. The UNIX shell enables scripting and execution
of many UNIX and OneFS commands.
CLI Usage
Option explanation
The man isi or isi --help command is an important command for a new
administrator. These commands provide an explanation of the available isi
commands and command options. You can also view a basic description of any
command and its available options by typing the -h option after the command.
OneFS applies authentication and RBAC controls to API commands to ensure that
only authorized commands are run.
19A chief benefit of PAPI is its scripting simplicity, enabling customers to automate
their storage administration.
3: Some commands are not PAPI aware, meaning that RBAC roles do not apply.
These commands are internal, low-level commands that are available to
administrators through the CLI. Commands not PAPI aware: isi config, isi
get, isi set, and isi services
4: The number indicates the PAPI version. If an upgrade introduces a new version
of PAPI, some backward compatibility ensures that there is a grace period for old
scripts to be rewritten.
The Gen 6 front panel display is an LCD screen with five buttons that are used for
basic administration tasks20.
The Gen 6.5 front panel has limited functionality21 compared to the Gen 6.
20Some of them include adding the node to a cluster, checking node or drive
status, events, cluster details, capacity, IP and MAC addresses.
21You can join a node to a cluster and the panel display node name after the node
has joined the cluster.
Course Summary
Course Summary
Now that you have completed this course, you should be able to:
→ Discuss installation engagement actions.
→ Explain the use of PEQ in implementation.
→ Describe PowerScale nodes.
→ Identify the PowerScale node internal and external networking components.
→ Explain the PowerScale cluster management tools.
Electrostatic Discharge
Electrostatic Discharge is a major cause of damage to electronic components and
potentially dangerous to the installer. To avoid ESD damage, review ESD
procedures before arriving at the customer site and adhere to the precautions when
onsite.
Antistatic Packaging:
Leave components in
antistatic packaging until
time to install.
PowerScale Nodes
Individual PowerScale nodes provide the data storage capacity and processing
power of the PowerScale scale-out NAS platform. All of the nodes are peers to
each other and so there is no single 'master' node and no single 'administrative
node'.
• No single master
• No single point of administration
Administration can be done from any node in the cluster as each node provides
network connectivity, storage, memory, non-volatile RAM (NVDIMM) and
processing power found in the Central Processing Units (CPUs). There are also
different node configurations, compute, and capacity. These varied configurations
can be mixed and matched to meet specific business needs.
Each contains.
• Disks
• Processor
• Cache
Tip: Gen 6 nodes can exist within the same cluster. Every
PowerScale node is equal to every other PowerScale node of the
same type in a cluster. No one specific node is a controller or filer.
F-Series
The F-series nodes sit at the top of both performance and capacity with all-flash
arrays for ultra-compute and high capacity. The all flash platforms can accomplish
250-300k protocol operations per chassis and get 15 GB/s aggregate read
throughput from the chassis. Even when the cluster scales, the latency remains
predictable.
• F80022
• F81023
22 The F800 is suitable for workflows that require extreme performance and
efficiency. It is an all-flash array with ultra-high performance. The F800 sits at the
top of both the performance and capacity platform offerings when implementing the
15.4TB model, giving it the distinction of being both the fastest and densest Gen 6
node.
23 The F810 is suitable for workflows that require extreme performance and
efficiency. The F810 also provides high-speed inline data deduplication and in-line
data compression. It delivers up to 3:1 efficiency, depending on your specific
dataset and workload.
H-Series
After F-series nodes, next in terms of computing power are the H-series nodes.
These are hybrid storage platforms that are highly flexible and strike a balance
between large capacity and high-performance storage to provide support for a
broad range of enterprise file workloads.
• H40024
• H50025
• H560026
• H60027
25The H500 is a versatile hybrid platform that delivers up to 5 GB/s bandwidth per
chassis with a capacity ranging from 120 TB to 720 TB per chassis. It is an ideal
choice for organizations looking to consolidate and support a broad range of file
workloads on a single platform. H500 is comparable to a top of the line X410,
combining a high compute performance node with SATA drives. The whole Gen 6
architecture is inherently modular and flexible with respect to its specifications.
26The H5600 combines massive scalability – 960 TB per chassis and up to 8 GB/s
bandwidth in an efficient, highly dense, deep 4U chassis. The H5600 delivers inline
data compression and deduplication. It is designed to support a wide range of
demanding, large-scale file applications and workloads.
A-Series
The A-series nodes namely have lesser compute power compared to other nodes
and are designed for data archival purposes. The archive platforms can be
combined with new or existing all-flash and hybrid storage systems into a single
cluster that provides an efficient tiered storage solution.
• A20028
• A200029
28The A200 is an ideal active archive storage solution that combines near-primary
accessibility, value and ease of use.
29The A2000 is an ideal solution for high density, deep archive storage that
safeguards data efficiently for long-term retention. The A2000 is capable of
containing 80, 10TB drives for 800TBs of storage by using a deeper chassis with
longer drive sleds containing more drives in each sled.
OneFS CLI
The command-line interface runs "isi" commands to configure, monitor, and
manage the cluster. Access to the command-line interface is through a secure shell
(SSH) connection to any node in the cluster.
PAPI
The customer uses OneFS application programming interface (API) to automate
the retrieval of the most detailed network traffic statistics. It is divided into two
functional areas: One area enables cluster configuration, management, and
monitoring functionality, and the other area enables operations on files and
directories on the cluster.
Serial Console
The serial console is used for initial cluster configurations by establishing serial
access to the node designated as node 1.
WebUI
The browser-based OneFS web administration interface provides secure access
with OneFS-supported browsers. This interface is used to view robust graphical
monitoring displays and to perform cluster-management tasks.
PARTICIPANT GUIDE
PARTICIPANT GUIDE
Table of Contents
Appendix ................................................................................................. 89
The graphic shows the PowerScale Solutions Expert certification track. You can
leverage the Dell Technologies Proven Professional program to realize your full
potential. A combination of technology-focused and role-based training and exams
to cover concepts and principles as well as the full range of Dell Technologies'
hardware, software, and solutions. You can accelerate your career and your
organization’s capabilities.
PowerScale Solutions
(C) - Classroom
Prerequisite Skills
To understand the content and successfully complete this course, a student must
have a suitable knowledge base or skill set. The student must have an
understanding of:
• Current PowerScale hardware portfolio and the OneFS operating system
• PowerScale Concepts
• Isilon InfiniBand to Ethernet Backend Conversion
Course Objectives
Module Objectives
Installation References
The Isilon Info Hubs page has links to specific OneFS software, hardware, and
troubleshooting pages. The OneFS in hubs have documents for installation-related
areas.
When working with PowerScale equipment, it is critical to ensure you adhere to the
following precautions.
6
1
2
5
3
1: The AC supply circuit for PowerScale nodes must supply the total current that is
specified on the label of the node. All AC power supply connections must be
properly grounded. Connections that are not directly connected to the branch
circuit, such as nodes that are connected to a power strip, must also be properly
grounded. Do not overload the branch circuit of the AC supply that provides power
to the rack holding PowerScale nodes. The total rack load should not exceed 80%
of the branch circuit rating. For high availability, the left and right sides of any rack
must receive power from separate branch feed circuits. To help protect the system
from sudden increases or decreases in electrical power, use a surge suppressor,
line conditioner, or uninterruptible power supply or UPS.
2: To avoid personal injury or damage to the hardware, always use two people to
lift or move a node or chassis. A Gen 6 chassis can weigh more than 200 lbs. It is
recommended to use a lift to install the components into the rack. If a lift is not
available, you must remove all drive sleds and compute modules from the chassis
before lifting. Even when lifting an empty chassis, never attempt to lift and install
with fewer than two people.
3:
Electrostatic Discharge
4: If you install PowerScale nodes in a rack that is not bolted to the floor, use both
front and side stabilizers. Installing PowerScale nodes in an unbolted rack without
these stabilizers could cause the rack to tip over, potentially resulting in bodily
injury. Use only approved replacement parts and equipment.
6: You can install racks in raised or nonraised floor data centers capable of
supporting that system. It is your responsibility to ensure that data center floor can
support the weight of the system. A fully populated rack with A2000 chassis’
weighs about 3,500 lbs (1,590 kg). If the floor is rated at less than 3,500 lbs, then
additional care and planning must be taken. Some data center floors have different
static load vs. dynamic (rolling) load specifications, and sectional weight and load
point limits. This becomes important while moving preracked solutions around the
data center.
1Failure to heed these warnings may also void the product warranty. Only trained
and qualified personnel should install or replace equipment. Select the button
options for specific information. Always refer to the current Site Preparation and
Planning Guide for proper procedures and environmental information.
SolVe
Before you arrive at a client site, remember to read the call notes and follow the
processes that are detailed in them. Check if there are any special instructions from
PowerScale Technical Support that you must follow.
1: Download SolVe Desktop application on the system. Go to the Tools and Sites
section, choose SolVe. And select SolVe Desktop Executable. Depending on the
browser used, you may be presented with security dialogue boxes. Take the
needed actions to launch the executable.
2:
Click through the Setup wizard and then select Install. Clicking Finish launches
the SolVe Desktop. SolVe must be authorized for use. Select OK. A few general
items2.
2 Notice the dialog in the lower left showing the version. This area also shows the
progress when upgrading and downloading content. Also notice in the lower right
the service topics. Once connected, many of articles that are shown may not be
3: From the menu, select Authorize and download the list of available products.
Adhere to the instructions shown, that is to leave SolVe open, enter credentials,
this is using SSO, and open the keychain file. Select OK. And then go to
downloads and open the keychain file.
4:
Next are the Release Notes. Review and then close this window. Bring back the
SolVe. Notice the dialog3 in the lower left indicating the keychain is loaded, that
means you are authorized, and content is updated. Now, scroll down, and click
PowerScale to gather the PowrScale content.
5: Click OK. Again, note the progress in the lower left. Once the download is
complete, you see that the PowerScale image has changed. Tools that are
downloaded appear in the upper left corner of the screen without the green arrow
present.
6: Now you can click PowerScale and view the available procedures. If updates are
available for download, you see an information icon, click the icon, and approve the
updated content download.
3The icons with a green arrow indicate that the user must click the icon in order to
download the tool.
4The view is dependent upon Partner Type. A service partner sees what an
employee sees, a direct sales partner sees what a customer sees, and an
ASP/ASN partner sees products depending upon credentials.
SolVe Online
Select each panel for a description for using SolVe Online and generating a
procedure.
Home
The graphic shows the SolVe Online home page. The home page list the latest
advisories and shows the products. The product folders contain any procedure
routes you accessed.
Procedure Routes
SolVe Online retains a history of the procedure routes you accessed. The graphic
shows route. You can download the document from the route.
The example shows the routes to replacing an A200 node and setting up a 40U-P
rack.
Procedure routes.
New Route
When you click on the top level product on the home page, SolVe Online starts the
menu for a new route.
As an example, you want the procedure to replace a drive in an H400 node. Here,
you select Replacement Procedures.
Product Topics
From the Isilon platform page, open the Replacement Procedures topic. The
example shows both the PowerScale and the Gen 6 topics expanded. Select the
H400 topic.
Procedure Generator
The generator launches and choose the Replace Drive procedure, and then Next.
Generator Steps
Step 2 asks for the usage information and step 3 shows the procedure to generate.
Procedure Document
Generating the document may take a few minutes. Once complete, the PDF is
downloaded, and you return to the home page. On the home page, the new route is
listed.
Script:
Let us see how easy it is to generate a document. For this exercise, the SolVe
Desktop is installed, and you step through the process to get the Isilon Site
Preparation and Planning Guide. Shown is the shortcut on the desktop. Click the
shortcut to launch the tool.
License Agreement:
SolVe:
The SolVe Desktop launches. Now click the Isilon image. Here we get the
procedures. Expand Node Procedures.
Node Procedures:
Select X210. Select Site Planning and Preparation Guide, then select Next.
Selection:
We keep the destination as shown. Select Generate. Note the information and then
click OK.
Generated:
A word doc is shown that gives some general information. Scroll down. The image
of the document that is shown here is a link to that document. Click the image. The
PDF is opened.
The installation process starts before you arrive at the customer site. When
planning an installation, there are a few items that you must know before you
arrive.
IP Address Verification
The Customer Engineer (CE) typically works with the Implementation Specialist
(IS), who completes any custom configurations after the initial configuration is
done. CEs must verify that all internal networks configured for a PowerScale cluster
use IP version 4 addresses.
Use the pre-engagement questionnaire (PEQ) throughout the installation and initial
configuration process.
Make sure you discuss with the customer the plan for the spacing of the Gen 5
nodes or Gen 6 chassis5.
5 PowerScale hardware can take from 1U to 4U. All Gen 6 chassis are 4U and the
recommended maximum number that can be installed in a standard 19" rack is 12,
with the additional space needed for switches or cabling. Up to 10 chassis can fit in
a 42U rack. Keep in mind that Gen 6 nodes based on the High and Ultra compute
modules will contain high line only (240V) power supplies. When used in an
environment where only low-line power (120V) is available, two rack-mounted step-
up transformers will be required per Gen6 chassis, each of which will consume 2U
additional rack space per chassis and thus reduce the amount of space available
for nodes. Racks should be populated from bottom to top to improve rack stability
and decrease the risk of tipping.
Heavier nodes6
Pairs of step-up transformers7
Ethernet backend8
1U accelerator nodes and SRS gateway servers9
7If needed, should be placed adjacent to their associated chassis, one above the
chassis and the other below (to prevent cabling issues).
8 A new Gen 6 or Gen 6.5 cluster will have Ethernet switches for the BE network.
Gen 5 Node processor Gen 6 Node processor Gen 6.5 Node processor
hierarchy hierarchy hierarchy
- HD-Series - A2000
The node on which you initially run the Configuration Wizard is designated with a
node ID of 112. If selecting between two node types, such as an H500 and F800,
9 1U accelerator nodes and SRS gateway servers should go at the top of the rack.
10 Any gaps between nodes or switches require a filler panel to ensure proper
airflow.
11The cable management tray option requires 1U of rack space per node, so you
will need to plan for the additional space.
with equal OneFS versions, choose the node type with the greater processor power
as node 1.
The processor speeds of each Gen 5, Gen 6, and Gen 6.5 node type are shown in
descending order.
Serial Number
Serial Number
Movie:
The web version of this content contains a movie.
Script:
For A100 accelerator nodes, you must remove the face plate from the front of the
node. Press both latches in simultaneously until they click. Then remove the face
plate. Locate the black plastic tab on the upper left of the node and slide it out. The
serial number is printed on the blue label. When you are done, slide the tab back in
and replace the face plate.
Serial Number
The service tag is in the front of the system or may be on a sticker on the chassis
of the system. The mini Enterprise Service Tag (EST) is found on the back of the
system. Dell uses the information to route support calls to the appropriate
personnel.
1
2 3 4
5 6
1: Pull out the information tag in front of the system to view the Express Service
Code and Service Tag.
5: Service Tag.
Prepare Site
Module Objectives
You can use a standard ANSI/EIA RS310D 19-inch rack system14 in a PowerScale
cluster installation. The rack cabinet must have full earth ground to provide reliable
grounding. The rack should have dual power supplies with a power distribution that
allows load balancing and switches between the two if there is an outage. Use a
rack cabinet that has dual power distribution units, one on each side.
14The default PDU configuration includes 4 PDUs. You receive power cables to
support the maximum configurations and are likely to have extra cords as part of
your shipment.
Generate the appropriate guide using SolVe. The guide explains the tools that are
required, unpacking, and building the rack.
The next step is to attach power cables to the power distribution units on each side
of the cabinet.
15You must connect power cords to unpopulated P1 and P2 connectors on the four
power distribution units within the cabinet.
17The default PDU configuration includes 4 PDUs. You receive power cords to
support the maximum configurations and are likely to have extra cords as part of
your shipment.
40U-P cabinet PDUs do not include a power on/Off switch. Ensure that the circuit
breakers on each PDU are UP, and OFF, position until ready to supply AC power 18
to the unit.
18Equipment that is correctly installed within the rack cabinet is grounded through
the AC power cables and connectors. In general, supplemental grounding is not
required.
19If your site requires external grounding (for example, to a common grounding
network beneath the site floor), you can use the grounding lugs that are provided
on each of the lower power distribution panels of the cabinet.
The customer is responsible for bolting the rack to the floor. If a third-party rack is
used, be sure to secure the rack to the subfloor with a stabilizer bracket or some
other method. Use both front and side stabilizers when installing nodes in a rack
that is not bolted to the floor. Installing in an unbolted rack without these stabilizers
could cause the rack to tip over, potentially resulting in bodily injury. Use only
approved replacement parts and equipment.
1 2 3
5 4
1: If the rack is not bolted to the floor, it is recommended that you secure your rack
cabinet to the floor with stabilizing brackets or use anti-tip devices. The stabilizing
brackets are also important if you are installing or removing a server in the upper
half of the cabinet when the lower half is empty.
2: Sometimes, the trim strips are unattached from the cabinet, or was removed to
detach the shipping brackets. Once the cabinet is secured to the floor stabilization
brackets, install the lower trim strip to the bottom front of the cabinet.
3: The seismic bracket is installed on the front, back, and sides of the cabinet.
5: The anti-move bracket is installed on the front and back of the cabinet.
Once the rack is built, you are ready to unpack the components. First, check the
ShockWatch meter to ensure it has not been activated. If the ShockWatch meter
has been activated, be sure to contact support before continuing. Ensure that you
have sufficient clearance around the rack cabinet. If components are still packed
onto a pallet, you need approximately eight ft of clearance to position the pallet.
Open the shipping container and remove the PowerScale components.
More information20.
20To avoid personal injury or component damage, use two people to lift and move
a Gen 6 chassis. Inspect each component for any sign of damage. If the
components appear damaged in any way, notify PowerScale technical support. Do
not use a damaged component.
After you have built the rack, you will want to repack and return the shipping
material as it is reused for cost and environmental reasons. Leave the job site of
the customer as clean as or cleaner than it was when you arrived.
1: The first step is to lay each of the outer sleeves flat, then fold it lengthwise
before collapsing the three sections.
2: The second step is to stack the shipping brackets on the pallet. After that, place
ramp section B upside-down on the pallet, then stack section A, right-side up. Then
stack the two folded outer sleeves above the ramp sections.
3: The third step is to place ramp section B upside-down on the pallet, then stack
section A, right-side up. Then stack the two folded outer sleeves above the ramp
sections. Next, apply the two collar sections to either side of the pallet, hold them in
place with the package hood.
4: Finally band the unit with shipping straps and return it to Dell EMC.
Module Objectives
Sliding Rails
Once you have built your rack cabinet, which is unpacked the node and verified its
contents, and repacked the shipping material, you can start installing the hardware.
Click each tab to learn more.
Sliding Rails
The first step is to attach the sliding rails. A sliding rail system is used to attach the
node or chassis to the rack or cabinet and allow you easy access to the
components. The sliding rail kit is compatible with rack cabinets with the following
hole types:
Adjustable Rails
1U, 2U, and 4U rails for nodes and chassis adjust in length from 24 inches to 36
inches to accommodate a variety of cabinet depths. The rails are not left-specific or
right-specific and can be installed on either side of the rack. The video
demonstrates installing the rails for a Gen 6 chassis.
Movie:
The web version of this content contains a movie.
Pre-Threaded Holes
Movie:
The web version of this content contains a movie.
Considerations
Pre-Threaded Holes Script: To install the rail kit in a rack with prethreaded holes,
replace the preassembled step alignment pins in the ends of the outer rails with the
proper shoulder alignment pins. Adjust the outer slide rail rear slide bracket to fit
the depth of the rack cabinet, ensuring that the alignment pins protrude through the
rack mounting rails. Secure each outer rail to the rack cabinet using two 8, 32 x
0.75” knurled thumbscrews in the center holes of the slide rail.
Installing 1U Nodes
Before you begin, verify that the rail kit contains all the necessary components.
Each 1U slide rail consists of a slide bracket, an outer rail, an intermediate rail, and
an inner rail. Change the rail alignment posts if needed. Download and view the
SolVe Desktop procedure for the most recent instruction details.
1: Remove the inner rails from the rail assemblies on both rails.
3: Install both the slide rails in the rack. Each assembly contains a right and a left
slide rail.
4: You are now ready to install the node on the rails in the rack, pushing in until the
slide rails are locked. Remember, the enclosure is heavy and should be installed
into or removed from a rack by two people.
5: To further secure the rail assembly and server in the cabinet, insert and tighten a
small stabilizer screw directly behind each bezel latch.
Installing 2U Nodes
Movie:
Installing 2U Nodes Transcript: First, verify that the rail kit contains all the
necessary components. The 2U node rail kit includes:
Each slide rail consists of a slide bracket, an outer rail, an intermediate rail, and an
inner rail. To remove the inner rails from the rail assemblies, fully extend the slide
rails to the open and locked positions. Press the release button on the inner slide
rail and pull it from the intermediate slide rail until they separate.
Next, you attach the inner rails to the sides of the node. Align the large end of the
keyhole slots of the inner rail over the shoulder screws on the side of the node
chassis. It may be necessary to loosen the shoulder screws to slide the rails on.
Push the inner slide rail until it locks into place. Tighten the shoulder screws. Then,
attach the second slide rail on the other side the same way. To attach the rail to the
rack, first determine the rack cabinet that you are using. See the Rail Kit Installation
Guide for specific instructions for different types of rack cabinets. Here, we are
using a rack cabinet with 3/8” square holes. (Note: If using a rack with prethreaded
holes, replace the preassembled step alignment pins in the ends of the outer rails
with the proper shoulder alignment pins before proceeding to the next step.)
First, place the rear slide bracket into the back of the rack. Adjust the outer slide rail
to fit the depth of the rack cabinet, ensuring that the alignment pins protrude
through holes in the rack mounting rails and that the outer and rear slide brackets
are level. Attach to the same holes on both ends of the rack. Place the node
retaining bracket on the front of the right-side rail so that the middle hole covers the
top alignment pin. Then secure the bracket with one of the 8-32 x 0.75-inch knurled
thumbscrews. Secure all other ends of the rail to the rack by inserting the 8-32 x
0.75-inch knurled thumbscrews. Tighten the screws. Repeat the process for the
second rail.
Next, install the node in the rack. However, Do Not Continue until you confirm that
both rails are secured to the rack and all mounting screws are in place and
tightened. Failing to do so could result in injury or damage to the node, and to avoid
injury, use two people to lift and move the node.
Fully extend each intermediate slide rail until the rail is fully open and locked into
place. With a person on each side of the node, keep the node level with the slide
rails and align the ends of the inner slide rails with the ends of the intermediate
slide rails. Slide the inner slide rails on each side of the node into the intermediate
slide rails, and then retract the slide rails until the node is fully inserted in the rack.
There is an audible click when the node is secure. Confirm that the inner slide rails
attached to the node are inserted correctly and firmly secured to the intermediate
slide rails that are attached to the rack.
Secure the node to the rack cabinet using either the 6-32 x 0.25-inch knurled
chassis retaining screws or chassis retaining screws and retainer bracket
depending on the node type. Tighten the screws.
Installing 4U Nodes
Before you begin, unpack and verify that the rail kit contains all the necessary
components. Each slide rail consists of an inner rail, an outer rail, and an
intermediate rail. To attach the rail to the rack, first determine the rack cabinet that
you are using. See the Rail Kit Installation Guide for specific instructions for
different types of rack cabinets. The graphic shows the installation for a rack
cabinet with 3/8” square holes.
1: Place the rear slide bracket into the back of the rack. Adjust the outer slide rail to
fit the depth of the rack cabinet, ensuring that the alignment pins protrude through
holes in the rack mounting rails and that the outer and rear slide brackets are level
and attached to the same holes on both ends of the rack. Secure all other ends of
the rail to the rack by inserting the 8-32 x 0.75-inch knurled thumbscrews. Tighten
the screws. Repeat the process for the second rail.
2: Attach a clip-on nut retainer to the third hole above the slide rail on each side of
the rack. The next step is to install the node in the rack. However, do not continue
until you confirm that both rails are secured to the rack and all mounting screws are
in place and tightened. Failing to do so could result in injury or damage to the node.
Remember to always use two people to lift a node. If a 4U node is shipped with
hard drives already installed, it can weight over 200 lbs. Use of a mechanical lift is
recommended. To use the cable management tray, be sure to leave 1U space
under the node for the cable management tray. For ease of installation, install the
Cable Management Tray before installing the main chassis.
3: Once the node is installed in the rack, insert the drives into the node.
4: Once the drives are inserted into the node, insert the front panel.
Important: 4U nodes ship with the inner slide rails pre-installed on the
node. If using a rack with prethreaded holes, replace the
preassembled step alignment pins in the ends of the outer rails with
the proper shoulder alignment pins before proceeding to the next
step.
A chassis containing drives and nodes can weigh more than 200 pounds. You
should attach the chassis to a lift to install it in a rack. If a lift is not available, you
must remove all drive sleds and compute modules from the chassis before
attempting to lift it. Even when the chassis is empty, never attempt to lift and install
the chassis with fewer than two people.
1. The first step in installing the chassis is to align the chassis with the rails that
are attached to the rack.
2. Slide the first few inches of the back of the chassis onto the supporting ledge of
the rails.
3. Release the lift casters and carefully slide the chassis into the cabinet as far as
the lift allows.
4. Secure the lift casters on the floor.
5. Carefully push the node off the lift arms and into the rack. Ensure to leave the
lift under the chassis until the chassis is safely balanced and secured within the
cabinet.
6. Install two mounting screws at the top and bottom of each rail to secure the
chassis to the rack.
1: Back-end switches
• Provided by PowerScale.
• Switch ports face rear21.
• Connect each switch to separate power circuit.
• For multiple power connectors22, connect each cable to different power circuit.
• Each pair of switch supports a single cluster23.
21 Install switches into the rack with the ports facing to the rear of the rack. In a
redundant switch deployment, each switch should connect to a separate power
circuit for redundancy.
22If the switch has multiple power connectors, connect each of the switch’s power
cables to a different power circuit. Some switches have additional power
connectors to support optional redundant power supplies that may not be installed.
Ensure that each power cable is plugged into a connector that actually provides
power to the switch.
• Switches24:
• QLogic 72-port
• Mellanox 36-port
• High-speed, unmanaged fabric.
• Greater range of sizes.
• Identical functionality for Gen5 and Gen 6 nodes.
• Switches25:
• Celestica
• Arista
• High-speed, managed fabric.
• Limited monitoring.
23 Each pair of back-end switches supports a single cluster. Only one cluster is
allowed to be connected to the same pair of back-end switches. Connecting cables
to the switch is covered later.
24
InfiniBand switches are still supported for Gen 5 nodes and for mixed Gen 5 and
Gen 6 node clusters.
25 A pure Gen 6 cluster uses Ethernet switches (Celestica and Arista initially) for its
back-end network. When adding Gen 6 nodes to an existing Gen 5 cluster, the Gen
6 nodes connect to the InfiniBand switches. Follow the manufacturer’s instructions
for mounting the switches in the rack. If the customer is building a new cluster of all
Gen 6 nodes, the back-end switches must be Ethernet.
• No small switches:
• 24-port minimum26
• 252 max27 (at launch).
26 The new Ethernet back-end is only supported on the Gen 6 nodes. The minimum
allowable switch size is a 24-port switch.
Legend:
• S - 1:1 supported
• LC - Line card required
• BO - Breakout cable (1:4)
• AN - Autonegotiate to lower speed
• NS - Not supported
• XS - Supported, but not recommended
Dell Z9100 BO S XS BO XS S
Dell Z9264 BO S XS BO XS S
Dell S4148 S NS S AN NS NS
Dell S4112 S NS S AN NS NS
Celestica BO S BO BO/AN AN NS
D4040
Celestica S NS S AN NS NS
D2060
Installing Components
Module Objectives
Installation Guides
You can generate installation guides using SolVe. The example shows the first step
in generating an F200 node install guide.
You may need to install chassis components after the chassis is racked. For
instance, if no lift is available to install the chassis into the rack, then remove the
compute modules and drive sleds of the node. Once removed, install the chassis,
using two people, and then reinstall the compute modules and drive sleds of the
node. Select each tab to view a short demonstration on installing the drives, drive
sleds, and compute modules.
This short video demonstrates installing drives into a drive sled and installing the
sled into a chassis. See the student guide for the video transcript.
Movie:
The web version of this content contains a movie.
Compute Module
This video demonstrates how to install the compute module of a node into a
chassis. Refer the student guide for the video transcript.
Movie:
The web version of this content contains a movie.
Install Drive and Sled Script: For 3.5" drives, push the drive and paddle card
straight down into the empty connector. Insert the tabs on the drive sled cover into
the slots in the drive sled. For 2.5” drives, with the top cover at an angle, insert the
cover's three tabs into the slots on the sled. Next lower the open side of the cover
until it is flat against the drive sled. Slide the cover forward to lock it onto the sled.
Swing the blue sled cover latches back into place and press until you feel the
latches click. Ensure that the drive sled handle is open. With two hands, slide the
drive sled into the same bay you removed it from. Push the drive sled handle back
into the face of the sled to secure the drive sled in the bay. Swing the display
assembly back against the chassis until you feel it click into place. Align the front
bezel with the front of the chassis, then push until you feel the bezel snap into
place.
Install Compute Module Script: Keep the lever in the open position until the
compute module is pushed all the way into the bay. Support the node with both
hands and slide it into the node bay. Push the release lever in against the compute
module back panel. You can feel the lever pull the module into place in the bay. If
the lever does not pull the compute module into the bay, pull the lever back into the
open position. Next, ensure that the module is pushed all the way into the bay, then
push the lever in against the compute module again. Tighten the thumbscrew on
the release lever to secure the lever in place.
The F200 has four SAS SSDs numbered from 0 to 3 and uses no SAS expander.
The F600 drive bays 0 and 1 are empty due to the internal cabling layout of the
node. The F600 has eight NVMe SSDs that use bays 2 to 9.
Bay 0 and bay 1 - Bay 2 and Bay 4 and Bay 6 and Bay 8 and
NOT USED bay 3 bay 5 bay 7 bay 9
1
3
8
4
5 9
6 10
7
11
1: CPU2 is available on the F600. On the F200, this area is the location of the
NVDIMM battery.
5: Disk drives. The F200 uses SAS SSDs, and the F600 uses NVMe SSDs. Hard
drives are supplied in hot swappable hard drive carriers that fit in the hard drive
slots. You can check the status indicator on the carrier to identify a faulted hard
drive.
7: Backplane. Label each hard drive before removing from the node to ensure that
the drives are returned to the same drive bay.
8: Power supply units (PSU). The node needs one PSU for normal operation.
Remove and replace only one PSU at a time in a node that is powered on. PSUs
must be of the same type and have the same maximum output power.
9: Internal Dual SD Module (IDSDM). The F200 and F600 nodes use one of the
two micro SD slots. The micro SD shows up in OneFS as a USB device (da0).
isi_hwmon monitors IDSDM. IDSDM in OneFS is used for secondary backup for
PSI information and temporary storage for FW upgrade binaries.
10: Frontend NIC. The F600 uses PCIe slot 3 for front-end connectivity, and the
F200 uses the rack network daughter card (rNDC).
F200 rNDC.
11: Backend. The backend NIC is installed in slot 1 for both the F200 and F600.
The F200 has a 25 GbE NIC while the F600 uses a 100 GbE NIC.
Bezels in A-Series devices are accelerator node specific and may not appear as
shown. Bezels may include a keylock. All bezels include tabs on either side that
you press in to release the bezel and its latches. Push the ends, not the middle, of
the bezel, press the bezel onto the latch brackets until it snaps into place. The
video demonstrates installing the bezel on a Gen 6 chassis.
Movie:
Module Objectives
Once the system is racked and stacked, move on to connecting the node to the
internal and external networks.
Network
A PowerScale cluster uses separate internal and external networks for backend
and front-end connectivity.
Internal Network28
External Network29
28The internal network is dedicated for intracluster data and messages, providing
point-to-point connectivity with microsecond latency. Nodes connect to the internal
network of a cluster to communicate with other cluster nodes. Nodes connect to the
external client network to communicate with the rest of the world.
29 Gen 6 nodes can use either InfiniBand or Ethernet switches for back-end
connectivity. You cannot mix Ethernet and InfiniBand back-ends within one cluster.
If Gen 6 nodes are added to a cluster with Gen 5 nodes, the backend uses
InfiniBand switches only.
InfiniBand30
Hybrid Cables31
The graphic shows a Gen 6, four node cluster with an InfiniBand backend. There
are no special considerations when connecting to an Ethernet back-end. The two
steps are:
30All current Gen 5 and Gen 6 nodes and InfiniBand switches use 40 Gb/sec Quad
Data Rate (QDR) InfiniBand, which uses QSFP type cables. Legacy nodes and
switches use 10 Gb/sec Single Data Rate (SDR) or 20 Gb/sec Double Data Rate
(DDR) InfiniBand which uses CX4-type cables.
31 Hybrid cables are available to connect Legacy nodes to current IB switches and
the opposite way. Reference the PowerScale Supportability and Compatibility
Guide if working with legacy switches and nodes to determine the cable type
needed. Nodes with QDR InfiniBand adapters are incompatible with SDR
InfiniBand switches.
1. Connect an InfiniBand or Ethernet cable between the int-a port and the switch
(switch 1) for the Internal A network.
2. If the network topology supports a second internal network, connect the int-b
port to a separate network switch (switch 2) for the int-b network.
Gen 6.5 only supports Ethernet. Each Gen F200 and F600 node provides two ports
for backend connectivity. The PCIe slot 1 is used. F600 nodes use 100GbE or
25GbE ports. F200 nodes use 10GbE or 25GbE ports.
Breakout Cables
The 40 GbE and 100 GbE connections contain 4 individual lines of 10 GbE and 25
GbE.
Most switches support breaking out a QSFP port into four SFP ports using a 1:4
breakout cable.
The backend is managed automatically when the switch detects the cable type as a
breakout cable.
Each cable and the cabling should adhere to the cabling recommendations listed:
• Color code the cables to depict their type and function within the environment.
• Both ends of the cables should be clearly labeled depicting cable usage and
connection points in accordance to the labeling convention.
• Labeling should be done with correct type of label to sustain environmental
stress such as heat and humidity.
• Attached labels should not constrict the cable.
• The minimum bend radius of an InfiniBand cable is 20 cm.
• Avoid placing cables along sharp edges.
• Do not zip and constrict cables.
• Avoid using zip ties for the cable bundling. Instead use Velcro ties.
• Bundle different cable types separately and keep a good distance between
power and InfiniBand cables.
• Ensure that your cabling support is well anchored and able to withstand cable
weight.
• Restrict cables from dangling between rack spaces.
• Retain a sufficient loop on the cable in order to facilitate maintenance work, for
example, when sliding out a node.
• Align cable placements in order to facilitate future addition of cables along the
path.
External networks connect the cluster to the outside world. Subnets are used in
external networks to manage connections more efficiently. Specifying external
network subnets depends on the topology of the network32. The Ethernet cable
connects the node to the external network of cluster so the node can communicate
with external clients. Use an Ethernet cable to connect the ext-1 port on the node to
the switch for the external network.
Gen 6 nodes33 have two ports, ext-1 and ext-2, as the graphic depicts.
Gen 6.5
Protocols: NFS, SMB, S3,
2x10 GbE, 2x25 GbE, 2x 40
HTTP, FTP, HDFS, SWIFT
GbE, 2x100 GbE
32In a basic network topology, one in which all nodes reside on a single subnet,
only one external subnet is required.
33
The Gen 6 node can connect to one or more external Ethernet networks using 10
GbE or 40 GbE connections.
The graphic shows a closer look at the external and internal connectivity.
Slot 1 is used for backend communication on both the F200 and F600.
Slot 3 is used for the F600 2x 25 GbE or 2x 100 GbE front-end network
connections.
The rack network daughter card (rNDC) is used for the F200 2x 25 GbE front-end
network connections.
PCIe slot 1 - used for all BE PCIe slot 3 - used for F600
communication FE
FE FE
The graphic shows the R640 and does not represent the F200 and F600 PCIe and rNDC
configuration.
Connecting Power
Nodes and chassis contain redundant power supplies to ensure that the node
remains powered in case a power supply fails. Connect the power supply cables to
the node power connectors, and then connect each power supply cable to a
separate power source.
• Gen 6
o Compute has single power
supply.34
o Node-pair provides redundant
power.35
o Nodes automatically powers
on.36
• Gen 6.5
The video demonstrates connecting An LED shows the current state of each power
the power cord to a Gen 6 node. supply.
Movie:
The web version of this content contains a movie.
35If the power supply fails, the node receives power from the power supply of the
adjacent node.
36Gen 6 nodes automatically power on when they are connected to power. When
setting up new nodes, if you are not yet ready for them to be powered on, do not
connect the power cables.
Connecting Power Cord Transcript: Connect the power cable to the power
supply. Rotate the metal bail down over the power cable to hold the cord in place.
Configuration Manager
For initial configuration, access the CLI by establishing a serial connection to the
node designated as node 1. The serial console gives you serial access when you
cannot or do not want to use the network. Other reasons for accessing using a
serial connection may be for troubleshooting, site rules, a network outage, so on.
Serial Port37
37 The serial port is usually a male DB9 connector. This port is called the service
port. Connect a serial null modem cable between a serial port of a local client, such
as a laptop, and the node service port. Connect to the node designated as node 1.
As most laptops today no longer have serial ports, you might need to use a USB-
to-serial converter. On the local client, launch a serial terminal emulator.
• Data bits = 8
• Parity = none
• Stop bits = 1
• Flow control = hardware
Course Summary
Course Summary
Now that you have completed this course, you should be able to:
→ Plan an installation.
→ Prepare the site.
→ Install the node in the rack.
→ Install components.
→ Cable the node to a network.
Electrostatic Discharge
Electrostatic Discharge is a major cause of damage to electronic components and
potentially dangerous to the installer. To avoid ESD damage, review ESD
procedures before arriving at the customer site and adhere to the precautions when
onsite.
Antistatic Packaging:
Leave components in
antistatic packaging until
time to install.
PARTICIPANT GUIDE
PARTICIPANT GUIDE
Table of Contents
PowerScale Implementation
PowerScale Implementation
PowerScale Implementation
PowerScale Implementation
PowerScale Implementation
The graphic shows the PowerScale Solutions Expert certification track. You can
leverage the Dell Technologies Proven Professional program to realize your full
potential. A combination of technology-focused and role-based training and exams
to cover concepts and principles as well as the full range of Dell Technologies'
hardware, software, and solutions. You can accelerate your career and your
organization’s capabilities.
PowerScale Solutions
(C) - Classroom
PowerScale Implementation
Prerequisite Skills
To understand the content and successfully complete this course, a student must
have a suitable knowledge base or skill set. The student must have an
understanding of:
• Current PowerScale hardware portfolio and the OneFS operating system
• PowerScale Concepts
PowerScale Implementation
Course Objectives
PowerScale Implementation
PowerScale Implementation
Module Objectives
PowerScale Implementation
Implementation Phases
1: The first phase is planning. During this phase, gather all required information for
the installation. A Pre- Engagement Questionnaire (PEQ) document outlines all the
information to gather. The customer completes the document and returns it to Dell
EMC before the installation date. The document ensures that the project team has
the required information for the design, installation, and configuration of the
PowerScale cluster.
2: The second phase of the project is the implementation and testing phase. During
this phase, the physical installation and initial configuration of the PowerScale
cluster is carried out and remote support configuration is done. Now, the testing
and verification of the cluster is performed to ensure that the cluster is ready for
use. In addition, the cluster is ready for advanced software installation and
configuration.
PowerScale Implementation
The PowerScale PEQ is the replacement for the Configuration Guide. The PEQ
defines the Professional Services project installation parameters and facilitates the
communication between the responsible resources. For this scenario, the required
information is in the Hardware tab.
PowerScale Implementation
To configure a PowerScale cluster for the first time, access the CLI by establishing
a serial connection to the node designated as node 1. The serial console gives
serial access when you cannot use the network. Other reasons for accessing using
a serial connection may be for troubleshooting, site rules, a network outage, and so
on. The serial port is usually a male DB9 connector. This port is called the service
port1.
1 Connect a serial null modem cable between a local system, such as a laptop, and
the service port on the node designated as node 1. As most laptops today no
longer have serial ports, you must use a USB-to-serial converter, available at many
online retailers. On the local system, launch a serial terminal emulator such as
mRemoteNG.
PowerScale Implementation
Creating a Cluster
Movie:
Link:
https://edutube.emc.com/Player.aspx?vno=zyK3csP1f1h6F0cke3Ivog==&autoplay
=true
PowerScale Implementation
PowerScale Implementation
Module Objectives
PowerScale Implementation
Movie:
Script:
The first step before adding a node to a cluster is to verify which OneFS version is
running. Methods to check that the version include the front panel display, CLI, and
WebUI. Let us start with the CLI. As shown here, I am logged into the cluster. Run
the isi version command. Next, we can use the WebUI. Here I am logging in. The
version is shown in the upper right side of the browser. If the new node is
compatible with the OneFS version on the existing cluster you are joining, there is
no need to reimage the new node or nodes. The node downloads the correct
version of OneFS from the cluster during the join process and reimages
automatically. In this demonstration, a new cluster was created.
There are several ways to add a node to a cluster. You can use the node front
panel display, the Configuration Wizard, or the WebUI. We show adding a node
using the Configuration Wizard and then another node using the WebUI.
Here I have established a serial connection the node I want to add. At the
Configuration Wizard prompt, type 2 and press ENTER to join the node to an
existing cluster. Type the number of the cluster you want to add the node to, and
press ENTER. In this example, I have only one cluster that is indexed as 1. The
node obtains the proper version of OneFS from the cluster if its version of OneFS
differs from that of the cluster. If a reimage is needed, allow the operation to
complete before attempting to add additional nodes.
If the addition is successful, the node boots to a login prompt. Log in as root and
run isi status to verify cluster membership and health. If there are additional nodes
to add, wait 5 minutes between node additions to allow the cluster time to stabilize
and incorporate the newly joined nodes.
Next let us add the third node to the cluster using the WebUI. Here I am logging in
as root. On the Dashboard page, we can see that there are only two nodes in this
cluster. Remember that Gen 5 must have a minimum of three nodes and Gen 6
needs four nodes to form a cluster. Navigate to Cluster Management, Hardware
Configuration. Here we see one available node. Select the Add node in the Actions
PowerScale Implementation
We saw how to check the OneFS version and how to add nodes to a cluster using
the Configuration Wizard and the WebUI. This concludes the demo.
PowerScale Implementation
Movie:
The video shows how to remove a node from the cluster using the WebUI.
2Before smart failing multiple nodes, consider whether the cluster has enough free
space to maintain protection on the data in the cluster. Removing too many nodes
causes the cluster to lose quorum. Data is restriped to other nodes in the cluster as
part of the process.
3If the node is still connected to the back-end network when the process
completes, it automatically initiates a reformat - otherwise, it can be reformatted
manually.
4The use cases are to remove older nodes during a life-cycle technology refresh.
First, to move a node from one cluster to another cluster. Second, if PowerScale
Support has identified a critical hardware failure.
PowerScale Implementation
Script:
We start by logging into the WebUI. For this demonstration, I am smartfailing node
ID 3 shown on the dashboard. Navigate to Cluster Management, Hardware
Configuration. Select the Remove Nodes tab. Select Node 3 and then Submit.
Next, I will Confirm the smartfail. Back on the Dashboard, we can see the node
status in the process of smartfailing. After several minutes, the Dashboard updates
showing the node is no longer a member of the cluster.
PowerScale Implementation
Reformat a Node
Movie:
Use case5.
Script:
5 The use case for reformatting a node is when resetting or repurposing nodes, for
example, the case of testing cluster nodes. This performs the reset quickly so the
nodes can be used for other testing. Another use case example is when destroying
a cluster to reset it to prepare a SyncIQ failback for SmartLock or WORM
directories. WORM directories cannot be overwritten, they can only be removed by
reformatting the cluster. This also applies to enterprise SmartLock directories and
compliance SmartLock directories. If SyncIQ has not been configured, these
considerations do not apply.
PowerScale Implementation
I am connected to the node using the console port. You can also connect using an
SSH session to the node. I begin by logging in as root. Here I use the
isi_reformat_node --reboot command and typing yes to understand I am sure I
want to reformat. Typing yes again. The node reboots and as seen here, the
Configuration Wizard displays.
PowerScale Implementation
Reimage a Node
Reimaging a node performs the same node reset functionality as reformatting the
node. The difference is that the OneFS operating system is reinstalled during the
process. Connect to the reimaging node over the console port or an SSH session.
Reimaging a node can be done using a USB flash drive or by copying the OneFS
install file locally on the node. If reimaging using a USB flash drive containing the
OneFS image, a limited number of USB flash drives are supported.
Copy the OneFS install file to the node and run the isi_reimage command to
reimage a node without using a flash drive.
PowerScale Implementation
The use case is to install a new OneFS image on the cluster. In certain
circumstances, nodes with a OneFS version that is installed from the factory may
differ from the OneFS version on the cluster. Do not always upgrade or downgrade
automatically when the node is added to the cluster. In this case, the nodes must
be reimaged to match the cluster OneFS version.
SDCZ36-16G 16 GB
SDCZ36-32G 32 GB
DT101G2/32GBZ 32 GB
PowerScale Implementation
PowerScale Implementation
Module Objectives
PowerScale Implementation
After completing the installation, update the install database. Browse to the Dell
EMC Product Registration and Install Base Maintenance service portal at the URL
shown in the browser. Select the Product Registration and Install Base
Maintenance option and then select the IB Status Change option to open the form.
Next complete the form and then submit. Shown in the graphic are sources for
PowerScale technical support.
PowerScale Implementation
Connectivity Tests
The final step of the initial configuration process is to perform connectivity tests.
The four tests - WebUI, HTTP, DNS, and client are performed from a client that is
connected over Ethernet to the cluster.
The first test, PowerScale WebUI test, can be done by launching the browser6
to the cluster from the client.
The other two tests, DNS, and client are performed post implementation7.
6Depending on the browser, you may be prompted with security information. Log in
over the administration interface to complete the test.
7Once the cluster is configured for DNS, you can test by pinging the cluster by
name. This is also true for the HTTP test, use a browser and the URL of the Fully
Qualified Domain Name and path. For example,
https://clustername.company.com/IFS. The client test can check access and
authorization. From a Windows client, map a drive to a share on the cluster. Then
drag and drop a file from the client to the mapped drive. Ensure that the transfer
completes successfully. From an NFS client, mount the export directory that is
presented from the cluster. Then copy a file from the client to the NFS mount point
and ensure that the transfer completes successfully.
PowerScale Implementation
PowerScale Implementation
Module Objectives
PowerScale Implementation
Movie:
Script:
This demonstration shows how to create a compliance mode cluster. You can
configure a node or cluster in Compliance mode by selecting the SmartLock
Compliance mode option in the Configuration Wizard. Here I am connected to the
node designated as node 1 with the node that is booted to the Configuration
Wizard. A compliance mode cluster must be created during the initial cluster
configuration process. Once you set a cluster to Compliance mode, you have to
reformat the node in order to return it to a standard cluster mode. If there is data on
the node, returning to standard cluster mode without data loss requires a
Professional Services engagement.
The first step is to select option 4, to reboot the node into SmartLock Compliance
mode. Next type yes to verify the option and understand that root access to the
node will be disabled. I press enter, and the node reboots.
The node boots back to the Configuration Wizard. Note the new wording on option
4, Disable SmartLock Compliance mode, indicating the node is in compliance
mode. This is the last opportunity to back out of compliance mode configuration
without reformatting the node.
Here I enter 1, to create a cluster. We are asked for a SmartLock license. I enter
that.
Accept the EULA, and then we set the password for compadmin and admin. The
rest of the configuration is no different than the standard cluster mode installs.
Remember to see the Configuration Guide for the information to enter. For this
PowerScale Implementation
At the login prompt, let us go ahead and login as compadmin. I run the isi status
command.
As part of this demonstration lets,’s sees how to add a node to a compliant cluster.
I am connected serially to the node I want to join. Once again, select option 4 to
reboot in SmartLock Compliance mode. Once the boot completes, select option 2
in the Configuration Wizard.
PowerScale Implementation
EMC Secure Remote Support (or ESRS) is a two-way, secure, IP-based customer
service remote support system that allows 24x7 remote monitoring of EMC
products. ESRS replaced SupportIQ as the PowerScale remote support system.
So, let us see what the ESRS environment looks like. I start with the customer side.
Here I show a heterogeneous Dell EMC environment as ESRS is the tool that
many Dell EMC platforms use. I add the virtual edition of ESRS. This is the
gateway and at least one ESRS gateway server must be installed and configured.
The gateway server acts as the single point of entry and exit for IP-based remote
support activities and monitoring notifications. ESRS can be configured for
redundancy with more than one ESRS instance installed, allowing reports through
ESRS if hardware or partial data environment failure.
The environment has some firewalls. A few notes about the optional policy
manager. You need the Policy Manager if fulfilling requirements for authentication,
authorization, and auditing. By implementing the Policy Manager, you can enable
monitoring on a node-by-node basis, allow or deny remote support sessions, and
PowerScale Implementation
Next I draw out the other half of the ESRS, the Dell EMC environment. Now, ESRS
functions as a communications Broker agent between the managed devices, the
Policy Manager, and Dell EMC. Dell EMC does not establish inbound network
communications to the systems. This is a security measure, benefiting customers
in secure sites while permitting limited, controlled outbound communications.
When the ESRS server retrieves a remote access request from the EMC
Enterprise, the policies that are configured on the Policy Manager access the
request and the ESRS server enforces these requests.
Communications between the customer site and Dell EMC support flow over an
encrypted HTTPS connection, which means that sensitive information does not
traverse the Internet unprotected.
On the Dell EMC support side, only authorized support representatives have
access to the customer systems or their information at all. All communication with
Dell EMC initiates from ESRS on port 443 or 8443 outbound from the customer site
to Dell EMC support services.
Enabling ESRS:
PowerScale Implementation
Before enabling and configuring ESRS, you must ensure the OneFS software has
a signed license and an ESRS gateway server is installed and configured. The IP
address pools that handle gateway connections must exist in system and must
belong to a subnet under groupnet0, which is the default system groupnet. Shown
is the OneFS 8.0 WebUI, Cluster Management, General Settings, ESRS page.
Later OneFS versions may prompt to update the license. Select Enable ESRS
Gateway Support, complete the fields, and Save Changes.
ESRS Documentation:
These documents and others about ESRS can be found on the EMC Online
Support site (support.emc.com) and you can use the SolVe Desktop to download
documents. Remember to always check for the latest updates and Release Notes.
PowerScale Implementation
Hadoop Overview
4 5
1
2 6
8 Though Hadoop has emerged as a tool of choice for big data analytics, there are
also reasons to use it in typical enterprise environments. To analyze existing data
for improving processes and performance depending on your business model.
PowerScale Implementation
2: Hadoop has two core components: HDFS and MapReduce. HDFS is a scalable
file system that is used in the Hadoop cluster. MapReduce is the compute
algorithm that analyzes the data and collects the answers from the query.
3: NameNode:
• Holds the location information for every file in the cluster. The file system metes
data.
Secondary NameNode:
5: Map Step:
• The primary node takes the input, divides it into smaller subproblems, and
distributes them to worker nodes. The worker node processes the smaller
problem and passes the answer back to its primary node.
Reduce Step:
• The primary node then collects the answers to all the subproblems and
combines them in some way to form the output. The answer to the problem it
was originally trying to solve.
6: Task Tracker:
• A node in the cluster that accepts tasks - Map, Reduce, and Shuffle operations
from a Job Tracker
PowerScale Implementation
Scale out NAS platform of PowerScale provides native support for the HDFS
protocol.
1: The NameNode on the PowerScale cluster gives the solution a complete and
automated failover process. If the node running as the NameNode fails, another
PowerScale node immediately picks up the function of the NameNode. No data or
metadata would be lost since the distributed nature of PowerScale spreads the
metadata across the cluster. There is no downtime if a failure occurs and most
importantly there is no need for administrative intervention to failover the
NameNode.
PowerScale Implementation
Configuring Hadoop
There are the five major steps to configure HDFS for a Hadoop solution.
1: First, you have to validate the OneFS version and license activation. This can be
verified from the WebUI by going to the Cluster Management, Licensing page.
Consult the PEQ for the license key.
2: Next you prepare the cluster which includes configuring DNS, SmartConnect,
and Access Zones to allow for the Hadoop cluster to connect to the PowerScale
OneFS cluster.
3: When preparing your Hadoop distribution (that is, Cloudera), do NOT select
HDFS service. PowerScale is used as storage and the native HDFS service is no
longer needed.
4: When configuring your Hadoop distribution (that is, Cloudera), refer to the
PowerScale OneFS installation guide to configure the Hadoop cluster elements, to
use the PowerScale cluster as its HDFS clients, and then tweak the HDFS
PowerScale Implementation
configuration of OneFS. Leave the PowerScale role as default during the role
assignment process.
PowerScale Implementation
The terms “upgrade” and “update” are often used interchangeably. They describe
the process of changing to a newer version of the OneFS operating system (OS) or
changing to newer node or disk firmware versions. However, the terms “upgrade”
and “update” do not mean the same thing.
Upgrade is used when the OneFS operating system version changes. Upgrades
apply code fixes, new operating system features, and functionality. Often times an
upgrade changes the underlying structure or operations of the OneFS file system. It
marks a strategic step forward for the product and for the functionality that it offers
to production clusters.
Update is used when new firmware versions are installed on the nodes or on the
disk drives. Firmware updates are applied to fix (patch) firmware code issues and
to standardize all the nodes or disks onto the same firmware9 code level.
9 Often when receiving a new node or a new drive, the firmware that is shipped on
the new hardware is a different version than the firmware running on the production
cluster. Newer firmware versions add or change cluster functionality, sometimes
they are used to apply code fixes and add support for new hardware.
PowerScale Implementation
Customers may ask about the OneFS version and an understanding of the version
identifiers. OneFS versions are broken down in the significance of the changes that
are made to the release. A change to the first and second numbers of the release
code identifies the major upgrades.
• First digit changes indicate major changes and potential core modifications to
the underlying operating system and data structures.
• The second digit indicates significant changes or enhancements, but usually do
not involve the core structure. Typically, significant features changes or
additions are in these releases.
• The last two numbers indicate minor releases.
− The third digit indicates smaller or less impactful feature enhancements. The
support of new nodes can be in the third digit release.
− The fourth digit denotes minor functionality and is used for issue fixes.
PowerScale Implementation
Upgrades
Rolling Upgrade
Simultaneous Upgrade
Simultaneous upgrades are faster than rolling upgrades, but reboot all nodes at the
same time, thus incurring an interruption in data access.
PowerScale Implementation
OneFS 8.2.0 enables the pausing and resuming of a OneFS upgrade. Pause and
resume are useful when the maintenance window ends. The upgrade can be
paused, and then resumed in a later window. The commands are isi upgrade
pause and isi upgrade resume.
Pausing is not immediate. Upgrade is in a pausing state until the current upgrading
node completes. Other nodes do not upgrade until the upgrade is resumed.
Pausing state can be viewed with isi upgrade view, or
isi_upgrade_status, or viewing the pause file data.
PowerScale Implementation
Below are the supported upgrade paths to OneFS 8.1 and OneFS 8.2. Upgrades
from versions earlier than OneFS 8.0 are simultaneous. Upgrading from OneFS 8.0
to OneFS 8.1 can be a rolling or simultaneous upgrade.
PowerScale Implementation
PowerScale Implementation
A standard part in each node installation or change activity is to bring the node up
to a minimum level. If required or requested to correct any significant issues,
OneFS patches can be applied. For the nodes, it includes bringing the node
firmware and drive firmware up to the latest version.
More information10.
Patches
A patch is a software update that addresses known issues before those updates
are incorporated into a maintenance release (MR)11.
10 For OneFS versions (OneFS 7.1.1 and above), it may include installing the drive
support package (DSP) to automate and simplify drive firmware updates. For all
drives to be on the same firmware version within each node. If possible, all nodes
should use the same firmware versions. When nodes run the latest version and all
drives are on the same version, customers experience fewer issues and there is a
reduction in future downtime or node outages.
PowerScale Implementation
There are two types of patches, standard patch and rollup patch.
• A standard patch addresses one or more known issues for a major, minor, or
MR release of OneFS. Some patches contain minor enhancements or
additional logging functionality that can help Dell EMC Technical Support
troubleshoot issues with the cluster.
• Rollup patches address multiple issues that are related to one component of
OneFS functionality, such as SMB. It might also contain fixes from previous,
standard patches that addressed issues that are related to that component.
Firmware
Firmware is the software code that makes hardware components12 contained within
the node work. Updates are used to manage the code versions13 applied to each of
12Included are components such as the drives, the motherboard bios, the disk
controller, the InfiniBand adapter and the network adapters. The firmware is stored
on, and runs inside, the individual hardware components.
PowerScale Implementation
the hardware components. Firmware updates are performed across the cluster to
standardize all node components on the same version to eliminate potential
incompatibility issues and simplify support.
The isi devices drive firmware command to view and update firmware.
More Information14.
13Code versions are updated to reflect code fixes or code enhancements for
additional stability.
14 Firmware updates are recommended as part of the process for new cluster
installations. The commands and process used to perform firmware updates are
different from the commands and process used for upgrades. When upgrading
PowerScale Implementation
Disable to add
drives manually
Enabled by default
The graphic shows the WebUI Cluster management, Automatic Replacement Recognition page.
When replacing a drive in a node, OneFS automatically formats the drive, adds it to
the cluster, and updates the drive firmware. The new drive firmware matches the
current drive support package that is installed on the cluster. The drive firmware is
not updated for the entire cluster, only for the new drive. If you prefer to format and
add drives manually, disable Automatic Replacement Recognition.
PowerScale Implementation
isi_upgrade_helper
PowerScale Implementation
isi_upgrade_helper Default
isi_upgrade_helper
1 - Rolling
PowerScale Implementation
Release Notes
Release notes contain installation instructions for drive and node firmware
upgrades and patches. Reading them carefully is an essential step in determining
whether your business benefits from the new features and functionality 15 of a new
OneFS version.
The release notes help you determine if the upgrade has an impact and if so,
provide a solution if needed. The contents of the release notes are similar across
releases - they provide release and upgrade information, new features,
modifications and enhancements, resolved issues, known issues and installation
instructions.
PowerScale Implementation
PowerScale Implementation
Course Summary
PowerScale Implementation
Course Summary
Now that you have completed this course, you should be able to:
→ Create a new cluster.
→ Join a node to an existing cluster.
→ Complete a configuration.
→ Complete other configuration options.
PowerScale Implementation
PARTICIPANT GUIDE
PARTICIPANT GUIDE
Table of Contents
The graphic shows the PowerScale Solutions Expert certification track. You can
leverage the Dell Technologies Proven Professional program to realize your full
potential. A combination of technology-focused and role-based training and exams
to cover concepts and principles as well as the full range of Dell Technologies'
hardware, software, and solutions. You can accelerate your career and your
organization’s capabilities.
PowerScale Solutions
(C) - Classroom
Prerequisite Skills
To understand the content and successfully complete this course, a student must
have a suitable knowledge base or skill set. The student must have an
understanding of:
• Current PowerScale hardware portfolio and the OneFS operating system
• PowerScale Concepts
Course Objectives
Module Objectives
The graphic shows a few basic reminders that are common to all hardware
maintenance procedures.
If you encounter any difficulties while performing this task, immediately contact Dell
EMC Technical Support.
3 6
4 8
2 7
1: Customer Replaceable Units (CRUs) are removed without shutting down the
node. This means you can accomplish CRU replacements and CRU hardware
upgrades while the node is powered up as long as the correct procedure is
followed. Field Replaceable Units (FRUs), on the other hand, require the node to
be powered off. If you must power off a node, always shut it down properly as
described in the replacement guide.
2: On Generation 6 nodes (Gen 6), before disconnecting any cables, ensure that
the compute module's Do Not Remove LED is off. When the LED is white or On,
this indicates the node’s journal is still active. The Do Not Remove LED is on the
right side of the compute module and looks like a symbol of a hand with a slash
through it. Do not disconnect any cables until this LED is off.
3: On Generation 5 nodes (Gen 5), never power off a node by pressing the power
button or removing both power cables unless you are expressly directed to do so
by Dell EMC Technical Support.
4:
Use the Solve Desktop tool to get the most recent, full instructions for the
procedure. These instructions are frequently updated based on feedback from the
field, so ensure to consult the instruction documents prior to every engagement,
even if you have already performed the service requested previously.
SolVe Desktop has been revised and updated to SolVe Online. It is a knowledge
management-led standard procedure for DELL-EMC field, service partners, and
customers. Click here for an overview on SolVe Desktop or SolVe Online.
6: Save the packaging from the replacement part. Use this packaging to return the
failed part to Dell EMC. A return label is included with the replacement part.
7: If the customer and/or Dell EMC technical support request Failure Analysis on
the replaced part, be sure to attach a filled out FA label to the return box, and
complete an FA request ticket in the WWFA system. Provide the FA ticket number
to your Support contact and/or add it to the SR in a comment.
8: After all work is complete, partner personnel should submit the Partner
Notification Form (PNF) to allow Dell EMC Technical Support to update the install
database. Dell EMC personnel should update the install database directly by going
to the Dell EMC Business Services website. In the Post Sales area, click Install
Base Group, complete, and submit the form.
Electrostatic Discharge
Anti-static Packaging:
Leave components in
anti-static packaging until
time to install.
Cold Serviceable
Blue Handles
Hot Serviceable
Terracotta Handles
Preparing a Node
Module Objectives
We’ll start off by covering how to safely power down a node. Remember, on a Gen
5 node, never power down a node by pressing the power button unless explicitly
instructed to do so by Dell EMC Technical Support. To power down a node, first
connect to the cluster. This can be done using SSH or the serial port. If using a
terminal emulator utility with a serial port connection, here are the settings to use:
In this exercise, we will shut down node 3. The next step after connecting to the
cluster is to get the IP address of the node to shut down. Here we are connected
and logged in to the cluster on node 1.
Click in the box and type isi status -q to get the node’s IP address.
Log on to the node that you want to shut down and type the command to shut the
node down: shutdown -p now.
To check that the node is powered down, we are connected to another node in the
cluster. Click in the box and type isi status -q again.
The shutdown node has a status of D--- (Down) in the command output.
Note:
• Gen 6 nodes do not support serial flow control - the flow control
setting should be set to 'None' when connecting to Gen 6
hardware.
• isi network interfaces list -v command can also
be used to get the node's IP address.
• isi config command can also be used to shut down a
node.
• It is recommended to run isi_flush command prior to
performing any shutdown to flush the cache.
Module Objectives
FRU
You can watch the videos of replacement procedures in the next few slides.
Gen 6
Movie:
Link:
https://edutube.emc.com/vlearning/launch/cPBJf1qfSHOzdb2Q|@$@|pA1wg==/vid
eodetails=false,comments=false,launch=yes
Node: Gen 6
Part 1
When facing the back of the chassis the compute modules are labeled right to left,
one to four as shown. Because compute modules are installed in pairs that are
called “node-pairs”, the minimum cluster size has increased from three to four
nodes and additional nodes must be added in node-pairs. The graphic shows that
the Node-pairs are either the left half or right half of chassis.
Part 2
Link:
https://edutube.emc.com/Player.aspx?vno=S3kFK4UC82Qmnc2PUec1iA==&autop
lay=true
• Transfer the internal components from the failed compute module to the
replacement unit, except for the battery
Remove the replacement node from the shipping package and inspect it for any
sign of damage. Notify Dell EMC Isilon Technical Support if the node appears
damaged in any way. Do not install a damaged node. Do not discard the shipping
container and packaging. You'll use them to return the failed node to Dell EMC.
Power down the node by following the instructions in the replacement guide.
Label the network cables connected to the back of the node to ensure that you can
reconnect them correctly. Before you disconnect any cables from a node, make
sure Do Not Remove LED is not lit. When the Do Not Remove LED is off
disconnect all cables from the back of the node. If there are transceivers connected
to your network cables, remove them from the node. You might see LEDs on inside
the node, even after you have removed the power cord. That is because the node
next to it is supplying redundant power. On the back of the chassis, loosen the
orange thumbscrew that secures the node release lever. To eject the node from the
node bay, pull the orange handle away from the node. Slowly pull the node out of
the bay. Support the node at front and back with both hands as you remove it from
the bay. Place the node on an ESD protected work surface next to the replacement
node.
Position the node with the fans facing you. The blue release handle should be
under the fans. Place the heel of your hand on the gray connectors above the fans
and grab the blue release handle with the fingertips of your other hand. Make sure
you are not pressing down on the top of the node with the heel of your hand as that
will keep the node lid from popping up when you pull the release handle. Pull on
the blue release handle to lift the lid up from the node. You will feel an initial pop as
the blue release handle pulls away from the node. Pull on the release handle until
you feel a second pop to raise the lid up off the node. Lift the lid straight up off the
node and place it next to the body of the node. Repeat for the replacement node.
Inside the left side of the node body, just behind the fans, locate the blue touch
point label. Place the thumb of your left hand on the blue touch point and press the
side of the node away from the fans. The metal tab that holds the fans in place will
flex away from the fans so you can remove them. Slide the fans straight up out of
the node with your right hand. Repeat for the replacement node
Locate the two blue tabs for removing the HBA riser. There's a sliding tab at the
back of the riser, and a fixed tab at the front. Complete the following three steps at
the same time. To free the back end of the riser, push the sliding tab in the
direction of the arrow on the tab. To free the front end, pull the riser away from the
locking pin on the side of the chassis with the fixed tab. Lift up on the tabs to
unseat the riser and pull it straight up out of the node.
Remove both the internal and external NICs from the HBA riser. When you are
looking down the length of the HBA riser, with the battery pack close to you, the
internal NIC is on your left, closest to the bottom of the riser. Remove the retaining
screw that secures the internal NIC to the chassis and set it aside. Pull the NIC
straight up out of its slot. Make a note that this is the internal NIC. Repeat for the
external NIC.
Disconnect the battery pack and remove it from the HBA riser. Press in on the
locking tab and disconnect the battery cable. Push in on the retaining tabs on the
bottom edge of the riser and lift up to free one side of the battery pack. Roll the
battery pack away from the riser to free the other side of the pack and remove it.
Unsnap and open the black retaining tab at the end of the M.2 vault card. Lift the
free end of the card at an angle and pull the card away from the connector.
After you remove the M.2 vault card, re-install the battery pack in the HBA riser
from the failed node. The replacement node already contains a battery. Hook the
two battery pack feet closest to the battery cable into the slots on the riser. Roll the
battery pack down until it is flat against the M.2 vault card, and then push in on the
retaining tabs until they click into the slots.
Slide the HBA riser into the node and secure it in place. Align the metal tab next to
the sliding tab on the riser with the slot on the node chassis. Slide the riser
downward until you seat the riser in the chassis. When you push the riser down to
seat it, you will see the sliding tab click forward and back as it secures the riser in
place. Make sure that the locking pin next to the fixed tab at the front of the riser
aligns with the locking slot in the chassis. The locking pin might sit away from the
side of the chassis. You can pinch the side of the chassis and the riser together to
make sure that the locking pin aligns with the slot on the chassis. When you install
the fans, the side of the fan module will hold the locking pin in place.
Remove the HBA riser from the replacement node using the same technique as
before. Remove the battery pack using the same technique as before.
Unsnap and open the black retaining tab at the end of the M.2 vault card. Insert the
connecting end of the M.2 vault card at an angle into the connector on the new
HBA riser card. Lower the other end of the M.2 vault until the card lies flat against
the HBA riser. Snap the retaining tab closed over the end of the card. Re-install the
battery using the same technique as before.
Locate the slot where you will install the internal NIC. Align the bottom of the card
with the appropriate slot and push the NIC into the slot. Secure the card to the
chassis using the retaining screw. Repeat for the external NIC. If you're installing a
10Gb NIC, the card is shorter than the internal NIC. You must install it in the middle
slot, right next to the internal NIC. If you're installing a 40Gb NIC, the card looks
just like the internal NIC. You must install it in the far-right slot, closest to the blue
HBA riser release tab.
Transfer all DIMMs from the failed node to the replacement node. In the
replacement node press down on the DIMM retaining tabs. Do the same in the
failed node for the first DIMM. Pull the DIMM straight up to remove it from the slot.
Make note of the slot from which the DIMM is removed. Transfer it into the
corresponding slot in the lid of the replacement node. Align the DIMM with the slot
and press down until the retaining tabs snap into place. Push on the retaining tabs
to make sure they are closed. Repeat for all remaining DIMMs.
With the label on top of the fans facing you, insert the rails on either side of the fans
into the slots on the sides of the node. Press down on the fans until you feel them
click into place. Repeat for the replacement node.
Make sure the blue release handle below the fans is pulled out completely. Place
the node lid onto the body of the node. You can use the cutouts on the side of the
lid to align the lid with the node body. Make sure that the lid is not in contact with
the HBA riser or any other internal components, otherwise you might damage
something when you secure the lid. Apply gentle pressure to the top of the lid with
one hand as you push in the blue release handle with the other hand. You'll feel the
lid pull down onto the node as you push in the release handle. If you do not feel the
lid pull down onto the node, pull the release handle back out and make sure that
the lid is properly aligned with the node body. Brace one hand against the back of
the node and push the blue release handle all the way in to secure the lid to the
node body. Repeat for the replacement node.
Keep the lever in the open position until the node is pushed all the way into the
bay. Support the node with both hands and slide it into the node bay. Push the
release lever in against the node back panel. You can feel the lever pull the node
into place in the bay. If you do not feel the lever pull the node into the bay, pull the
lever back into the open position, make sure that the node is pushed all the way
into the node bay, then push the lever in against the node again. Tighten the
thumbscrew on the release lever to secure the lever in place. Locate the labels on
the network cables and connect them to the correct ports on the back of the node.
Locate the power cable and connect it to the back of the node. Drop the metal bale
down over the power cord to secure the connector in place. The node will
automatically power up when you connect the power cable.
• Gather logs
If you encounter any difficulties while performing this task, contact Dell EMC
Technical Support.
Gen 6.5
Movie:
Link:
https://edutube.emc.com/vlearning/launch/mxpAovnlxV1ZXK1wL4Mr1g==/videodet
ails=false,comments=false,launch=yes
Module Objectives
Node Compatibility
1First, Gen 6 nodes can be in a cluster with Gen 5 and Gen 4 nodes. Gen 5 nodes
and earlier are not compatible with a Gen 6 cluster using an Ethernet back-end.
This means you cannot add Gen 5 nodes to a Gen 6 cluster that uses an Ethernet
back-end, but you can have Gen 6 and Gen 5 nodes using the same InfiniBand
backend.
2 It enables you to transition slowly to the new hardware over time without a forklift
upgrade by allowing you to add one node at a time to an existing node pool. This is
more cost effective than adding the node minimum to start a new node pool with
the all new hardware. When a customer has grown the new node counts to
sufficient quantities, node compatibility can be disabled on an individual node pool.
3 Enabling SSD compatibility allows customers to replace older, smaller SSDs with
new, larger SSDs to allow more L3 cache space. This lets customers better utilize
storage resources. Every node in the pool must be the same model or of the same
SSD: Gen 6
Movie:
Link:
https://edutube.emc.com/Player.aspx?vno=Ti76c637o7LUEAJj9GMayw==&autopla
y=true
Script: If there is more than one cache SSD installed, review the cluster event
associated with the failed SSD to determine which SSD to replace. Refer to the
replacement guide for more information. Press up on the orange tab to free the
bottom of the protective cover from the node, then swing the cover up and remove
it. On the face of the SSD, press up on the orange release button to release the
SSD handle. Rotate the SSD handle downward until it is perpendicular to the
compute module. Pull the SSD from the node.
Install the new cache SSD into the back of the node. If both SSD bays are empty,
install the SSD into the bay on the right. Make sure the SSD handle is completely
open and insert the SSD into the empty drive bay. Rotate the SSD handle upward
to seat the SSD and lock it in place. Place the upper tab of the SSD cover into the
slot above the SSDs. Swing the bottom of the SSD cover down and press it up into
the back of the node until it clears the catches and rests securely in place.
series or family. The node pool must have the same number of SSDs per node in
every node if the OneFS version is prior to OneFS 8.0.
If you encounter any difficulties while performing this task, contact Dell EMC
Technical Support.
Regardless of the node type, each compute module slot pairs with five drive sled
bays. Depending on the length of the chassis and type of drive, this means that
each node can have up to thirty drives, or as few as fifteen with every sled in place.
Every node needs a consistent set of sled types, and drive types in each sled,
meaning you cannot mix-and-match different drives within a sled or different sleds
in node slots. There are three types of drive sleds. For 3.5" drives, there are long
and short sleds, and for 2.5" drives there is a short sled that contains up to six
drives. The 3.5" drives come with a paddle card that connects the drive into the
sled, while the 2.5" drives connect directly into the sled. The 3.5" drives fit into a
sled without paddle cards, but there will be no connection to the sled without the
paddle card.
Drives: Gen 6
Part 1
Internal to the 2.5" sled, there are individual fault lights for each drive. The yellow
LED associated with each drive is visible through holes in the top cover. A
supercapacitor can keep one light lit for around 10 minutes while the sled is out of
the chassis, but if more than one light is lit (indicating multiple drive failures) the lit
time is correspondingly reduced.
In the 3.5" drive sleds, the yellow drive fault LEDs are on the paddle cards, and
they are visible through the cover of the drive sled so that you can see which drive,
if any, needs replacement. The graphic shows the 3.5” short drive sled, the 3.5”
long sled has four LED viewing locations.
Part 2
Movie:
The web version of this content contains a movie.
Link:
https://edutube.emc.com/Player.aspx?vno=hfBe54RKpBcgPpEvb2/dXQ==&autopla
y=true
Part 1
The graphic shows the lights and their information for the drive sleds. All twenty
sleds can be individually serviced. Do not remove more than one sled per node at a
time on running nodes. The typical procedure is to go to a chassis where a fault
has been detected, inspect the sleds to see which one shows a fault light, press the
service request button, wait until the LED stops blinking and goes dark, then
remove the sled and replace the drive. Replace the sled. The node automatically
detects and configures the replacement drive.
The service request button informs the node that the sled will be removed, and the
node prepares for this by moving key boot information from drives on that sled,
suspending the drives in the sled from the cluster file system, and then spinning
them down. This is to maximize survivability in the event of further failures, and to
prevent cluster file system issues that are caused by multiple drives becoming
temporarily unavailable.
Power/Activity
Sled Fault
Part 2
Movie:
The web version of this content contains a movie.
Link:
https://edutube.emc.com/Player.aspx?vno=JvXxvcHiwtqgODunE/TzlA==&autoplay
=true
Note: If the suspend button is pressed and drives are detected, the
node attempts to rediscover the sled and rejoin its drives after 1 hour.
If the suspend button is pressed and drives are not detected or the
sled is still removed, the node automatically smartfail the drives after
15 minutes.
Movie:
Link:
https://edutube.emc.com/Player.aspx?vno=LW/ZI6kmMaqKY54dpE/lOw==&autopl
ay=true
Script: Gather logs by following the instructions in the replacement guide. Press
both latches of the front bezel simultaneously to release it. Align the front bezel with
the front of the chassis, then push until you feel the bezel snap into place.
If you encounter any difficulties while performing this task, contact Dell EMC
Technical Support.
Movie:
Link:
https://edutube.emc.com/Player.aspx?vno=oveIZf3k48xr4biiu/hZXg==&autoplay=tr
ue
Script: Lift the metal bale to free the power cord. Disconnect the power cord from
the power supply. You may see LEDs on inside the compute module, even after
you have removed the power cord. That is because the node next to it is supplying
redundant power. Press the orange retaining tab upward and pull the black handle
to slide the power supply out of the node.
Slide the new power supply unit into the open bay in the back of the node until you
feel the unit click into place. Connect the power cord to the power supply. Rotate
the metal bale down over the power cord to hold the cord in place.
Follow the instructions in the guide to complete the replacement procedure. If you
encounter any difficulties while performing this task, contact Dell EMC Technical
Support.
Gen 6.5
Movie:
Link:
https://edutube.emc.com/vlearning/launch/aYXb4cgLbhio9HDLNhC0Xg==/videodet
ails=false,comments=false,launch=yes
• To update the drive firmware on nodes without bootflash drives, download and
install the latest drive firmware package.
• Power cycling drives during a firmware update might return unexpected results.
As a best practice, do not restart or power off nodes when the drive firmware is
being updated in a cluster.
• To update the drive firmware for your entire cluster, run the following command:
# isi devices drive list (--node-lnn node-lnn-number)
• Note that you must wait for the current upgrade operation to complete before
initiating another.
• To confirm that a node has finished updating, run the following command: #
isi devices -d <node-number>. A drive that is still updating displays a
status of FWUPDATE.
• OneFS updates drive sequentially.
Module Objectives
Click each tab to learn how to generate the procedure through SolVe Online.
Step 1
Step 2
Note that the procedure is still generated even though you do not enter any
information.
Step 3
• Click GENERATE.
Step 4
• You can find the generated procedure in My Content tab in SolVe Online.
Step 5
• You will also receive a mail with the link for the procedure.
Course Summary
Course Summary
Now that you have completed this course, you should be able to:
→ Explain hardware maintenance procedures.
→ Prepare a node.
→ Replace Field Replaceable Units (FRUs).
→ Replace Customer Replaceable Units (CRUs).