NetBackup8301 Hadoop Guide
NetBackup8301 Hadoop Guide
Hadoop Administrator's
Guide
Release 8.3.0.1
Veritas Hadoop Administrator's Guide
Last updated: 2020-09-03
Legal Notice
Copyright © 2020 Veritas Technologies LLC. All rights reserved.
Veritas, the Veritas Logo, and NetBackup are trademarks or registered trademarks of Veritas
Technologies LLC or its affiliates in the U.S. and other countries. Other names may be
trademarks of their respective owners.
This product may contain third-party software for which Veritas is required to provide attribution
to the third party (“Third-party Programs”). Some of the Third-party Programs are available
under open source or free software licenses. The License Agreement accompanying the
Software does not alter any rights or obligations you may have under those open source or
free software licenses. Refer to the Third-party Legal Notices document accompanying this
Veritas product or available at:
https://www.veritas.com/about/legal/license-agreements
The product described in this document is distributed under licenses restricting its use, copying,
distribution, and decompilation/reverse engineering. No part of this document may be
reproduced in any form by any means without prior written authorization of Veritas Technologies
LLC and its licensors, if any.
The Licensed Software and Documentation are deemed to be commercial computer software
as defined in FAR 12.212 and subject to restricted rights as defined in FAR Section 52.227-19
"Commercial Computer Software - Restricted Rights" and DFARS 227.7202, et seq.
"Commercial Computer Software and Commercial Computer Software Documentation," as
applicable, and any successor regulations, whether delivered by Veritas as on premises or
hosted services. Any use, modification, reproduction release, performance, display or disclosure
of the Licensed Software and Documentation by the U.S. Government shall be solely in
accordance with the terms of this Agreement.
http://www.veritas.com
Technical Support
Technical Support maintains support centers globally. All support services will be delivered
in accordance with your support agreement and the then-current enterprise technical support
policies. For information about our support offerings and how to contact Technical Support,
visit our website:
https://www.veritas.com/support
You can manage your Veritas account information at the following URL:
https://my.veritas.com
If you have questions regarding an existing support agreement, please email the support
agreement administration team for your region as follows:
Japan [email protected]
Documentation
Make sure that you have the current version of the documentation. Each document displays
the date of the last update on page 2. The latest documentation is available on the Veritas
website:
https://sort.veritas.com/documents
Documentation feedback
Your feedback is important to us. Suggest improvements or report errors or omissions to the
documentation. Include the document title, document version, chapter title, and section title
of the text on which you are reporting. Send feedback to:
You can also see documentation information or ask a question on the Veritas community site:
http://www.veritas.com/community/
https://sort.veritas.com/data/support/SORT_Data_Sheet.pdf
Contents
Index .................................................................................................................... 63
Chapter 1
Introduction
This chapter includes the following topics:
■ Limitations
Hadoop plug-in Is
Hadoop cluster deployed on all the backup
hosts BigData policy
NameNode
Application_Type=hadoop
Backup Host 1
DataNode 1
Master server
DataNode 2
Backup Host 2
Media server
DataNode 3
Storage
Backup Host 3
DataNode n Parallel Streams
...
3 Discovery of 4 Workload
Backup job
workload for backup discovery file 1
is triggered.
2
6 Child
DataNode 2 job 2
Backup Host 2
DataNode 3 3
6 Child
job 3
Backup Host 3 Storage
DataNode n 7 Data is backed up in = Workload
parallel streams n distribution files
Hadoop Cluster
(Snapshot Enabled)
3. During discovery, the first backup host connects with the NameNode and
performs a discovery to get details of data that needs to be backed up.
4. A workload discovery file is created on the backup host. The workload discovery
file contains the details of the data that needs to be backed up from the different
DataNodes.
5. The backup host uses the workload discovery file and decides how the workload
is distributed amongst the backup hosts. Workload distribution files are created
for each backup host.
6. Individual child jobs are executed for each backup host. As specified in the
workload distribution files, data is backed up.
7. Data blocks are streamed simultaneously from different DataNodes to multiple
backup hosts.
The compound backup job is not completed until all the child jobs are completed.
After the child jobs are completed, NetBackup cleans all the snapshots from the
NameNode. Only after the cleanup activity is completed, the compound backup job
is completed.
See “About backing up a Hadoop cluster” on page 42.
2
Backup host connects
with NameNode
NameNode
1 Master server
Restore job
is triggered
DataNode 1
Backup host
DataNode 2 4
Objects are restored on Storage
Hadoop Cluster the associated datanodes
(Snapshot Enabled) 3
Restore
Starts
Terminology Definition
Discovery job When a backup job is executed, first a discovery job is created. The
discovery job communicates with the NameNode and gathers
information of the block that needs to be backed up and the associated
DataNodes. At the end of the discovery, the job populates a workload
discovery file that NetBackup then uses to distribute the workload
amongst the backup hosts.
Child job For backup, a separate child job is created for each backup host to
transfer data to the storage media. A child job can transfer data blocks
from multiple DataNodes.
Workload discovery During discovery, when the backup host communicates with the
file NameNode, a workload discovery file is created. The file contains
information about the data blocks to be backed up and the associated
DataNodes.
Parallel streams The NetBackup parallel streaming framework allows data blocks from
multiple DataNodes to be backed up using multiple backup hosts
simultaneously.
Backup host The backup host acts as a proxy client. All the backup and restore
operations are executed through the backup host.
Terminology Definition
Fail-over NameNode In a high-availability scenario, the NameNodes other than the primary
NameNode that are updated in the hadoop.conf file are referred
as fail-over NameNodes.
Terminology Definition
Limitations
Review the following limitations before you deploy the Hadoop plug-in:
■ Only RHEL and SUSE platforms are supported for Hadoop clusters and backup
hosts.
Introduction 14
Limitations
Task Reference
Pre-requisites and See “Pre-requisites for the Hadoop plug-in” on page 16.
requirements
Task Reference
Best practices See “Best practices for deploying the Hadoop plug-in” on page 17.
Verifying the
deployment
Task Reference
Task Reference
Configuring the See “Configuring the Hadoop plug-in using the Hadoop configuration
Hadoop plug-in file” on page 23.
using the Hadoop
See “Configuring NetBackup for a highly-available Hadoop cluster”
configuration file
on page 24.
Configuring the See “Configuration for a Hadoop cluster that uses Kerberos” on page 34.
backup hosts for
Hadoop clusters
that use Kerberos
Configuring See “Configuring NetBackup policies for Hadoop plug-in” on page 35.
NetBackup policies
for Hadoop plug-in
For more information on how to create a policy, see See “Creating a BigData backup
policy” on page 35.
To add a backup host
1 In the Backup Selections tab, click New and add the backup host in the
following format:
Backup_Host=<IP_address or hostname>
For more information on how to create a policy, See “Creating a BigData backup
policy” on page 35.
Alternatively, you can also add a backup host using the following command:
For Windows:
<Install_Path>\NetBackup\bin\admincmd\bpplinclude PolicyName -add
"Backup_Host=IP_address or hostname"
For UNIX:
/usr/openv/netbackup/bin/admincmd/bpplinclude PolicyName -add
"Backup_Host=IP_address or hostname"
For more information, See “Using NetBackup Command Line Interface (CLI)
to create a BigData policy for Hadoop clusters ” on page 38.
2 As a best practice, add the entries of all the NameNodes and DataNodes to
the/etc/hosts file on all the backup hosts. You must add the host name in
FQDN format.
OR
Add the appropriate DNS entries in the /etc/resolv.conf file.
Configuring NetBackup for Hadoop 21
Managing backup hosts
For UNIX:
/usr/openv/netbackup/bin/admincmd/bpplinclude PolicyName -delete
'Backup_Host=IP_address or hostname'
■ For Windows
The directory path to the command:
<Install_Path>\NetBackup\bin\admincmd\bpsetconfig
bpsetconfig -h masterserver
bpsetconfig> APP_PROXY_SERVER = clientname1.domain.org
bpsetconfig> APP_PROXY_SERVER = clientname2.domain.org
bpsetconfig>
Windows systems: <ctl-Z>
■ Hostname and port of the NameNode must be same as you have specified with
the http address parameter in the core-site.xml of the Hadoop cluster.
■ For password, provide any random value. For example, Hadoop.
Configuring NetBackup for Hadoop 23
Configuring the Hadoop plug-in using the Hadoop configuration file
Note: You must not provide a blank value for any of the parameters, or the backup
job fails.
Ensure that you configure all the required parameters to run the backup and restore
operations successfully.
Note: For non-HA environment, the fail-over parameters are not required.
{
"application_servers":
{
"hostname_of_the_primary_namenode":
{
"failover_namenodes":
[
{
"hostname":"hostname_of_failover_namenode",
"port":port_of_the_failover_namenode
}
],
"port":port_of_the_primary_namenode
}
},
"number_of_threads":number_of_threads
}
■ Hostname and port of the NameNode must be same as you have specified with
the http address parameter in the core-site.xml of the Hadoop cluster.
■ User name of the primary and fail-over NameNode must be same.
■ Do not provide a blank value for any of the parameters, or the backup job fails.
Configuring NetBackup for Hadoop 26
Configuring the Hadoop plug-in using the Hadoop configuration file
{
"application_servers":
{
"hostname_of_primary_namenode1":
{
"failover_namenodes":
[
{
"hostname": "hostname_of_failover_namenode1",
"port": port_of_failover_namenode1
}
],
"port":port_of_primary_namenode1
}
}
}
Configuring NetBackup for Hadoop 27
Configuring the Hadoop plug-in using the Hadoop configuration file
2 If you have multiple Hadoop clusters, use the same hadoop.conf file to update
the details. For example,
{
"application_servers":
{
"hostname_of_primary_namenode1":
{
"failover_namenodes":
[
{
"hostname": "hostname_of_failover_namenode1",
"port": port_of_failover_namenode1
}
],
"port"::port_of_primary_namenode1
},
"hostname_of_primary_namenode2":
{
"failover_namenodes":
[
{
"hostname": "hostname_of_failover_namenode2",
"port": port_of_failover_namenode2
}
],
"port":port_of_primary_namenode2
}
}
}
3 Copy this file to the following location on all the backup hosts:
/usr/openv/netbackup/
{
"application_servers": {
"hostname_of_namenode1":{
"port":port_of_namenode1
}
}
2 Copy this file to the following location on all the backup hosts:
/usr/openv/netbackup/
{
"number_of_threads": number_of_threads
}
{
"application_servers":
{
"hostname_of_namenode1":
{
"use_ssl":true
}
}
}
{
"application_servers":
{
"primary.host.com":
{
"use_ssl":true,
"failover_namenodes":
[
{
"hostname":"secondary.host.com",
"use_ssl":true,
"port":11111
}
]
}
}
}
ECA_TRUST_STORE_PATH Specifies the file path to the certificate bundle file that contains
all trusted root CA certificates.
If you have not configured the option, add all the required
Hadoop server CA certificates to the trust store and set the
option.
If you have not configured the option, add all the required
CRLs to the CRL cache and then set the option.
Set this value to YES when you have set the use_ssl as
true in the hadoop.conf file. The single value is applicable
to all Hadoop clusters when use_ssl is set to true.
HADOOP_CRL_CHECK Lets you validate the revocation status of the Hadoop server
certificate against the CRLs.
Usage Description
CRLs in the CRL cache are periodically updated with the CRLs in the directory that
is specified for ECA_CRL_PATH based on the ECA_CRL_PATH_SYNC_HOURS option.
If the ECA_CRL_CHECK or HADOOP_CRL_CHECK option is not set to DISABLE (or 0) and
the ECA_CRL_PATH option is not specified, NetBackup downloads the CRLs from
the URLs that are specified in the CRL distribution point (CDP) and uses them to
verify revocation status of the peer host's certificate.
Note: For validating the revocation status of a virtualization server certificate, the
VIRTUALIZATION_CRL_CHECK option is used.
Usage Description
Usage Description
How to use Use the nbgetconfig and the nbsetconfig commands to view,
add, or change the option.
HADOOP_SECURE_CONNECT_ENABLED = YES
Usage Description
Usage Description
ECA_TRUST_STORE_PATH=/tmp/cacert.pem
ECA_CRL_PATH=/tmp/backuphostdirectory
HADOOP_SECURE_CONNECT_ENABLED=YES/NO
HADOOP_CRL_CHECK=DISABLE / LEAF / CHAIN
■ Acquire the keytab file and copy it to a secure location on the backup host.
■ Ensure that the keytab has the required principal.
■ Manually update the krb5.conf file with the appropriate KDC server and realm
details.
Note: Hostname and port of the NameNode must be same as you have specified
with the http address parameter in the core-site.xml of the Hadoop cluster.
For more information on how to create a BigData policy, See “Creating a BigData
backup policy” on page 35.
Note: The directory or folder specified for backup selection while defining
BigData Policy with Application_Type=hadoop must not contain space or
comma in their names.
For more information on using NetBackup for big data applications, refer to the
Veritas NetBackup documentation page.
4 View the details about the new policy using the -L option.
bpplinfo policyname -L
For UNIX:
bpplinclude PolicyName -add 'Application_Type=hadoop'
7 Specify the backup host on which you want the backup operations to be
performed for Hadoop.
For Windows:
bpplinclude PolicyName -add "Backup_Host=IP_address or hostname"
For UNIX:
bpplinclude PolicyName -add 'Backup_Host=IP_address or hostname'
Note: The backup host must be a Linux computer. The backup host can be a
NetBackup client or a media server or a master server.
8 Specify the Hadoop directory or folder name that you want to backup.
For Windows:
bpplinclude PolicyName -add "/hdfsfoldername"
For UNIX:
bpplinclude PolicyName -add '/hdfsfoldername'
Note: Directory or folder used for backup selection while defining BigData
Policy with Application_Type=hadoop must not contain space or comma in
their names.
9 Modify and update the policy storage type for BigData policy.
bpplinfo PolicyName -residence STUName -modify
10 Specify the IP address or the host name of the NameNode for adding the client
details.
For Windows:
bpplclients PolicyName -M "MasterServerName" -add
"HadoopServerNameNode" "Linux" "RedHat"
For UNIX:
bpplclients PolicyName -M 'MasterServerName' -add
'HadoopServerNameNode' 'Linux' 'RedHat'
Configuring NetBackup for Hadoop 40
Disaster recovery of a Hadoop cluster
11 Assign a schedule for the created BigData policy as per your requirements.
bpplsched PolicyName -add Schedule_Name -cal 0 -rl 0 -st
sched_type -window 0 0
Task Description
After the Hadoop cluster and nodes are up, Perform the following tasks:
prepare the cluster for operations with
Update firewall settings so that the backup
NetBackup.
hosts can communicate with the Hadoop
cluster.
The backup hosts use the hadoop.conf file With this release, the following plug-in
to save the configuration settings of the settings can be configured
Hadoop plug-in. You need to create separate
■ See “Configuring NetBackup for a
file for each backup host and copy it to
highly-available Hadoop cluster”
/usr/openv/netbackup/. You need to
on page 24.
create the hadoop.conf file in JSON format.
■ See “Configuring number of threads for
backup hosts” on page 28.
Update the BigData policy with the original See “Configuring NetBackup policies for
NameNode name. Hadoop plug-in” on page 35.
Chapter 4
Performing backups and
restores of Hadoop
This chapter includes the following topics:
Task Reference
(Optional) See “Pre-requisite for running backup and restore operations for a
Complete the Hadoop cluster with Kerberos authentication” on page 43.
pre-requisite for
Kerberos
Best practices See “Best practices for backing up a Hadoop cluster” on page 43.
Performing backups and restores of Hadoop 43
About backing up a Hadoop cluster
Task Reference
Troubleshooting For discovery and cleanup related logs, review the following log file on
tips the first backup host that triggered the discovery.
/usr/openv/netbackup/logs/nbaapidiscv
For data transfer related logs, search for corresponding backup host
(using the hostname) in the log files on the master server.
Note: During the backup and restore operations, the TGT must be valid. Thus,
specify the TGT validity accordingly or renew it when required during the operation.
For example,
kinit -k -t /usr/openv/netbackup/nbusers/hdfs_mykeytabfile.keytab
[email protected]
■ Ensure that the local time on the HDFS nodes and the backup host are
synchronized with the NTP server.
■ Ensure that you have valid certificates for a Hadoop cluster that is enabled with
SSL (HTTPS).
Task Reference
Complete the See “Pre-requisite for running backup and restore operations for a
pre-requisites for Hadoop cluster with Kerberos authentication” on page 43.
Kerberos
Restoring Hadoop ■ See “Using the Restore Wizard to restore Hadoop data on the same
data on the same Hadoop cluster” on page 46.
NameNode or ■ See “Using the bprestore command to restore Hadoop data on the
Hadoop cluster same Hadoop cluster” on page 47.
Restoring Hadoop See “Restoring Hadoop data on an alternate Hadoop cluster” on page 49.
data to an
alternate
NameNode or
Hadoop cluster
Best practices See “Best practices for restoring a Hadoop cluster” on page 45.
Troubleshooting See “Troubleshooting restore issues for Hadoop data” on page 59.
tips
■ Ensure that the backup host contains a valid CRL that is not expired.
From the Policy type for restores list, choose BigData as the policy type
for restore.
Click Ok.
6 Go to the Backup History and select the backup images that you want to
restore.
7 In the Directory Structure pane, expand the Directory.
All the subsequent files and folders under the directory are displayed in the
Contents of Selected Directory pane.
8 In the Contents of Selected Directory pane, select the check box for the
Hadoop files that you want to restore.
9 Click Restore.
10 In the Restore Marked Files dialog box, select the destination for restore as
per your requirement.
■ Select Restore everything to its original location if you want to restore
your files to the same location where you performed your backup.
■ Select Restore everything to a different location if you want to restore
your files to a location which is not the same as your backup location.
Where,
-S master_server
Specifies a file (listfile) that contains a list of files to be restored and can be
used instead of the file names option. In listfile, list each file path must be on
a separate line.
-L progress_log
Specifies the name of whitelisted file path in which to write progress information.
-t 44
Where,
-S master_server
Specifies a file (listfile) that contains a list of files to be restored and can be
used instead of the file names option. In listfile, list each file path must be on
a separate line.
-L progress_log
Specifies the name of whitelisted file path in which to write progress information.
-t 44
Specifies the name of a file with name changes for alternate-path restores.
Change the /<source_folder_path> to /<destination_folder_path>
Note: NetBackup supports redirected restores only using the Command Line
Interface (CLI).
Performing backups and restores of Hadoop 50
About restoring a Hadoop cluster
Note: Make sure that you have added the credentials for the alternate NameNode
or Hadoop cluster in NetBackup master server and also completed the Whitelisting
tasks on NetBackup master server. For more information about how to add Hadoop
credentials in NetBackup and whitlelisting procedures, See “Adding Hadoop
credentials in NetBackup” on page 22. See “Whitelisting a NetBackup client on
NetBackup master server” on page 21.
Performing backups and restores of Hadoop 51
About restoring a Hadoop cluster
Parameter Value
Specifies a file (listfile) that contains a list of files to be restored and can be
used instead of the file names option. In listfile, list each file path must be on
a separate line.
-L progress_log
Specifies the name of whitelisted file path in which to write progress information.
-t 44
Specifies the name of a file with name changes for alternate-path restores.
Use the following form for entries in the rename file:
change backup_filepath to restore_filepath
ALT_APPLICATION_SERVER=<Application Server Name>
Note: Ensure that you have whitelisted all the file paths such as
<rename_file_path>, <progress_log_path> that are already not included as
a part of NetBackup install path.
Chapter 5
Troubleshooting
This chapter includes the following topics:
Area References
General logging See “About NetBackup for Hadoop debug logging” on page 54.
and debugging
Backup issues See “Troubleshooting backup issues for Hadoop data” on page 54.
Restore issues See “Troubleshooting restore issues for Hadoop data” on page 59.
To avoid issues See “Best practices for deploying the Hadoop plug-in” on page 17.
also review the
See “Best practices for backing up a Hadoop cluster” on page 43.
best practices
See “Best practices for restoring a Hadoop cluster” on page 45.
Troubleshooting 54
About NetBackup for Hadoop debug logging
■ See “Extended attributes (xattrs) and Access Control Lists (ACLs) are not backed
up or restored for Hadoop” on page 56.
■ See “Backup operation fails with error 6654” on page 57.
■ See “Backup operation fails with bpbrm error 8857” on page 57.
■ See “Backup operation fails with error 6617” on page 57.
■ See “Backup operation fails with error 6616” on page 57.
Extended attributes (xattrs) and Access Control Lists (ACLs) are not
backed up or restored for Hadoop
Extended attributes allow user applications to associate additional metadata with
a file or directory in Hadoop. By default, this is enabled on Hadoop Distributed File
System (HDFS).
Access Control Lists provide a way to set different permissions for specific named
users or named groups, in addition to the standard permissions. By default, this is
disabled on HDFS.
Hadoop plug-ins do not capture extended attributes or Access Control Lists (ACLs)
of an object during backup and hence these are not set on the restored files or
folders.
Workaround:
If the extended attributes are set on any of the files or directories that is backed up
using the BigData policy with Application_Type = hadoop, then, you have to
explicitly set the extended attributes on the restored data.
Extended attributes can be set using the Hadoop shell commands such as fs
-getfattr and hadoop fs -setfattr.
If the Access Control Lists (ACLs) are enabled and set on any of the files or
directories that is backed up using the BigData policy with Application_Type =
hadoop, then, you have to explicitly set the ACLs on the restored data.
ACLs can be set using the Hadoop shell commands such as hadoop fs -getfacl
and hadoop fs -setfacl.
Troubleshooting 57
Troubleshooting backup issues for Hadoop data
Verify that the backup host has valid Ticket Granting Ticket (TGT) in case of
Kerberos enabled Hadoop cluster.
Workaround:
Renew the TGT.
Workaround:
Verify the hadoop.conf file to ensure that blank values or incorrect syntax is not
used with the parameter values.
/data/1
/data/2
Workaround
To view the available data that can be restored from an incremental backup image,
select the related full backup images along with the incremental backup images.
Extended attributes (xattrs) and Access Control Lists (ACLs) are not
backed up or restored for Hadoop
For more information about this issue, See “Extended attributes (xattrs) and Access
Control Lists (ACLs) are not backed up or restored for Hadoop” on page 56.
Restore operation fails when Hadoop plug-in files are missing on the
backup host
When a restore job is triggered on a backup host which does not have Hadoop
plug-in files installed, the restore operation fails with the following error:
{
"application_servers":
{
"primary.host.com":
{
"use_ssl":true
"failover_namenodes":
[
{
"hostname":"secondary.host.com",
"use_ssl":true
"port":11111
}
],
"port":11111
Troubleshooting 62
Troubleshooting restore issues for Hadoop data
}
},
"number_of_threads":5
}
Index
A N
Adding NetBackup
backup host 19 debug logging 54
server and client requirements 16
B NetBackup Appliance
backup host 22
Backup 44
Hadoop 42
backup 9 O
BigData policy overview
Command Line Interface 38 backup 7
NetBackup Administration Console 36 configuration 7
Policies utility 37 deployment 7
Policy Configuration Wizard 36 installation 7
restore 7
C
compatibility P
supported operating system 16 parallel streaming framework 7
Creating policies
BigData backup policy 35 configuring 35
Preparing
D Hadoop 16
disaster recovery 40
R
H Removing
backup host 19
Hadoop credentials
Restore
adding 22
bprestore command 47
Hadoop 44
K restore 10
Kerberos Restoring
post installation 34 alternate NameNode 49
kerberos Hadoop cluster 46
backup 43
restore 43
T
terms 11
L Troubleshoot
License backup 54
Hadoop 16 troubleshooting
Limitations 13 restore 60
Index 64
W
Whitelisting
backuphost 21