optimize-storage-compute-platforms-in-vmware-vsphere-environments-best-practices-guide
optimize-storage-compute-platforms-in-vmware-vsphere-environments-best-practices-guide
MK-SL-145-04
September 2024
© 2024 Hitachi Vantara LLC. All rights reserved.
No part of this publication may be reproduced or transmitted in any form or by any means, electronic or mechanical, including copying and recording,
or stored in a database or retrieval system for commercial purposes without the express written permission of Hitachi, Ltd., Hitachi Vantara, Ltd., or
Hitachi Vantara Corporation (collectively “Hitachi”). Licensee may make copies of the Materials provided that any such copy is: (i) created as an
essential step in utilization of the Software as licensed and is used in no other manner; or (ii) used for archival purposes. Licensee may not make any
other copies of the Materials. “Materials” mean text, data, photographs, graphics, audio, video and documents.
Hitachi reserves the right to make changes to this Material at any time without notice and assumes no responsibility for its use. The Materials contain
the most current information available at the time of publication.
Some of the features described in the Materials might not be currently available. Refer to the most recent product announcement for information about
feature and product availability, or contact Hitachi Vantara LLC at https://support.hitachivantara.com/en_us/contact-us.html.
Notice: Hitachi products and services can be ordered only under the terms and conditions of the applicable Hitachi agreements. The use of Hitachi
products is governed by the terms of your agreements with Hitachi Vantara LLC.
By using this software, you agree that you are responsible for:
1. Acquiring the relevant consents as may be required under local privacy laws or otherwise from authorized employees and other individuals; and
2. Verifying that your data continues to be held, retrieved, deleted, or otherwise processed in accordance with relevant laws.
Notice on Export Controls. The technical data and technology inherent in this Document may be subject to U.S. export control laws, including the
U.S. Export Administration Act and its associated regulations, and may be subject to export or import regulations in other countries. Reader agrees to
comply strictly with all such regulations and acknowledges that Reader has the responsibility to obtain licenses to export, re-export, or import the
Document and any Compliant Products.
Hitachi and Lumada are trademarks or registered trademarks of Hitachi, Ltd., in the United States and other countries.
AIX, DB2, DS6000, DS8000, Enterprise Storage Server, eServer, FICON, FlashCopy, GDPS, HyperSwap, IBM, OS/390, PowerHA, PowerPC, S/390,
System z9, System z10, Tivoli, z/OS, z9, z10, z13, z14, z15, z16, z/VM, and z/VSE are registered trademarks or trademarks of International Business
Machines Corporation.
Active Directory, ActiveX, Bing, Excel, Hyper-V, Internet Explorer, the Internet Explorer logo, Microsoft, Microsoft Edge, the Microsoft corporate logo,
the Microsoft Edge logo, MS-DOS, Outlook, PowerPoint, SharePoint, Silverlight, SmartScreen, SQL Server, Visual Basic, Visual C++, Visual Studio,
Windows, the Windows logo, Windows Azure, Windows PowerShell, Windows Server, the Windows start button, and Windows Vista are registered
trademarks or trademarks of Microsoft Corporation. Microsoft product screen shots are reprinted with permission from Microsoft Corporation.
All other trademarks, service marks, and company names in this document or website are properties of their respective owners.
Copyright and license information for third-party and open source software used in Hitachi Vantara products can be found in the product
documentation, at https://www.hitachivantara.com/en-us/company/legal.html.
Feedback
Hitachi Vantara welcomes your feedback. Please share your thoughts by sending an email message to [email protected]. To assist the
routing of this message, use the paper number in the subject and the title of this white paper in the text.
Revision history
MK-SL-145-04 ■ Added a new section Ransomware and data protection recommendations. September 2024
■ Removed content that was not current, updated with VSP One messaging.
■ Updated hyperlinks.
Contents
Optimize Hitachi Storage, Solutions, and Platforms in VMware vSphere Environments 3
Hitachi storage capabilities defined on array-side and advertised by
VASA scheme......................................................................................... 22
Clustered VMDK with vSphere 7.0+.............................................................26
iSCSI............................................................................................................ 26
iSCSI provisioning...................................................................................26
Multipathing with iSCSI........................................................................... 26
VMware vSphere storage optimizations and capacity management............27
UNMAP................................................................................................... 27
VMware vSphere Storage DRS.............................................................. 27
VMware vSphere Storage I/O Control (SIOC)........................................ 28
VMware Aria Operations......................................................................... 29
Hitachi storage resource management........................................................ 29
Dynamic Tiering and Active Flash...........................................................29
Capacity savings, deduplication, and compression with Hitachi
Storage....................................................................................................30
Deduplication recommendations and considerations............................. 30
VMware Site Recovery Manager best practices.......................................... 32
Standard storage SRM and stretched storage SRM with global-
active device best practices.................................................................... 32
VMware vSphere Metro Storage Cluster with global-active device best
practices....................................................................................................... 33
Changes in multipathing and path configuration best practice............... 33
Uniform and non-uniform host access.................................................... 34
3 Data Center (3DC) with VMware Site Recovery Manager best
practices....................................................................................................... 34
Ransomware and data protection recommendations................................... 36
Data protection............................................................................................. 36
Backup and recovery.............................................................................. 37
Ransomware...........................................................................................38
High availability....................................................................................... 40
Disaster recovery.................................................................................... 40
VMware Cloud Foundation (VCF) and external storage.............................. 41
Kubernetes and persistent storage options with Hitachi Virtual Storage
Platform and UCP.........................................................................................43
Contents
Optimize Hitachi Storage, Solutions, and Platforms in VMware vSphere Environments 4
Best Practices Guide
Hitachi Vantara LLC, a subsidiary of Hitachi, Ltd., provides various datacenter infrastructure
components to enable IT environments to support a VMware ecosystem. This includes mid-
range and enterprise storage, converged, and hyperconverged infrastructure as well as a
suite of software and software integrations to enable a robust automated operational
environment.
This document outlines most of the best practices to implement in a VMware vSphere
Foundation (VVF), Virtual Desktop Infrastructure (VDI), or VMware Cloud Foundation (VCF)
environment with Hitachi storage or a converged Hitachi Unified Compute Platform (UCP).
This includes the associated software integrations into various VMware vCenter and Aria
management stacks. These aid in building a VMware environment that provides the
performance, scalability, reliability, usability, resilience, and recoverability expected when
paired with Hitachi products.
Hitachi is a Pinnacle Partner in the Broadcom Advantage Partner Program, a participant in
VMware Ready Partner programs for Storage Infrastructure Services, and a Value-Added
OEM (VAO) partner. Together, Hitachi and VMware by Broadcom (mentioned simply as
“VMware” here onwards in this paper) are committed to providing innovative, business-
enabling technology, with end-to-end virtualization solutions for the software-defined
datacenter.
These best practices cover the Hitachi storage and converged products listed in the following
table.
Hardware Product
Some of the Hitachi software products covered by these best practices are listed in the
following table.
Software Product
Software Product
Note: Testing to develop these best practices was in a lab environment. Many
factors affect production environments beyond prediction or duplication in a lab
environment. Follow the recommended practice of conducting proof-of-concept
testing for acceptable results in a non-production, isolated test environment that
otherwise matches your production environment before your production
implementation of this solution.
For more information see the Storage Plug-in for VMware vCenter User Guide at https://
docs.hitachivantara.com/v/u/en-us/adapters-and-drivers/4.10.x/mk-92adptr047.
For more information, see the Ops Center Protector VMware Application Guide at https://
docs.hitachivantara.com/r/en-us/ops-center-protector/7.7.x/mk-99prt004/vrealize-
orchestrator-integration.
These performance enhancements move the I/O load from the dependent VMware vCenter
host platform into the storage controller. By offloading storage related operations off to the
storage subsystem, it speeds up the datastore and VMDK provisioning operations. This frees
virtualization management for more critical tasks.
When used with VMware vSphere (including VVF), as well as VMware Cloud Foundation
(VCF), Hitachi storage supports the following API primitives:
■ XCOPY or Fullcopy — This primitive enables the storage system to make full copies of
data within the storage system without having the VMware ESXi host read and write the
data.
■ Write Same or Block zeroing — This primitive enables storage systems to zero out many
blocks to speed provisioning of virtual machines.
■ Atomic Test & Set (ATS) or Hardware-assisted locking — This primitive provides an
alternative means to protect the metadata for VMFS cluster file systems, thereby
improving the scalability of large VMware ESXi host farms sharing a datastore.
■ Thin provisioning stun — This primitive enables the storage system to notify the VMware
ESXi host when thin provisioned volumes reach a certain capacity utilization threshold.
When enabled, this allows the ESXi host to take preventive measures to maintain virtual
machine integrity.
■ UNMAP — This primitive enables a VMware ESXi host to inform the Hitachi storage
system that space can be reclaimed that previously had been occupied by a virtual
machine that has been migrated to another datastore or deleted.
LUN size
The following lists the current maximum LUN/datastore size for VMware vSphere and Hitachi
Storage:
■ The maximum LUN size for VMware vSphere 7.x or 8.x is 64 TB.
■ The maximum LUN size for Hitachi Virtual Storage Platform is 256 TB with replication.
■ The LUN must be within a dynamic provisioning pool, or Dynamic Drive Protection pool for
VSP One Block series.
■ The maximum VMDK size is 62 TB (vVol-VMDK or RDM).
Using multiple smaller sized LUNs tend to provide higher aggregated I/O performance by
reducing the concentration of a storage processor (MPB). It also reduces the recovery time in
the event of a disaster. Take these points into consideration if using with larger LUNs. In
some environments, the convenience of using larger LUNs might outweigh the relatively
minor performance disadvantage.
Keep in mind that recovery is typically quicker with smaller LUNs. So, use the size that
maximizes usage of MPB resources per LUN for workloads. Use the VMware integrated
adapters or plugins Hitachi Vantara provides, such as UCP Advisor and vSphere Plugin to
automate datastore and LUN management.
LUN distribution
The general recommendation is to distribute LUNs and workloads so that each host has 2-8
paths to each LDEV. This prevents workload pressure on a small set target ports to become a
potential performance bottleneck.
You should isolate production, and critical systems, to dedicated ports to avoid contention
from other hosts workloads. However, presenting the same LUN to too many target ports
could also introduce additional problems with slower error recovery.
Follow these best practices:
■ Each host bus adapter physical port (HBA) should only see one instance of each LUN
■ The number of paths should typically not exceed the number of HBA ports for better
reliability and recovery
■ Two to four paths to each LUN provides the optimal performance for most workload
environments
VMware vSphere Storage APIs Array Integration (VAAI) — Atomic Test and Set
(ATS)
A change in the VMFS heartbeat update method was introduced in VMware VMFS 5, and
this optimization results in a significant increase in the volume of ATS commands the ESXi
kernel issues to the storage system and causes increased load on the storage system. Under
certain circumstances, the VMFS heartbeat using ATS might fail with false ATS miscompare
events. This causes the ESXi kernel to verify again its access to VMFS datastores. This
leads to “Lost access to datastore” messages.
The resolution of this issue is implemented in VMFS 6. The following setting is recommended
for a VMware vSphere environment:
■ Set the ATS heartbeat OFF for vSphere 6.0 or later with VMFS 5.
■ Keep the default ATS heartbeat ON for vSphere 6.0 or later with VMFS6 without global-
active device configured.
■ Set the ATS heartbeat OFF for vSphere 6.0 or later with VMFS 6 with global-active device
configured.
See ESXi host loses connectivity (2113956) for more details and how to turn off ATS.
Conventional Fibre Channel and iSCSI requires an LU mapping for a port to manage an
access route between the host and the logical volume. NVMe-oF, on the other hand, requires
the following system components to be configured on the storage system between the host
and the logical volume.
■ NVM subsystem: A flash memory storage control system that supports the NVMe-oF
communication protocol with one or more namespaces, and one or more communication
ports (NVM subsystem ports).
■ Namespace: A flash memory space formatted into a logical block.
■ NVM subsystem port: A Fibre Channel port set to NVMe mode.
■ Host identification (host NQN): Host name qualifier.
■ Host-namespace path: Access permission to the namespace for each host NQN
registered on the NVM subsystem.
■ There is no concept of Host mode options in an NVMe subsystem.
At the time of publishing this document, global-active device with remote NVMe-oF is not
supported.
For detailed implementation guidelines, see the Provisioning Guide for VSP One Block at
https://docs.hitachivantara.com/r/en-us/svos/10.2.x/mk-23vsp1b012/planning-for-port-
connections/configuration-of-nvme-of-and-nvme/tcp-for-the-host-and-the-storage-system.
Zoning
Use zoning to enable access control in a SAN environment. Through zoning, a SAN
administrator can configure which HBA WWPNs on the VMware ESXi host can connect to
which WWPNs on the Hitachi storage processors.
The VMware ESXi host port in the Fibre Channel HBA is referred to as the initiator. The
storage processor port in the Hitachi storage system is referred to as the target.
You can break zoning down into the following different configurations:
■ Single Initiator to Single Target (SI-ST) Zoning — This configuration allows one initiator to
be zoned to only one target. This configuration is the most resilient configuration, as traffic
originating from another Initiator on the SAN will have less impact than the initiator in this
zone.
■ Brocade Peer Zoning — This configuration allows a single zone to provide a Principal–
Pupil relationship where all pupils can communicate with the principal but not with each
other. This provides the same zone-security as SI-ST zoning but with the administrative
benefit of a reduction of number of zones. This is the preferred configuration in a Brocade
fabric configuration.
■ Cisco Smart Zoning — This implementation is preferred in a Cisco environment where
NX-OS can eliminate initiator to initiator and target to target communication.
■ Single Initiator to Multiple Target (SI-MT) Zoning — This configuration allows one initiator
to be zoned to multiple targets in a single zone.
■ Multi Initiator Zoning — This configuration is never recommended . This configuration
allows multiple initiators to be zoned to multiple targets in a single zone. This exposes all
initiators and targets to all traffic in the zone. Events such as a malfunctioning HBA could
affect all initiators and targets in the zone and either negatively affect performance for all
or bring down the Fibre Channel network completely.
Hitachi generally recommends the following:
■ For utmost availability with slightly higher administrative cost, recommends SI-ST zoning.
Brocade Peer Zoning and Cisco Smart Zoning is supported to reduce admin burden.
■ Each HBA port should only see one instance of each LUN. This is primarily based on
years of experience with fabrics and to avoid potential availability issues where host HBA
ports can be overrun leading to performance issues and error recovery with fabric path
issues (transient or otherwise) is faster and less impactful to hosts.
■ See Recommended Multi-Path Settings for Hitachi Storage at https://
knowledge.hitachivantara.com/Knowledge/Storage/Recommended_Multi-
Path_Settings_for_Hitachi_Storage?
_ga=2.106482883.176345872.1547835671-1693012570.1547243188 for more
information.
Optionally, do the following:
■ Use SI-MT with Brocade Peer Zoning or Cisco Smart Zoning and follow same LUN
presentation recommendation previously.
■ Regarding SI-MT, an example to use is provided within Cisco and Hitachi Adaptive
Solutions for Converged Infrastructure and Cisco and Hitachi Adaptive Solutions for SAP
HANA Tailored Data Center Integration.
Zoning is configured as SI-MT with Cisco Smart Zoning to optimize traffic intended to be
specific to the initiator (Cisco UCS host vHBA) and the targets (Hitachi Virtual Storage
Platform controller ports).
Using SI-MT zoning provides reduced administrative overhead versus configuring traditional
SI-ST zoning, and results in the same SAN switching efficiency when configured with Smart
Zoning. See the Cisco and Hitachi Adaptive Solutions for Converged Infrastructure Design
Guide at https://www.cisco.com/c/en/us/td/docs/unified_computing/ucs/UCS_CVDs/
cisco_hitachi_adaptivesolutions_ci_design.html for more details.
Multipathing
Multipathing allows a VMware ESXi host to use more than one physical path between the
ESXi host and the storage array. Multipathing provides load balancing. This is the process of
distributing I/O across multiple physical paths, to reduce or remove potential bottlenecks.
Multipathing also provides redundancy and fault tolerance in case of a failure of any element
in the SAN network, such as an adapter, switch, or cable. The ESXi host can switch to
another physical path that does not use the failed component. This process of path switching
to avoid failed components is known as path failover.
To support path switching with a Fibre Channel SAN, the ESXi host typically has two or more
HBAs available from which the storage array can be reached. It also has full fault tolerance
that uses two or more switches. Additionally, for full fault tolerance, two storage processors
on Hitachi storage systems should be utilized so that the HBA can use a different path to
reach the disk array.
Available multipathing policies supported by ESXi hosts are Round Robin, Most Recently
Used, Fixed, and Hitachi Dynamic Link Manager.
Hitachi recommends using the Round Robin Multipathing PSP policy (VMW_PSP_RR) and
using the SATP default of active-active (VMW_SATP_DEFAULT_AA). In a global-active
device configuration, Round Robin Multipathing PSP and using ALUA SATP
(VMW_SATP_ALUA) are recommended options. This multipathing policy takes advantage of
all available paths and bandwidth. Taking advantage of all available paths assures maximum
performance from the SAN infrastructure. Note, with vSphere 6.7U1 and vSphere 6.5 P03 or
later, the round robin multipathing policy became the default setting as part of the SATP claim
rules for Hitachi storage.
In a global-active device configuration without ALUA configured, the fixed policy is the
preferred PSP to ensure writes are sent to the preferred side.
As part of the VMware ESXi Round Robin Path Selection Plug-in (PSP), there is an I/O
quantity value when a path change is triggered that is known as the limit. After reaching that
I/O limit, the PSP selects the next path in the list.
The default I/O limit is 1000 but can be adjusted if needed to improve performance.
Specifically, it can be adjusted to reduce latency seen by the ESXi host when the storage
system does not see latency.
The general recommendation for the PSP limit is to continue to use the default value of 1000
in typical VMware mixed environments with multiple ESXi hosts with multiple datastores. It
has been observed that to set value of 1 or 20 potentially provides the optimum value for an
additional 3-5% latency improvement potentially reducing path error detection.
For reference, here is information on various multipath policies:
■ Round Robin (VMware) — This policy sends a set number of I/O down the first available
path, then sends the same set number of I/O down the next available path. This repeats
through all available paths, and then starts over again and repeats. If a path becomes
unavailable, it is skipped until it becomes available again.
■ Most Recently Used (VMware) — This policy uses the last successfully used path. If the
last successfully used path is not available, then path failover occurs, and the next
available path is used. This new path continues to be used until it is no longer available,
even if the previously used path becomes available again.
■ Fixed (VMware) — This policy has a preferred path that can be set by the VMware
vCenter administrator. This path is used until it becomes unavailable. Then, it fails over to
the next available path until it becomes unavailable. In which case, the path fails over to
the next available path, or the preferred path when it becomes available again. If the
preferred path does become available again, then the system fails back to the preferred
path.
■ Hitachi Dynamic Link Manager —VMware ESXi also supports third party path selection
policies. Hitachi Dynamic Link Manager is multipathing software from Hitachi that
integrates with global-active device on Hitachi Virtual Storage Platform to provide load
balancing and path failover capabilities for servers.
■ vVols can share an existing pool or pools and resource groups with VMFS datastores or
have dedicated pools and resource groups. The best practice to date has been to create a
separate resource group and pool that can share an underlying parity group or use a
dedicated parity group if available.
■ Communication from the VASA provider to Hitachi storage for vVols operations is using
the service processor (SVP). You can deploy the SVP as a preconfigured physical 1U
node (or nodes) or installed as a virtual machine appliance.
■ To ensure the high availability of vVol out-of-band management operations, treat the VASA
appliance as if they are availability deployment modes used for a vCenter appliance
(VCSA) or NSX- T appliance. The VASA provider from Hitachi supports the following
availability features:
● VMware vSphere Fault Tolerance (FT)
● VMware vSphere High Availability (HA)
The VASA provider from VMware also supports monitoring of its application service level
under the VMware vSphere High Availability configuration. By enabling the monitoring of
the virtual machine and application option within VMware vSphere High Availability, the
VASA provider will automatically restart if the VASA provider service stops.
■ When Storage Provider for VMware vCenter or SVP becomes unavailable, only storage
management control operations related to the virtual volume are affected, such as clone
virtual machines, snapshot operations, and power on operations. This out-of-band
communication does not affect virtual machine I/O, as the I/O flows through the fabric data
path using protocol endpoints.
Policy-based management is premised on the fact that storage exposes a set of capabilities.
That is, a resource that provides Tier 1 IOPS and encryption and remote replication services.
Virtual machine administrators define virtual machine storage policies policy for virtual
machine producers selecting one or more of these capabilities and values that are exposed
by the VASA provider. When a virtual machine is provisioned and assigned a virtual machine
storage policy, vCenter asks the VASA provider to create that virtual machine and its VMDKs
in storage resources or services that match that virtual machine storage policy.
The following figures show an example of a storage capability profile definition by VASA
provider in the web user interface, an example of virtual machine storage policy definition by
VMware vSphere Web Client for both placement and replication capabilities, and the visibility
of those vVols and their associated storage policy assignment.
Note that replication services for Hitachi VASA Provider are enabled with Ops Center
Protector.
2. Create a VMware vCenter virtual machine storage policy with capabilities of Tier 3 IOPS
+latency for these virtual machine services.
3. Review the capability schema in the VASA web user interface to see how VASA maps
storage policies to tiers created with Dynamic Tiering.
When you assign this policy to these aging or less relevant virtual machines, the VASA
provider will detect that policy change and automatically tier those warm or cold virtual
machines or VMDKs to the lowest legacy tier in this configuration. This frees up that tier 1 for
the next revenue generating app.
Hitachi Storage (VASA) provider enabling tag-based storage policy for VMFS
datastores
The tag-based storage policy provides an outcome for VMFS datastores similar to what
storage policy-based management (SPBM) does for VMware Virtual Volumes.
You can set a storage capability profile on the pool that is serving the VMFS datastores or
you can customize the storage capability profile of an individual LUN. Hitachi Virtual Storage
Platform automatically tags the datastores in VMware vCenter for existing and new
datastores and these will appear in vCenter as tags.
Like SPBM for VMware vVols, you can create a virtual machine storage policy using tags to
allow virtual machine producers to find the right datastore that matches their requirements. It
also allows you to create custom capabilities (such as rack location, availability zone) to
augment the base capabilities provided.
iSCSI
This section describes volume provisioning using iSCSI.
iSCSI provisioning
iSCSI initiators and targets use TCP to create relationships called sessions. The initiator sees
one logical connection to the target. An iSCSI session might also contain multiple logical
connections.
From a VMware vSphere host perspective, these sessions might also be thought of in term of
paths between the initiator and target. Having multiple connections per session enables
bandwidth aggregation and can also provide load balancing.
Although vSphere does not support multiple connections per session, by configuring multiple
sessions, you can configure load balancing and ensure path redundancy.
UNMAP
VMFS 6 automatically issues the UNMAP command to release free storage space in
background on thin-provisioned storage arrays that support UNMAP operations.
The main requirements to take advantage of UNMAP are listed below:
■ Use Thin provisioned VMDKs backed with thin provisioned LDEVs/LUs or vVols
datastores.
■ VMFS 6 datastores.
■ In-Guest UNMAP support:
● Linux guest OS with hardware version 13 or later to present SCSI-4.
● Windows 2012 R2 OS or later with VM hardware version 11 or later.
● In-Guest automated UNMAP also supported for VMware vVol datastores.
■ Storage: VSP E series, VSP G series, VSP F series, VSP 5000 series, or VSP One Block
series.
■ Ensure Host Mode Options (HMO) 54 and HMO 63 are set to ON.
See the blog Dealing With VMware Datastore Space Management On VSP Storage - Part 2
at https://community.hitachivantara.com/s/article/Dealing-with-VMware-Datastore-space-
management-on-VSP-Storage-part-2 for additional details.
In VMFS 6, VMware enabled additional performance parameters (low, medium, high, and
fixed) to determine the UNMAP throughput that arrays would receive. Setting automatic
space reclamation settings to a fixed rate at 100 MBps provides a reasonable combination of
UNMAP rate and storage processor (MPU) utilization. Always monitor MPU usage before and
after changing the UNMAP rate.
For VMFS 5, manual UNMAP is still supported with the following command:
This helps improve the speed, capacity, and cost of performance. Dynamic Tiering improves
underlying storage resources with following conditions:
■ Trigger — Monitor the I/O load per page and relocate the page to the optimal tier.
■ Time — Define a user-specified period of at least 30 minutes.
● Real-time with the active flash.
■ Granularity — Relocate the storage tier with a page size of 42 MB.
In addition, active flash monitors a page's access frequency level in real time to promote
pages that suddenly became busy from a slower media to high-performance flash media.
In a VMware environment, many workloads tend to be highly random with smaller block size.
This might not be suitable for deduplication, even with an all flash configuration. Dynamic
Tiering with active flash could be a good option to improve capacity and cost by efficiently
using the flash tier minimally.
Microsoft Office® Because there are many identical file copies, deduplication
is effective.
Multiple Server VMs in the Deduplication is effective if you have many server VMs with
same storage pool the same OS deployed in the same datastore or multiple
datastores from the same storage pool.
Standard storage SRM and stretched storage SRM with global-active device best
practices
VMware vCenter Site Recovery Manager integrates tightly with Hitachi storage systems using
either Hitachi Storage Replication Adapter or Hitachi Ops Center Protector. This provides
centralized management of recovery plans. Tight integration between storage systems,
VMware vCenter, VMware vCenter Site Recovery Manager, and Hitachi storage replication
adapters ensures a coordinated recovery for large, business critical environments.
Remote data replication is a key function in building out stable and reliable disaster recovery
environments. Replicating data to a remote secondary site represents the most effective
insurance policy against a catastrophic failure. Although you can perform data replication at
the server level, you can perform data replication more effectively within the storage
infrastructure.
The following are the two types of underlying storage configurations supported by VMware
Site Recovery Manager:
■ Standard storage (active-standby solution) leveraging Hitachi True Copy or Hitachi
Universal Replicator
■ Stretched storage (active-active solution) leveraging global-active device on Hitachi Virtual
Storage Platform the following table lists the differences between the two types of storage.
Table 2 Comparison between standard storage and stretched storage
Business Site failover is required with During planned migration such as site
continuity down time even though maintenance, no disruption and down time
planned migration such as occurs by using Cross-vCenter vMotion with
site maintenance is being Stretched storage is made up by global-
conducted. active device and Hitachi Storage
Replication Adapter.
Storage Site failover is required due When primary storage failure occurs, no site
availability to primary storage failure. failover is required by using the cross path
to remote site storage which is virtualized as
It costs application
a single stretched storage and volume
downtime.
across the sites powered by global-active
device technology.
You can decide, depending on required RPO, which results in which replication type you
want. For more information on the SRAs, see the following:
■ Storage Replication Adapter for VMware vCenter Site Recovery Manager - Deployment
Guide at https://docs.hitachivantara.com/v/u/en-us/adapters-and-drivers/2.5.1/
mk-09rm6745.
■ Ops Center Protector VMware Application Guide at https://docs.hitachivantara.com/r/en-
us/ops-center-protector/7.7.x/mk-99prt004/site-recovery-manager-integration
Here is an example of enabling a SATP rule set as ALUA for Hitachi devices and selecting
PSP as round robin on the VMware ESXi side:
■ Hitachi recommends using RR (round robin) instead of MRU (most recently used).
■ With VMware vSphere 6.7U1 and vSphere 6.5 P03 or later, this ALUA SATP rule is set by
default.
Hitachi Universal Replicator with delta resync functionality establishes storage remote
replication from the primary data center to the third data center and from the secondary data
center to the third data center, respectively. This is called the global-active device 3DC delta
resync environment, as shown in the following illustration.
To maintain adequate service level agreements, ensure journals are adequately sized to
avoid any throttling of IOPS to maintain replication SLAs if Hitachi Universal Replicator inflow
control is set to enabled. If Universal Replicator inflow control is set to disabled, host I/O is
prioritized and the Universal Replicator relationship is suspended.
Installing VMware Site Recovery Manager in this 3DC environment gives you the
orchestrated and repeatable planned or unplanned migration or disaster recovery operations
using a tested and proven recovery plan. This enables end-to-end virtual machine protection
across 3 data centers. As a normal state, VMware SRM protects the virtual machine between
the primary and the third data center.
This solution is based on VMware Metro Storage Cluster, which clusters the primary and the
secondary data centers within a single VMware vCenter data center object and uses
stretched storage cluster powered by global-active device.
When the primary data center goes down, the virtual machine can be restarted on the
secondary data center, leveraging VMware vSphere High Availability fail over functionality as
a VMware Metro Storage configuration. During failover from the primary to secondary
datacenter, storage remote replication established from the primary to the third data center is
also automatically failed over to the other one established from the secondary to the third
data center by leveraging delta resync functionality.
For a global-active device 3DC delta resync environment solution, the virtual machine
protected by VMware SRM can follow this data center failover movement and re-protect the
virtual machine between the secondary and third data center with minimal effort.
Note: As a normal state, VMware SRM protects the virtual machine between the
primary and third data center. When the primary data center goes down, storage
remote replication can automatically failover though, the re-protection of virtual
machines by SRM requires some manual operation to switch the source and
target datacenter. To do this, switching the command control interface
configuration file and restarting the Hitachi Open Remote Control Manager
daemon is required.
From a data protection standpoint, Hitachi has partnered with VM2020, and using their
CyberVR with our latest VSP One Block and Thin Image Advanced (HTI Advanced)
snapshots, we offer together a fully orchestrated protection and recovery solution that is
simple, quick, and resource-efficient. We call it “The World’s Fastest Ransomware Recovery
form Immutable Snapshots for VMware Environments.”
Data protection
Hitachi Vantara provides robust data protection offerings with our top four partners namely
Commvault (HDPS), Veeam, Veritas, and VMware VM2020. This includes areas in Backup
and Recovery, Ransomware, High Availability and Disaster Recovery. This section describes
the best practices in terms of the level of integration of each product.
Commvault
Hitachi Data Protection Suite (HDPS) solution integration with VMware focuses on two areas:
■ Virtual Storage Platform (VSP) storage system as the target primary backup repository.
■ Virtual Storage Platform One Object as the target secondary backup repository using
Auxiliary Copy.
The VSP storage system is deployed using Fibre Channel and presented to VMware/vCenter
as a datastore where VMs, folders, and files are configured as source. When formatting a
datastore as VMFS, it is recommended to use the latest VMFS version. HDPS also supports
eligible data type for backup namely virtual machines (Linux/Windows), VM templates, VMDK
files, Virtual RDMs, GPT or dynamic disk volumes, vSphere tags on virtual machines, and VM
custom attributes to name a few.
HDPS also offers IntelliSnap backup operations which create a point-in-time snapshot of data
to be used for various data protection operations. IntelliSnap backup works in conjunction
with software and hardware snapshot engines to provide snapshot functionality for data
protection operations. In Hitachi cases, VSP is integrated with Hitachi Thin Image. For
environments leveraging Fibre Channel storage like VSP, it is recommended to install the
Virtual Server Agent and Media Agent on a physical server.
Veeam
Veeam solution integration with VMware brings the power of protecting virtualized workloads.
VSP presented as a datastore in VMware connected using Fibre Channel or ISCSI is used by
Veeam as a source data/workload. There are two use cases for VSP:
■ VSP as the Veeam Primary Repository (Standalone)
■ VSP as part of the Veeam Scale-Out Repository (SOBR) – Performance Tier
Veeam Best Practices for VMware:
1. 3-2-1-1-0 Protection Rules: Ensure that three copies of data comply with the
conventional aspect of this rule. Maintain three (3) copies of data (original data and at
least two copies), Use two (2) different types of media for storage; one (1) copy offsite,
one (1) copy offline, air-gapped or immutable and zero (0) errors with SureBackup
Recovery verification.
2. Select the appropriate backup mode. It is important to test which backup mode best
suits your environment. Network mode (NBD) and direct storage access (excluding
backup from storage snapshots) both use vSphere APIs – Data Protection (formerly
known as vStorage API for Data Protection or VADP) which can impact backup
performance. Veeam has the ability to bypass this API in these scenarios namely:
3.2 Direct NFS (Network File System), like direct storage access
Veritas
Veritas NetBackup integration with VMware delivers strong protection of various data
sources. NetBackup uses VSP as the target primary backup storage as well as the source of
the VMware datastore for workloads. It is recommended that datastores are created using the
latest version of VMFS.
VM2020/CyberVR
CyberVR is a software solution that is deployed as a virtual machine (either a pre-configured
OVA or installed in an existing/new dedicated virtual Windows server. VSP provides the
LDEVs presented to the vCenter/ESXi Cluster as datastores used as a source of production
VMs. It is recommended that datastores are created using the latest version of VMFS.
The following are considerations for CyberVR with VMware:
■ It is recommended that you back up CyberVR VM regularly to safeguard against failures/
correction.
■ The integration of CyberVR to VMware is that CyberVR connect to vCenter Server and
Ops Center Protector from APIs/SDKs. CyberVR depends on LDEV snapshots generated
by Ops Center Protector and leverages the snap-of-snap feature of Protector to make
sure that the snapshots themselves are not impacted/tampered.
■ It is recommended that a dedicated user (a service account) be created within Ops Center
for CyberVR so Ops Center Protector can view/audit active taken by CyberVR.
Make sure to use naming conventions to be able to map VMware ESXi compute hosts to
equivalent Ops Center Protector Host Groups. Make sure that existing and active Dataflows
taking snapshots of LDEVs are present.
Ransomware
Commvault
Commvault ransomware solutions are built on responsiveness, innovation and rapid
execution. By strengthening organization's resiliency security posture and stay one step
ahead of ransomware with proactive data security best practice. Commvault has put in
protection and monitoring capabilities aimed explicitly against malware, including
ransomware, for fast recovery.
The following are considerations for Commvault with VMware:
■ It is recommended that you have a multi-layered security plan that can protect, detect, and
recover.
■ Commvault leverages immutable data storage mechanisms to protect backup copies from
unauthorized modifications or deletions.
■ Commvault assists in rapid data recovery from clean backup copies, minimizing downtime
or disruption in business operations. The capability to perform a granular restore allows
restoring specific files/apps/entire systems and use of automated workflows to accelerate
recovery times.
■ Ensure control to access with hardened security protocols and zero-trust access controls.
Veeam
Hitachi Vantara and Veeam collaborated to build stronger product solutions that addresses
the challenges of ransomware protection and cyber resilience. The combined products such
as VSP, HCP Anywhere, VSP One Object, and Veeam Data Platform provides most effective
defense against ransomware long with proven ransomware recovery.
The following are considerations for Veeam with VMware:
■ It is recommended that you have an end-to-end solution for ransomware aligned to the
NIST v2.0 framework which covers a multi-layered approach offering ransomware
protection to build a secure and resilient infrastructure.
■ With VMware VMs as a source data, it is recommended that these are scanned before
backup through VBR malware and threat detection methods.
■ Make sure to use a VBR scale-out repository where data is stored in primary storage
(VSP) and tiered to secondary storage (HCP Anywhere/VSP One Object) for immutability
storage protecting data from any modification/deletion from ransomware.
■ Enforce physical and logical security to VMware virtual machines by providing access
control, assigning appropriate roles/permissions, encryption (in-flight and at-rest), and
isolation/air-gapping.
■ Consider Veeam approach to Zero Trust Data Resilience (ZTDR) which provide critical
techniques to fight risk by separating backup management software and backing up
storage into separate resilience zones or security domains.
■ Consider deploying Veeam One to bring clear visibility into VMware virtual machines and
Veeam-protected cloud and physical workloads. This provides powerful infrastructure
monitoring and reporting (provides alerts for potential ransomware), intelligent diagnostics
and automated solutions, infrastructure utilization, and capacity planning.
Veritas
Veritas NetBackup integration with VMware delivers strong protection of various data
sources. NetBackup uses VSP as the target primary backup storage as well as the source of
the VMware datastore for workloads. It is recommended that datastores are created using the
latest version of VMFS.
The following are considerations for Veritas NetBackup with VMware:
■ VMware virtual machines is one of the data sources that Veritas NetBackup supports to
back up and recover from any potential ransomware attacks. Consider the combination of
VSP snapshot technology which provides fast, efficient data recovery and VSP One
Object which offers another layer of protection by having an immutable, secure, tamper-
proof data archive.
■ Make sure that VMware virtual machine data is scanned for ransomware using the
NetBackup Anomaly Detection Engine
VM2020/CyberVR
The combination of VM2020 CyberVR, Hitachi storage systems, and Hitachi Ops Center
Protector provides a solution that addresses ransomware attacks by leveraging Hitachi
snapshot technology and CyberVR automation to create an air-gapped test environment that
helps protect the VMware virtual machine's environment.
The following is a consideration for VM2020/CyberVR with VMware:
■ The CyberVR approach to recover VMware virtual machines from ransomware attacks
uses high-level steps such as pre-recovery, isolated recovery, re-protection, and re-
connection.
High availability
Commvault
Commvault has the capability to provide service availability and business continuity by
making sure individual Commvault components such as CommServe Server, Media Agent
Server, and Web Server are configured for high availability.
Veeam
The need for addressing common causes of outages/downtime such as power outages,
connectivity issues, hardware failure, and human error or natural disaster is a big challenge.
The pursuit of high availability, particularly in virtual environments such as VMware, is
complex but is a necessary undertaking.
The following are considerations for Veeam with VMware:
■ Employ cluster-aware file systems such as GFS2 or OCFS2.
■ Use load balancing algorithms like weighted round robin or least connection which offer
traffic distribution, and optimize resource usage and response times.
■ Leverage enhanced VM migration techniques.
■ Automate fault detection and recovery in virtual environments by developing specialized
monitoring systems for the virtual network infrastructure.
■ Storage I/O and network I/O control can help prioritize access to storage and network
resources.
Disaster recovery
VM2020/CyberVR
VM2020/CyberVR has the capability to run in multi-site architecture to address disaster
recovery.
VMware Cloud Foundation (VCF) supports Hitachi SAN storage (VMFS and vVols) used as
principal or supplemental storage. The use cases vary by customer, ranging from flexibility to
scale storage footprint, mission critical applications with stringent RPO/RTO requirements, to
simply matching business outcomes to suitable tier of storage.
VCF expanded their external storage offerings for principal or supplemental storage with
support for vVols and VMFS to augment native vSAN datastores. The best practice is to
deploy the VASA provider OVA from Hitachi in the VCF management domain and each VCF
workload domain VMware vCenter can register to that single VASA provider.
The following figure shows the workload domain options available on Hitachi Unified
Compute Platform RS with VMware Cloud Foundation. SeeThe Easier Path to the Hybrid
Cloud - Using Hitachi Virtual Storage Platform with VMware Cloud Foundation and VMware
Virtual Volumes Reference Architecture Guide at https://docs.hitachivantara.com/v/u/en-us/
application-optimized-solutions/mk-sl-222 for more information.
For VMware Cloud Foundation with metro storage stretched cluster solution, see the UCP RS
with VMware Cloud Foundation Supports Metro Storage Cluster Datastores from Virtual
Storage Platform Lab Validation Report at https://docs.hitachivantara.com/v/u/en-us/
application-optimized-solutions/mk-sl-239.