Data Protection For FlashStack With Cohesity Reference Architecture
Data Protection For FlashStack With Cohesity Reference Architecture
Executive Summary...........................................................................................................................................3
Audience...............................................................................................................................................................3
What is FlashStack?..........................................................................................................................................4
Topology................................................................................................................................................................5
Cohesity Overview.............................................................................................................................................11
Conclusions..........................................................................................................................................................57
We will discuss the design principles of building a next-generation, “holistic and cohesive” data center compute, network,
and storage solution that combines the simplicity, resilience, and performance of next-generation flash storage from Pure
Storage® with compute and networking from Cisco® and secondary storage from Cohesity®.
This document specifically provides insight on application-centric, pre-validated architectures that deliver high performance.
Its focus is to ensure that the recommended architectures can be deployed quickly and scaled seamlessly. We believe that
this in turn provides maximum agility while also reducing IT capex and opex, because bottom line costs are more often than
not the most important factor in industry.
1. W
e will showcase and certify that Cohesity and Pure Storage deliver an optimal blend of performance, user-
friendliness, non-disruptive operations, and data recoverability from the micro to the macro for multiple workloads.
2. W
e will describe a Pure FlashStack™ solution reference architecture that illustrates how these primary and
secondary storage solutions, along with Cisco UCS® and Nexus®, fit together to provide an optimal Converged
Infrastructure (CI) for a VMware® Horizon solution that includes the ability to rapidly recover the entire
environment – all the way down to individual user files and everything in between.
3. F
inally, we will showcase how this unique FlashStack CI solution brings the best of both primary and secondary
storage, hence delivering value to applications in tier 0 through tier 4 workloads.
In summary, we discuss all aspects of running high-performance VMware and physical workloads while focusing on
protecting these workloads at every level in the solution stack.
Incumbent CI solutions tend to focus on the performance of tier 0-1 workloads while looking at tier 2-4 data protection
and data management workloads as an afterthought. However, we strongly believe that integrating data protection at the
ground level is the key to a successful deployment. This paper will go deep into the proposed architecture to support the
claim. The Flashstack solution combined with Cohesity is built to be optimized at each of these tiers and hence delivers the
ultimate value to the end customer by simplifying adoption of CI with a well thought-out design for all workloads running in
the modern data center.
While this document’s primary test case will focus on a VMware Horizon persistent VDI environment, it is important to
note that the use cases shown here are extensible throughout the data center to databases, VSI, physical host workloads
and beyond.
Audience
The target audience for this document includes storage and virtualization administrators, consulting data center architects,
field engineers, and desktop specialists who want to implement a next-generation data center. A working knowledge of
VMware vSphere, VMware Horizon, Login VSI, server, storage, backups, networks, and data center design is assumed but is
not a prerequisite to read and understand this document.
This complexity is one of the main motivations for enterprises to consider moving their applications to the cloud. Cloud
offers the prospect of seamless integration and scale and supposedly takes away the headache of managing siloed
infrastructure IT. However, major downsides of this approach are that the cloud can be extraordinarily expensive to operate
at scale and the performance of the most important applications can be inconsistent.
There is a strong need for an overall solution that is optimized to work out-of-the box with different component parts,
delivers simplicity in management by consolidating multiple workloads, and leverages the best aspects of the cloud.
Just as virtualization solved the hardware sprawl problem, and hyper-convergence helped resolve the complexities of
scale by converging virtualization, compute, network, and storage, and the next big challenges to address are silos and
fragmentation.
What is FlashStack?
FlashStack CI (Converged Infrastructure) is a flexible, all-flash converged infrastructure solution that brings the flash
revolution to your data center, faster. It combines the latest in compute, network, storage hardware, and virtualization
software into a single, integrated architecture that speeds time to deployment, lowers overall IT costs, and reduces
deployment risk. Highly efficient components coupled with simple, intelligent software reduce the costs associated with
power, cooling, and data center space. Based on 100 percent flash storage, FlashStack CI provides the performance and
reliability that business-critical applications demand.
The hardware foundation of FlashStack CI includes Pure Storage FlashArray//M, Cisco UCS Blade Servers, Cisco Nexus
ethernet switches and Cisco MDS fibre channel switches. VMware vSphere provides the virtualization technology and
Cohesity provides the secondary storage layer for data protection and recovery.
FlashStack CI is available from qualified FlashStack Partners who help to provide an excellent converged infrastructure
ownership experience. FlashStack Partners have the knowledge and experience necessary to help streamline the sizing,
procurement, and delivery of your entire system while simply, easily, and non-disruptively growing with your organization
as needs change. A single point of contact for support issues for all components of the FlashStack solution ensures a rapid
response to any issues in the stack - and often preemptively, before an issue arises.
Both hardware and software components are combined into a single integrated unit that helps facilitate faster deployments
and lowers overall IT costs. Repeatable, turnkey designs for the most common workloads means that your users are online
faster and data center administrators are focused on more important automation tasks up the stack.
The above topology diagram shows the recommended connectivity and components used in this whitepaper.
Pure Storage supports mixed Fibre-Channel and iSCSI connectivity, which has the benefit of segregating
backup traffic from production workload traffic, thereby enhancing environment resiliency, isolating network
traffic and minimizing Recovery Time Objective (RTO) with Cohesity. FlashStack CI has no single point of
failure and maintains 100% performance through component failure allowing customers to non-disruptively
upgrade and/or add hardware to the solution stack transparently to users.
Today, Cohesity supports iSCSI connectivity, which has been denoted in the above topology diagram. Multiple
connections to redundant Nexus switches ensure both throughput as well as protection from any single
path or component failure. Worth noting is that mixed FC and iSCSI connectivity with Pure Storage is not
a requirement for this design; however, it is recommended as it segregates backup traffic from production
traffic, further improving the resilience of the design. For Pure Storage, using only iSCSI or Fibre Channel
connectivity is fully supported and performant as well.
The port connections shown here from the Pure Storage GUI show the mixed FC and iSCSI connectivity used in this design.
System Requirements
Here we show the supported operating systems for the integration defined in this document.
For detailed supportability information and release notes for Cohesity version 3.5.1, please refer to:
http://pubs.cohesity.com/3_5_1/PDFs/ReleaseNotes.pdf
For detailed supportability information and release notes for Pure Storage, please refer to:
https://support.purestorage.com/FlashArray/Getting_Started/Compatibility_Matrix
• Reserve eight IP Addresses (four Node IP Addresses and four VIP Addresses) on a single subnet, the Cluster
Subnet
• Verify that the Cluster Subnet can communicate with the vCenter Server Subnet
• Reserve four 10GbE ports on one switch (recommended for best performance)
• Reserve four 10GbE ports on another switch (recommended for best performance)
• Reserve four IP Addresses for IPMI Interfaces. They can be in the IPMI Subnet or the Cluster Subnet.
• Reserve four 1GbE ports for the IPMI Interfaces on a switch
• Enable multicast traffic on your network (required for Cohesity’s auto-discovery function)
From a network perspective, Pure Storage requires the following as tested in this design guide:
• Reserve three IP addresses (one IP per controller, one IP as VIP) on a single subnet
• Verify that the array can communicate with the vCenter subnet as well as the Cohesity appliance
• Reserve two 10GbE ports on one switch for Pure Storage to Cohesity backup traffic (recommended for best
performance and resiliency)
• Reserve two 10GbE ports on a different switch for Pure Storage to Cohesity backup traffic (recommended for best
performance and resiliency)
• Reserve six 16GB Fibre Channel ports on a Fabric Interconnect for production network traffic (recommended for
best performance and resiliency)
• Reserve six 16GB Fibre Channel ports on a different Fabric Interconnect for production network traffic
(recommended for best performance and resiliency)
Note that there is no restriction on using 100% FC or iSCSI in our design for Pure Storage. The suggested configuration
above was used in order to separate production and backup network traffic and improve the resiliency of the design by
implementing multiple data paths.
The 6332-16UP Fabric Interconnect is the management and communication backbone for Cisco UCS B-Series Blade Servers,
C-Series Rack Servers, and 5100 Series Blade Server Chassis. All servers attached to a 6332-16UP Fabric Interconnect
become part of one highly available management domain. The 6332-16UP is a core part of the Cisco Unified Computing
System (Cisco UCS), and is typically deployed in redundant pairs. Because it supports unified fabric, the Cisco UCS 6300
Series Fabric Interconnect provides both LAN and SAN connectivity for all servers within its domain. The 6332-16UP offers
40 ports in one rack unit (RU), including:
• 24 40-Gigabit Ethernet and Fibre Channel over Ethernet (FCoE)
• 16 1- and 10-Gbps and FCoE or 4-,8-, and 16-Gbps Fibre Channel unified ports
Cisco Fabric Extender technology scales up to 20 chassis in a single unified system without additional complexity. This
means that customers can eliminate dedicated chassis management and blade switches, as well as reduce cabling.
Cisco’s first blade-server chassis offering, the Cisco UCS 5108 Blade Server Chassis is six rack units (6RU) high, can mount in
an industry-standard 19-inch rack, and uses standard front-to-back cooling. A chassis can accommodate up to eight half-
width or four full-width Cisco UCS B-Series Blade Server form factors within the same chassis.
Optimized for data center or cloud, the Cisco UCS B200 M4 can quickly deploy stateless physical and virtual workloads with
the programmability of the UCS Manager and the simplified server access of SingleConnect technology.
The UCS B200 M4 is built with the Intel® Xeon® E5-2600 v4 and v3 processor family, up to 1.5 TB of memory (with 64 GB
DIMMs), up to two drives, and up to 80 Gbps total bandwidth. It offers exceptional levels of performance, flexibility, and I/O
throughput to run the most demanding applications.
In addition, Cisco UCS has the architectural advantage of not having to power and cool switches in each blade chassis.
Having a larger power budget available for blades allows Cisco to design uncompromised expandability and capabilities in
its blade servers.
The UCS B200 M4 Blade Server delivers:
• Suitability for a wide range of applications and workload requirements
• Highest-performing CPU and memory options without constraints in configuration, power, or cooling
• Half-width form factor offering industry-leading benefits
• Latest features of Cisco UCS Virtual Interface Cards (VICs)
REST API
Cloud Tier
Physical Servers
VMware
App Servers
Database
Cohesity enables enterprises to take control of their increasingly complex storage environments through a hyperconverged
secondary storage infrastructure. The Cohesity DataPlatform can consolidate multiple use cases such as data protection,
test/dev, file services, and analytics onto a single web-scale platform. Today, 80% of enterprise data resides in siloed
secondary storage systems. These islands of storage create inefficiency, waste, and overhead expense. Cohesity consolidates
silos into a highly resilient and fully redundant web-scale storage fabric, dramatically improving IT productivity and reducing
long-term storage operating costs.
Cohesity DataPlatform, integrated with Pure Storage FlashArray//m, maintains all the value that comes from secondary
storage hyper-convergence while providing enterprise-class flash based storage for performance and reliability. FlashArray
//m is designed to deliver exceptional performance, expandability, and efficiency for storage and I/O-intensive infrastructure
workloads. This makes the FlashArray//m platform ideal for databases, real-time big data analytics, virtualization, and bare-
metal applications.
Cohesity complements FlashArray//m with a distributed platform architecture for data protection and data management
that is designed for high availability. The Cohesity nodes have a shared-nothing topology and there are no single points
of failure or inherent bottlenecks. Consequently, both performance and capacity can scale linearly as more physical nodes
are added to the cluster. The distributed file system spans across all nodes in the cluster and natively provides global
deduplication, compression, and encryption.
In addition, Copy Data Management is built into Cohesity’s DataPlatform. It allows for creation of zero-space instant
writeable clones from backups. Therefore, the same platform can provision and orchestrate test & development
environments. The entire platform is exposed through a single pane of glass that manages all secondary storage workflows
spanning multiple clusters, geographical sites, and the cloud.
By consolidating Tier 2-4 storage workflows into a single unified, scale-out web-scale platform, Cohesity reduces overall
TCO and improves business agility for enterprise customers. The solution eliminates fragmentation, significantly improves
storage efficiency, and, with its inbuilt MapReduce-based analytics, shines light into data which otherwise would have
remained dark.
Cohesity leverages VMware’s vSphere API for Data Protection and Change Block Tracking (CBT) mechanism to ensure
consistent and storage-efficient protection of VM data while keeping the data fully hydrated, indexed, and instantly
available.
Deep dives on the Cohesity architecture and SnapTree(TM) that supports these capabilities can be found here:
http://www.cohesity.com/wp-content/uploads/2015/10/Cohesity-Architecture-WhitePaper.pdf
https://www.cohesity.com/resource-assets/solution-brief/Cohesity-SnapTree-Solution-Brief.pdf
CBT is implemented at the VMware virtualization layer and can track virtual disk blocks that have been used and/or
changed since a previous snapshot. This allows for very efficient storage of incremental changes to the virtual disk reducing
both storage space as well as the overall time for backups to take place and in-turn replicate.
For further information and supported configurations for CBT, please visit:
https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1020128
By default, new VMs have CBT disabled. Cohesity’s DataProtect software will automatically detect this condition and enable
CBT for the VMs that are configured for protection, removing the need for administrators to track CBT status.
Non-CBT Backups
CBT backups are the most common form of Cohesity backups due to their space-efficient, fully-hydrated nature. However,
Cohesity also provides the ability to take non-CBT backups. A policy-based backup allows for non-CBT backups as well, to
provide an additional layer of data protection in case that is also required.
FlashArray//m’s performance can also make your business smarter by unleashing the power of real-time analytics,
driving customer loyalty, and creating new, innovative customer experiences that simply weren’t possible with disk. All by
transforming storage with FlashArray//m.
FlashArray//m leverages a chassis-based design with customizable modules, enabling both capacity and performance to be
independently improved over time with advances in compute and flash, to meet business needs today and tomorrow.
The FlashArray//M Operating Environment, Purity, provides a native, IP-based, asynchronous replication engine,
FlashRecover, which enables multi-site long-distance RPO. FlashRecover supports bi-directional replication, which facilitates
use of DR infrastructure for DR and production. FlashRecover leverages space-efficient, scalable snapshots to create point-
in-time replicas of volumes. The replication engine compares the last replicated snapshot to the new one and sends only
incremental data between the two snapshots, thus reducing network traffic. Snapshots are deduplicated for space efficiency.
Pure Storage FlashRecover snapshots are just new volumes with full capabilities. They inherit the data reduction
characteristics of their parent volume and can be restored instantaneously to any volume. However, this technology only
extends to the datastore level and recovery on a per file basis is not supported.
The full product line of the Pure Storage FlashArray//m can be seen in the below graphic.
Our test setup was configured to emulate a VMware Horizon 7 persistent full virtual machine environment. In a persistent
VDI environment, a virtual desktop is statically assigned to a single end-user and changes that user makes to the VM are
inherited between sessions. User data can be offloaded to an SMB share using a product like VMware Persona Management
or User Environment Manager, or managed via another 3rd party in instances where customers wish to manage their full
virtual machines via updates to the template. Worth noting is that Cohesity supports SMB/CIFS shares, making hosting user
data on the Cohesity appliance an option as well, with the same robust recovery operations. The data-reducing capabilities
of the Pure Storage array enable customers to use non-persistent and persistent desktops interchangeably without
significantly impacting the storage footprint required in either scenario. Full virtual desktops also have the advantage
of taking up far less array space, as they are more easily reduced than linked clones, as well as having significantly faster
creation times using VAAI.
A diagram depicting the entire VMware Horizon suite can be seen below.
OS Function Quantity
Not pictured below are physical Windows and Linux systems, which will be detailed in another section and were not used in
the VDI experiments.
The virtual machines listed above were split between two separate Pure Storage datastores. The 250 persistent desktops
and their template image under the ‘VDI’ cluster were hosted on a single 25TB datastore, while the infrastructure VMs, Login
VSI components, and Windows 10 template were hosted on a separate 5TB datastore for infrastructure components.
This environment can easily be linearly scaled to thousands of desktops by adding additional ESXi hosts and infrastructure
virtual servers (e.g. AD/DNS/DHCP, Horizon Connection Servers) in order to provide the necessary compute and load
balancing for the additional users. Since the purpose of this paper is to demonstrate the integration and workflows between
Pure Storage, Cisco UCS, and Cohesity, we elected to keep this ecosystem relatively small - though all actions performed
throughout can easily be leveraged for use with a much larger deployment and for workloads outside of, and mixed with,
VDI.
The Windows 10 user desktops included the standard set of Knowledge Worker tools used by Login VSI for simulating user
workflows, including: Office® 2013 Suite, Adobe Acrobat Reader®, Internet Explorer®, 7zip, and DoroPDF writer, amongst
others. We used the VMware Horizon OS Optimization tool to setup the parent VM properly.
From a networking perspective, two separate networks with two vNICs each (for failover) were used in our test environment:
one for ESXi host management traffic and Cohesity and Pure Storage integration, and a second network dedicated for
VDI desktop traffic between desktop VMs and the simulated Login VSI endpoint devices exercising the environment.
Segregating network traffic in this manner is an important consideration for both performance as well as resiliency concerns.
The screen capture below shows the configuration used for the persistent VDI desktop pool. Important to note is that for
these experiments we elected to use a persistent full virtual machine cloned from a template. All user state changes are
inherited between sessions using this model. Alternative methods for hosting user data such as Persona Management, User
Experience Manager, AppVolumes, and/or a third party tool would all be supportable options within this solution
As mentioned earlier, a vital piece of the test environment was setting up Login VSI to provide realistic, user-driven changes
to the end-user desktops that could then be protected and potentially restored by the Cohesity appliance in the event of
a VM or data file corruption. The diagram below shows the required connections and components we used to create a
realistic Horizon 7 production deployment. Login VSI provides the most realistic and industry-standard VDI benchmarking
available on the market. Further information about Login VSI can be found at their website: http://www.loginvsi.com.
With our Horizon 7 environment completed, the next step was to protect the various components, which will be detailed in
the following sections.
Further documentation about running VMware Horizon with FlashStack, including detailed best practices and Login VSI
performance testing results can be found at the following links:
http://www.cisco.com/c/en/us/td/docs/unified_computing/ucs/UCS_CVDs/ucs_flashstack_view62_5k.html
http://www.purestorage.com/content/dam/purestorage/pdf/whitepapers/FlashStackMini_for_VMware_Horizon.pdf
http://info.purestorage.com/rs/225-USM-292/images/Design%20Guide%20for%20Horizon%20View%206.2.pdf
In addition, please visit the FlashStack Cisco Design Zone to reference other Cisco Validated Designs and use cases for
FlashStack that this solution integrates with seamlessly:
http://www.cisco.com/c/en/us/solutions/enterprise/data-center-designs-cloud-computing/flash-stack.html
More detailed intelligence and array administration are easily accessible through the various tabs across the top of the GUI
and separately via Pure 1 manage. Pure Storage also features a restful API, enabling full integration with a wide ecosystem
of scripting languages and products. Pure Storage provides a full-featured vSphere web client plug-in that allows single
pane of glass administration for one or multiple arrays through the vSphere web client.
Advanced system details are available by clicking any one of these tabs.
We can initiate recovery workflows for files or folders from VMs or physical servers being protected, VM level recovery,
Microsoft SQL point in time database recoveries, and Pure Storage volume level restores. We will look at each of these
recovery options in detail in the following sections.
The next section discusses the Pure Storage FlashArray//m REST API integration with the Cohesity data protection and data
management platform.
Figure 24 : Cohesity Job and Policy definition for Pure //m Flash Array protection
©2017 Cohesity, All Rights Reserved 24
Figure 25 : Protected volume summary
The restore operation initiates the following background tasks to recover the Pure Storage Array//m volume back to the
original array or another array.
Restore
Step 1: Verify that the volume that is to be recovered does not already exist on the Pure array.
Step 2: Clone the backed up volume snapshot files on SnapFS to a new “recover view” on SnapFS.
Step 3: Fetch metadata about the backed up volume snapshot from the data on SnapFS.
Step 4: C
reate a new recovery volume on Pure (this is the volume where the backed up snapshot data will
be copied to).
Step 5: Connect the Cohesity nodes to the volume created in step (4).
Step 6: Back up all the data from the snapshot volume on SnapFS to the recovery volume.
A. A bridge node is assigned to do this data transfer.
B. This bridge node will do an iSCSI discovery on the Pure array.
C. It will then connect to the iSCSI target (if it’s not already connected).
D. It will then figure out which local block device corresponds to the volume that needs to be restored.
E. It will then proceed to copy the relevant bytes from SnapFS and write it to the local block device.
F. It will disconnect from the iSCSI target at the end of the restore if necessary.
Step 7 : After restore completes, the Cohesity nodes are disconnected from the volume created in step (4).
Select the respective point of recovery from the available recovery points for the volume.
Once the volume has been recovered, the task of connecting hosts can be performed.
To re-attach a recovered Pure Storage LUN to vSphere, the process is very straightforward:
1. From within the Pure Storage GUI (standalone or vSphere web client plug-in), attach Host and/or Host Groups to
the restored Volume.
2. Within the vSphere Web Client, re-add and re-signature the volume.
3. Re-add the virtual machines to the vSphere inventory.
Option 2 :
Search for files folders by browsing specific servers or desktops.
Each of these recovery options would initiate the following background tasks to recover files or folders.
Step 1: User manually triggers a file/folder recovery task either, by searching the files through the elasticsearch
database or via browsing VMs and its volumes.
Step 2: C
ohesity creates an internal view and clones the VM snapshot and mounts the view to the target ESXi
host(s).
Step 3: Cohesity attaches the cloned VMDK files to the target VM to which the files are being recovered.
Step 4: Cohesity deploys a helper utility onto the VM and triggers the restore process.
Step 5: The restore helper utility performs file copy from the attached disks (originally from the backup) onto the
recovery location. The utility additionally preserves the file attributes and other properties based on user
preferences.
Step 6: Once the file/folder copy completes, the disks are detached from the VM.
Step 7: Cohesity requests the datastore to unmount.
Step 8: ESXi host unmounts the datastore.
Step 9: Cohesity releases the view.
Here are the two recovery options presented in the Recover VM workflow:
Add the selected VM to the Shopping Cart, and then search for other VMs, if required, to make multiple VMs part of the
same recovery operation.
Figure 43©2017
: VMCohesity,
recovery All workflow - continued
Rights Reserved 35
Continue on with recovery to choose other recovery options.
Once the admin clicks on Finish, the following tasks are performed for a successful recovery of the VM.
Step 1: U
ser manually triggers a Cohesity VM recovery task and selects snapshot, target, networking settings, VM
name, and target datastore.
Step 2: Cohesity contacts VMware endpoint to validate current inventory and chosen recovery task settings.
Step 3: Cohesity creates an internal view, clones the VM snapshot, and mounts the view to the target ESXi host(s).
Step 4: C
reate a new VM object using the original VM configuration file and chosen recovery settings. Network
configuration changes take place at this step.
Step 5: VM is (optionally) powered on. (Note that the VM is now available for use.)
Step 2: T
he master distributes the task of backing up individual servers across different slaves. Essentially, all the
nodes in the cluster take part in the backup. If any of the nodes goes down either due to failure or due to
upgrade, the underlying architecture allows seamless migration of the tasks to a different node and does not
affect the backup runs.
Step 3: T
he Cohesity cluster contacts Cohesity agent running on the physical server, collects information regarding
volumes, and checks health.
Step 4: The Cohesity cluster then triggers a snapshot on the server via the agent.
Step 5: T
he Cohesity agent performs environment-specific steps. 5 {Windows} On Windows this results in a VSS
snapshot being taken. Additionally, it will also contact the Cohesity change block tracking driver to fetch the
blocks changed since the last backup in case of incremental backup. 5 {Linux} On Linux this triggers an LVM
snapshot of the volumes.
Step 6: C
ohesity backs up volumes from the server in a parallel fashion, distributing across several Cohesity nodes
but limiting to a number so that the primary remains unaffected.
Step 7: The volumes backed up will be saved as fully hydrated VHD/VHDx files onto Cohesity’s view.
Step 8: O
nce the snapshot is backed up, the Cohesity cluster will indicate to the agent that the backup completed
successfully.
Step 11: The Cohesity cluster proceeds to index the files on the VHD/VHDx files and wraps up the backup job run.
In four simple steps the agent is installed on the physical server. The server requires a reboot once the agent has been
installed as the agent enables CBT on the physical volumes.
Individual drive letters/volumes can be selected for data protection, or by default all volumes are protected.
Step 1: U
ser selects the files that he wishes to restore, either searching for the files or browsing the volumes on the
Cohesity UI, and also specifies preferences (whether to recover to original location or alternate, etc.)
Step 2: Additionally the user selects the snapshot (backup job run) from which to restore the files.
Step 3: T
he Cohesity cluster creates an internal view, clones the relevant VHD/VHDx files onto a Cohesity view, and
instructs the Cohesity agent to mount the View.
Step 4: O
n a Windows server, the Cohesity agent will mount the view as an SMB (or CIFS) share. It will be mounted
as an NFS mount on a Linux server.
Step 5: T
he Cohesity cluster now instructs the agent to mount the VHD/VHDx files as volumes within the physical
server.
Step 6: O
n Windows, the Cohesity agent uses Windows native support for VHD/VHDx files and mounts them as
volumes. The Linux Cohesity agent will mount them as filesystems.
Step 7: The Cohesity cluster now instructs the agent to restore files based on the user’s preferences.
Step 8: O
nce file restore completes, the Cohesity cluster instructs the agent to unmount the VHD/VHDx volumes
and also unmount the SMB/NFS mounts.
Step 9: The Cohesity cluster releases the internal view.
See below the workflow of the option to mount the volume to a physical server, so files or folders can be restored from the
volume as the admin deems necessary.
On the Windows server, browsing to the above directory provides the mounted volume and all the files contained
in the volume.
Now that the VM has been registered as a MS SQL server, we are able to protect this server using the Cohesity MS SQL
adapter. This adapter provides a point-in-time backup of the SQL server by backing up the database and the log backups.
Select MS SQL to use the Cohesity native adapter to create a data protection job for the View Composer SQL server that
was registered above.
Once the server is selected, the next step is to associate it with the SQL Policy.
In this case, the SQL database policy is set to protect the SQL server DB every 1 hour and, in addition, to capture logs every
15 mins. This combination of backups enables “point in time” restore.
SQL Restore
In this section, we will simulate the VMware Horizon Composer SQL database being accidently deleted and then restore it
using the built-in SQL Restore function of the Cohesity appliance.
In the screen capture below we can see the Composer Database (named vc) in SQL Server Management Studio.
Moving over to the Composer Server, we can see that the ODBC connection to the vc database now
fails and that the Horizon Composer service is unable to start.
Furthermore, we can see in the Horizon Connection server that the Composer server is offline, as the service is not able to
start due to the database being deleted. Already problem desktops are appearing.
To recover from this traditionally major outage, we move to the Cohesity appliance and begin the MS SQL
recovery operation.
As this database was fully backed up every hour, with incremental logs backed up, we have the ability to restore the SQL
database to any point over a very wide time interval. Since the database was not facing a corruption issue but instead
accidental deletion in the example, we elected to recover it to as recent a time as possible.
The recovery job progress can be monitored from the Cohesity appliance.
After about five and a half minutes, the SQL database has been restored.
Looking back atthe SQL server, the vc database has been recovered within the SQL Management Studio console.
This enables us to start the VMware Horizon 7 Composer service successfully, confirming that the ODBC connection is
operational and all problematic VDI desktops are resolved within the Connection Server console.
Conclusions
This integration paper represents the initial documentation of the value that our joint solution provides to the customer.
Subsequent papers will dive deeper into other common FlashStack use cases and provide an extended look at some
functionalities that were only covered briefly here.
We have demonstrated the seamless integration and ease by which today’s modern data center engineer can deploy,
manage, and recover a FlashStack Converged Infrastructure solution. Few workloads change as dynamically as a production
VDI environment – and few demand the flexibility to recover both relatively minor items, such as a single user file, as well as
significant components of the server infrastructure. We have proven that this is not only possible with the outlined solution
but that it is also optimized to provide maximum performance at scale, with near instant recovery of any piece of the
solution.
In this paper, the following backup and recovery procedures were covered:
Pure Storage and Cohesity together deliver the best of both worlds: all-flash performance for your mission-critical
applications and flash-driven consolidation of your secondary storage needs, including data protection with instant recovery
and an agile method for repurposing data for test/dev and file services.
Authors:
Kyle Grossmiller - Solutions Architect - Pure Storage
Damien Philip - Principal Solutions Architect - Cohesity
Reviewers :
Mayur Dewaikar - Product Management - Pure Storage
Vivek Agarwal - Business Development - Cohesity
Ravi Venkat - Sr. DataCenter Architect - Pure Storage
Raymond Mar - Strategic Alliance Manager - Pure Storage
Cohesity, Inc.
Address 451 El Camino Real, Santa Clara, CA 95050 @cohesity
Email contact@cohesity.com www.cohesity.com ©2017 Cohesity.©2017
All Rights Reserved.
Cohesity, All Rights Reserved
58.58