Backup and recovery is the process of duplicating data and storing it in a secure place in case of loss or damage, and then restoring that data to a location—the original one or a safe alternative—so it can be used again in operations. Ideally, this backup copy (often called a snapshot) is immutable—meaning it cannot be altered after it is created to protect against mutations such as ransomware. Backup and recovery is also a category of onsite and cloud-based technology solutions that automate and support this process, enabling organizations to protect and retain their data for business and compliance reasons.
Backups are often bucketed into three categories:
The key difference between backup and recovery is that the backup process is how you save and protect your production data and safely store it away so you have it for a later time, when you might need to use it.
Recovery is the process whereby you retrieve and restore that backup data to your production systems to avoid downtime.
Reliable backups and fast recovery together ensure business continuity and business resilience.
The amount of data organizations create, capture, and store has skyrocketed over the last decade. And analysts anticipate the amount of new data generated will grow at more than 50% compounded annually.
Because enterprises and people are storing data in more places, new categories of backup data recovery have emerged. These include:
For enterprises, a disaster is when a catastrophic event negatively impacts your people and/or your data. The event can be natural—a hurricane taking down a data center, for example—or human-made, such as a ransomware attack. Regardless of whether disasters are due to human error, hardware failure, malicious attacks, or natural events, the result is the same—data corruption or lost data that makes it difficult to ensure business continuity.
Disaster recovery is the process your IT organization goes through to restore data. Increasingly, organizations are setting aside a complete or full backup of entire environments—either on-premises or in the public cloud—to ensure all of their data can be made available quickly in the event of a catastrophe. Having a way to quickly recover lost or damaged data is crucial for business continuity, and using cloud storage in disaster recovery planning is ideal for backing up essential business data.
All of the data sources that your organization protects may at some time need to be recovered due to a data loss event. These include:
Data is essential to organizations of all types and sizes. You need a robust data backup and disaster recovery plan because it provides a roadmap for the people responsible for taking charge in a disaster scenario to know who is doing what and in what sequence to restore operational functionality. Your DR plan should include both people and processes, serving as a guide for employees to follow as they bring your business back up.
A robust data backup and disaster recovery plan should also ensure that your data is always protected—as and after you move it from day-to-day production systems for short—and long-term retention. With the best backup and disaster recovery plan, you will always have your data readily available should you need it.
Imagine if the data needed to operate your business, department, or agency was unavailable, even for a few minutes, never mind hours, days, or weeks. Customers would be unhappy. Employees would be, too. And in the case of ransomware, your entire business might even cease to exist. Effective backup and recovery of important data prevents all of these scenarios.
Yes, data deduplication is absolutely important in backups. Here’s why. Data is growing exponentially and organizations are retaining more data—for marketing, compliance, and more—than ever. Because of this, IT teams need to deploy techniques that will help their organizations reduce data footprints, keeping costs lower.
Advanced data reduction through deduplication enables more data to fit into the same hardware space—helping to reduce cost.
The most powerful and flexible global deduplication architecture is variable-length data deduplication technology that spans an entire cluster across various data sources rather than simply a single node, resulting in significant savings across the entire storage footprint.
With variable-length deduplication, the size is not fixed. Instead, the algorithm divides the data into chunks of varying sizes based on the data characteristics. The chunks are cut in a data-dependent way that results in variable-sized chunks and greater data reduction than fixed-size deduplication. The efficiency benefit of variable-length deduplication compounds over time as additional data is retained.
Integrated data compression adds a boost. Compression works well on a single file, but across files, there is a need for some macro-level data compression. Why? Because when two identical copies of a file are stored, compression can individually compress the files while deduplication can completely eliminate the need to store any data for the second copy. So adding compression to the deduped data further reduces data size.
This works by finding small byte patterns common between the deduplicated blocks. Based on the type of data being ingested, compression can provide no benefit for encrypted or random data or up to 5–10x compression for common log files. Deduplication ratios for VMs, databases, and file shares belong somewhere belong between that range.
Data powers your organization and your competitive advantage. That’s why backup and recovery is important. With a robust data backup and recovery strategy — and technology solution — in place, your organization can:
Modern, Comprehensive Backup and Recovery | Traditional Backup and Recovery |
---|---|
Low (or no) capital costs. Modern backup solutions are typically a single platform with low or no on-prem infrastructure footprint, keeping backup and recovery costs low. | High capital costs. Often IT must cobble together multiple, costly infrastructure point products for data backup which raises costs. |
Fast, accurate backups. Modern backup eliminates data silos and automates operations for faster, more accurate backups than traditional approaches. | Slow, error-prone backups. Traditional backup contributes to mass data fragmentation — having siloed data that requires manual operations and leads to greater backup errors than modern approaches. |
Set and forget policies. Once IT staff creates and approves policies, they are easily and automatically added to data sources as servers join the network. | Tedious policy setting. IT staff must create and manage a unique policy for each data source as it is added to the network. And if a server is added without IT being notified, the business risks data not being backed up. |
Instant and predictable recovery. Modern backup minimizes data loss and provides predictable recovery assurance with restores at scale, and to any point in time. | Unpredictable recovery. Traditional backup can be slow and error-prone, often bleeding into production time. |
Unlocks business value through complete data visibility. Because there are no longer data silos and all backups are completed on one platform, IT can see and gain insights from all of enterprise data and apps. | No access to business insights because data is dark or hidden. Because backups are completed using many products and data can easily be lost, IT has dark data that makes it impossible to use for business insights. |
Ransomware protection. Modern backups feature immutable snapshots and have minimal data center footprints, reducing attack surfaces. | Ransomware protection. Traditional backups do not include immutable snapshots and have large data center footprints, widening attack surfaces. |
The single biggest challenge with trying to put an enterprise-wide backup and recovery strategy in place is that data typically resides in numerous places: in on-premises systems, clouds, and at the edge. Mass data fragmentation from siloed hardware and software and incomplete visibility into enterprise data means that time that should be spent on business innovation is wasted managing and maintaining disconnected point solutions.
Cohesity provides a backup and recovery solution that converges multiple point products and backs up data whether it is stored on-prem, at the edge, or in the public cloud on a single multicloud data platform. By taking a complex operation and simplifying it for businesses, Cohesity ensures business continuity, minimizes data loss, and reduces the total cost of ownership (TCO).
As a Gartner Magic Quadrant Data Center Backup and Recovery Solutions, Cohesity’s business is the data business. Cohesity radically simplifies how organizations manage their data everywhere and derive more value from that data.
Organizations that choose Cohesity enjoy: