Storage efficiency has typically referred to the processes resulting in reduced storage and bandwidth capacity requirements. Compression, thin provisioning, data deduplication and even storage virtualization have had a huge impact on storage, IT efficiency and, ultimately, the total cost of ownership (TCO) of enterprise storage. These technologies are pervasive in data center services, such as production storage, backup, WAN optimization, and archiving.
In today’s post-virtualization data center, virtualized workloads with different IO streams are sharing the physical resources of the host. This results in random IO streams competing for resources—and the emergence of new efficiency requirements as the IOPS required to service virtual workloads has increased. Some band-aids applied to the IOPS problem include overprovisioning HDDs or investment in SSDs/flash—with both contributing to a higher cost per gigabyte of storage allocated to each virtual machine (VM).
Enter data efficiency. Data efficiency capitalizes on the familiar storage efficiency technologies, such as compression and deduplication, but executes them in a way that positively impacts both capacity AND IOPS—and costs—in today’s modern data centers.
SimpliVity delivers data efficiency in its OmniStack software, the foundation of its OmniCube hyperconverged infrastructure. The enabling technology is the Data Virtualization Platform (DVP). OmniStack DVP deduplicates, compresses and optimizes all data inline (before data is written to HDD) and maintains that state across all tiers of the infrastructure and all stages of the data lifecycle.
To help you understand data efficiency better, this blog series is designed to detail how data centers have evolved in the 21st century, what problems need to be addressed today, how using “yesterday’s technology” or “yesterday’s approaches” cannot solve the problem, and why it’s worthwhile to solve the problem. The series will also provide a foundational education about the technologies being applied by SimpliVity in order to arm IT professionals in making more informed decisions regarding data center infrastructure.
The next blog in the series will look at the data problem.
Here’s a rundown on the whole series … check back for live links!
- Data Efficiency Overview
- Defining The Data Problem
- Deduping IO
- Hardware-Assisted Deduplication
- Advantages of Data Efficiency: WAN Benefits
- Advantages of Data Efficiency: Cache Benefits
- Advantages of Data Efficiency: Data Protection Benefits
- Data Efficiency Across the Lifecycle
- Last Words on Data Efficiency