Click me
Transcribed

Why Deduplication Is Becoming a Must Have in Storage Environments

WHY DEDUPLICATION IS BECOMING A MUST HAVE IN ENTERPRISE STORAGE ENVIRONMENTS "Explosive data growth and increasingly stringent regulatory compliance laws demand that companies of all sizes look for storage solutions that not only have an extensive feature set but also are cost effective." – IDC Whitepaper DEMAND FOR ENTERPRISE ALTHOUGH ENTERPRISE STORAGE STORAGE HAS INCREASED RAPIDLY PRICES HAVE FALLEN -20% IN THE LAST 5 YEARS YEAR-OVER-YEAR Unit cost of storage hardware per gigabyte, $ Storage demand, cumulative sum in thousands of petabytes 20 4.5 4.0 15 100 3.5 3.0 10 2.5 2.0 1.5 1.0 2006 2007 2008 2009 2010 2011 .5 Petabyte = 1 quadrillion bytes which is enough space for about 20,000 Blu-ray movie rips. 2006 2007 2008 2009 2010 2011 "Primary data deduplication is proving to have the most dramatic impact on data footprint reduction and should be seriously considered." – IDC Whitepaper A HYPOTHETICAL CASE STUDY Assuming the cost per GB of storage (including system, software, RAID overhead, etc.) is $2 - $3 per GB, a 50 TB enterprise storage system would cost roughly $100,000. Let's see how Deduplication can cut that cost dramatically: OTHER USAGE OTHER USAGE OTHER USAGE VIRTUALIZATION FILE SHARE, ETC. EMAIL/STRUCTURED DATA FILE SHARE, ETC. DEDUP'D VIRTUALIZATION EMAIL/STRUCTURED DATA DEDUP’D ORIGINAL STORAGE REMAINING STORAGE REMAINING STORAGE ALLOCATION OF 50 TBs ALLOCATION OF 30 TBs ALLOCATION OF 20 TBs Assuming half of the 50 TBs of storage are being used to support a virtual server infrastructure, the level of redundancy is VERY HIGH For home directories / file shares, let's assume 15 TBs are assigned for that. The level of redundancy is less than that of a virtualized server, but a 3x deduplication efficiency rate is not uncommon. This means 15 TB of the required storage capacity could be reduced to 5 TB. For semi-structured data like email and structured data like databases represent 5 TBs of the storage capacity. For this set of data, you can expect a modest deduplication efficiency rate of 1.5:1 or 2:1. This means 5 TB of the 50 TB storage capacity could be reduced to 3.3 TB. in these environments. It is not uncommon to expect a 5x or greater level of storage efficiency enabled by deduplication. This means 25 of the 50 TB storage capacity could be reduced to 5 TB. ORIGINAL STORAGE REQUIREMENTS & COST: 50TBS @ $100,000 REQUIREMENT & COST AFTER DEDUPLICATION: 18.3TBS a $37,000 NET SAVINGS ON HYPOTHETICAL SCENARIO: $63,000 SOURCES: https://www.mckinseyquarterly.com/Meeting_the_demand_for_data_storage_2153 http://www.storageswitzerland.com/Articles/Entries/2012/12/3_The_ROI_of_ Primary_Storage_Deduplication.html http://www.tegile.com/resources/de-duplication-not-just-for-backup-anymore tegile WHY DEDUPLICATION IS BECOMING A MUST HAVE IN ENTERPRISE STORAGE ENVIRONMENTS "Explosive data growth and increasingly stringent regulatory compliance laws demand that companies of all sizes look for storage solutions that not only have an extensive feature set but also are cost effective." – IDC Whitepaper DEMAND FOR ENTERPRISE ALTHOUGH ENTERPRISE STORAGE STORAGE HAS INCREASED RAPIDLY PRICES HAVE FALLEN -20% IN THE LAST 5 YEARS YEAR-OVER-YEAR Unit cost of storage hardware per gigabyte, $ Storage demand, cumulative sum in thousands of petabytes 20 4.5 4.0 15 100 3.5 3.0 10 2.5 2.0 1.5 1.0 2006 2007 2008 2009 2010 2011 .5 Petabyte = 1 quadrillion bytes which is enough space for about 20,000 Blu-ray movie rips. 2006 2007 2008 2009 2010 2011 "Primary data deduplication is proving to have the most dramatic impact on data footprint reduction and should be seriously considered." – IDC Whitepaper A HYPOTHETICAL CASE STUDY Assuming the cost per GB of storage (including system, software, RAID overhead, etc.) is $2 - $3 per GB, a 50 TB enterprise storage system would cost roughly $100,000. Let's see how Deduplication can cut that cost dramatically: OTHER USAGE OTHER USAGE OTHER USAGE VIRTUALIZATION FILE SHARE, ETC. EMAIL/STRUCTURED DATA FILE SHARE, ETC. DEDUP'D VIRTUALIZATION EMAIL/STRUCTURED DATA DEDUP’D ORIGINAL STORAGE REMAINING STORAGE REMAINING STORAGE ALLOCATION OF 50 TBs ALLOCATION OF 30 TBs ALLOCATION OF 20 TBs Assuming half of the 50 TBs of storage are being used to support a virtual server infrastructure, the level of redundancy is VERY HIGH For home directories / file shares, let's assume 15 TBs are assigned for that. The level of redundancy is less than that of a virtualized server, but a 3x deduplication efficiency rate is not uncommon. This means 15 TB of the required storage capacity could be reduced to 5 TB. For semi-structured data like email and structured data like databases represent 5 TBs of the storage capacity. For this set of data, you can expect a modest deduplication efficiency rate of 1.5:1 or 2:1. This means 5 TB of the 50 TB storage capacity could be reduced to 3.3 TB. in these environments. It is not uncommon to expect a 5x or greater level of storage efficiency enabled by deduplication. This means 25 of the 50 TB storage capacity could be reduced to 5 TB. ORIGINAL STORAGE REQUIREMENTS & COST: 50TBS @ $100,000 REQUIREMENT & COST AFTER DEDUPLICATION: 18.3TBS a $37,000 NET SAVINGS ON HYPOTHETICAL SCENARIO: $63,000 SOURCES: https://www.mckinseyquarterly.com/Meeting_the_demand_for_data_storage_2153 http://www.storageswitzerland.com/Articles/Entries/2012/12/3_The_ROI_of_ Primary_Storage_Deduplication.html http://www.tegile.com/resources/de-duplication-not-just-for-backup-anymore tegile

Why Deduplication Is Becoming a Must Have in Storage Environments

shared by jessief on Mar 30
279 views
1 shares
0 comments
The inspiration behind this exercise was the growing adoption of virtual desktop infrastructure (VDI) in enterprise business environments. Virtual server environments tend to have a very high level of...

Publisher

Tegile

Category

Technology
Did you work on this visual? Claim credit!

Get a Quote

Embed Code

For hosted site:

Click the code to copy

For wordpress.com:

Click the code to copy
Customize size