Research: Data Deduplication

Oct 05, 2009


Squeeze Play: The Expanding Profile of Data Deduplication

Want proof that data deduplication is hot? Look no further than the recent bidding war between NetApp and EMC over dedupe specialist Data Domain. After a deal with NetApp seemed set, EMC swooped in and tendered more than $33 per share for Data Domain, for a total bid of $2.1 billion. That beat NetApp's best offer of $1.9 billion.

What recession?

The fact is, EMC knows CIOs always find a way to free up cash when a problem threatens to spiral out of control. And if you look up "out of control" in the dictionary, you might just see a picture of a bursting-at-the-seams SAN. In November 2008 we launched our most recent InformationWeek Analytics State of Storage poll, fielded in conjunction with Byte and Switch. Of the 328 business technology professionals we surveyed for that report, 25% said they manage less than 1 TB of  data. Fast forward to June 2009: Just 10% of the 437 respondents to our data deduplication survey manage less than 1 TB. Percentages at the very top end remain reasonably steady, but many more organizations seem on a fast track to the petabyte mark.

There are multiple technologies on the table that vendors say will help us get a handle on our storage, from both a security/compliance angle as well as a growth perspective. These include thin provisioning, data deduplication, improved compression methodologies, disk-to-disk-to-tape backup, encryption, massive array of idle disks (MAID), virtual tape libraries, larger disk drives (up to 2 TB at the time this article was written), and faster performing drives such as SSDs. In this report, we’ll touch on many of these and help sort out whether deduplication in particular makes sense in your environment.

Research Report