ExaGrid's Scalable Data Deduping NAS

Aiming for the SME market, ExaGrid System's built a line of data deduping NAS appliances with 5 models designed to protect from 1 Tb to 5 TB of source data. A year ago a vendor coming out with a deduping NAS would have been noteworthy on its own, but this market moves fast and I'm not that easily impressed any more. What makes the ExaGrid boxes intriguing now is that you can stack up to 5 appliances into a single grid with 34 TB of disk space and a data ingestion rate of over 2 TB per hour.

Howard Marks, Network Computing Blogger

March 25, 2008

1 Min Read
InformationWeek logo in a gray background | InformationWeek

Aiming for the SME market, ExaGrid System's built a line of data deduping NAS appliances with 5 models designed to protect from 1 Tb to 5 TB of source data. A year ago a vendor coming out with a deduping NAS would have been noteworthy on its own, but this market moves fast and I'm not that easily impressed any more. What makes the ExaGrid boxes intriguing now is that you can stack up to 5 appliances into a single grid with 34 TB of disk space and a data ingestion rate of over 2 TB per hour.ExaGrid takes a unique approach to data deduplication. Rather than take an arbitrary backup stream and break it into blocks, ExaGrid's software knowing the format of the backup files created by popular backup applications, including Backup Exec, ARCserve, and Galaxy, identifies each file that's been backed up and compares it with the version of that file from previous back ups. It then replaces the previous version of the file with a delta file describing the changes from the newest version. This preserves the latest versions intact, speeding restores from the latest back up. The down side to this approach is it can't eliminate duplicate data that appears in multiple files or even duplicate files.

Like appliances from competitors Quantum and Data Domain, the ExaGrid appliances can replicate deduped data from a branch office back to another ExaGrid appliance or grid. The data replication is delayed until after the post-process data deduplication completes, which could be hours after the back-up job runs.

About the Author

Howard Marks

Network Computing Blogger

Howard Marks is founder and chief scientist at Deepstorage LLC, a storage consultancy and independent test lab based in Santa Fe, N.M. and concentrating on storage and data center networking. In more than 25 years of consulting, Marks has designed and implemented storage systems, networks, management systems and Internet strategies at organizations including American Express, J.P. Morgan, Borden Foods, U.S. Tobacco, BBDO Worldwide, Foxwoods Resort Casino and the State University of New York at Purchase. The testing at DeepStorage Labs is informed by that real world experience.

He has been a frequent contributor to Network Computing and InformationWeek since 1999 and a speaker at industry conferences including Comnet, PC Expo, Interop and Microsoft's TechEd since 1990. He is the author of Networking Windows and co-author of Windows NT Unleashed (Sams).

He is co-host, with Ray Lucchesi of the monthly Greybeards on Storage podcast where the voices of experience discuss the latest issues in the storage world with industry leaders.  You can find the podcast at: http://www.deepstorage.net/NEW/GBoS

Never Miss a Beat: Get a snapshot of the issues affecting the IT industry straight to your inbox.

You May Also Like


More Insights