Sun Solaris To Linux: Size Matters After All

Adding a 128-bit file system to Solaris 10 will create "billions and billions" of unique addresses that can be used for storing data and files.

Charles Babcock, Editor at Large, Cloud

May 1, 2006

3 Min Read

Sun Microsystems is about to up the ante in the open-source race, adding a 128-bit file system to its Solaris 10 operating system so that users can leave behind any worries about ever exhausting their virtual memory space. By having a bigger bit count than everyone else, the number of unique addresses that can be created by the system for storing data and files can be larger than anyone else's. Virtual memory assumes an unlimited number of physical storage devices is available. What's needed are unique identifiers for the location of each file or data block stored, so the constraint on virtual memory is how many of those unique addresses can be created.

Linux, Microsoft Windows Vista, and most Unix systems employ either 32-bit or 64-bit file systems, which for the next 10 years are likely to be able to handle the needs of the largest computer systems.

To understand how big 32-bit systems really are, consider that 20-bit systems, once considered state-of-the-art, can create one million unique addresses. A 24-bit system increases the number of virtual addresses to 16 million. The jump to 32-bit is a leap to four billion unique addresses, with the ability to retrieve a file or data stored at any of the locations.

The shift to 64-bit file systems, currently under way, represents a quantum jump that many experts believe will serve existing systems as long as they can run. But Sun is talking about a 128-bit file system, which it concedes is unlikely to be needed for the next 10 years, when 64-bit systems start to run out of steam.

In terms of addressable virtual memory, 128-bit systems represent "a very large number," says Chris Ratcliffe, director of marketing for Solaris 10. The number of unique addresses that can be created by a 128-bit system is something like the late Carl Sagan's answer to the question of how many stars are there--billions and billions.

Sun calls Solaris' new feature the Zettabyte File System. A zettabyte is one sextillion bytes--which is a very large number. Because of that, as Solaris 10 is upgraded with its next release in June, it will offer a file system "with virtually unrestricted capacity," says Ratcliffe. A computer running ZFS could address "all the disks currently on the planet," he says.

But Sun also says size doesn't matter.

What's really important, Ratcliffe says, is that ZFS will have built-in data integrity and disk volume management capabilities, instead of requiring its users to add on storage software that tries to overcome the file system's shortcomings.

Through its built-in checksum capability, ZFS will assign a 64-bit checksum or bit count associated with data at a particular address. If the actual bit count doesn't match the checksum, the system has detected what is probably corrupted data, and recovery measures can be implemented. The checksum approach gives data stored by ZFS a data integrity rate close to 100% or "19 nines," in Ratcliffe's words.

ZFS will also include a copy-on-write mechanism that allows a backup copy of data to be available the instant a data error is detected, he says.

About the Author(s)

Charles Babcock

Editor at Large, Cloud

Charles Babcock is an editor-at-large for InformationWeek and author of Management Strategies for the Cloud Revolution, a McGraw-Hill book. He is the former editor-in-chief of Digital News, former software editor of Computerworld and former technology editor of Interactive Week. He is a graduate of Syracuse University where he obtained a bachelor's degree in journalism. He joined the publication in 2003.

Never Miss a Beat: Get a snapshot of the issues affecting the IT industry straight to your inbox.

You May Also Like

More Insights