Capacity optimization
From Wikipedia, the free encyclopedia
- See also: Deduplication
Capacity optimization technologies are similar to data compression technologies, but they look for redundancy of very large sequences of bytes across very large comparison windows. Typically using cryptographic hash functions as identifiers of unique sequences, long (8KB+) sequences are compared to the history of other such sequences, and where possible, the first uniquely stored version of a sequence is referenced rather than stored again.
Capacity optimization generally refers to the use of this kind of technology in a storage system. An example of this kind of system is the Venti file system [1] in the Plan9 open source OS. (Source code is available here). There are also implementations in networking (especially Wide Area networking), where they are sometimes called bandwidth optimization technologies. [2]
Commercial implementations of capacity optimization are most often found in backup/recovery storage, where storage of iterating versions of backups day to day creates an opportunity for reduction in space using this approach. The term was first used widely in 2005. [3]
Other non-vendor terms to describe this technology have included deduplication or factoring.