Storage@home
Original author(s) | Adam Beberg |
---|---|
Developer(s) | Stanford University / Adam Beberg |
Initial release | 2009-09-15 |
Stable release | 1.05 / 2009-12-02 |
Operating system | Microsoft Windows, Mac OS X, Linux[1] |
Platform | x86 |
Available in | English |
Type | Distributed Storage |
License | Proprietary |
Website | http://en.fah-addict.net/articles/articles-6.php |
Storage@home was a distributed storage infrastructure designed to store massive amounts of scientific data across a large host of volunteer machines.[2] The project was developed by some of the Folding@home team at Stanford University,[3] and is currently inactive.
Function
Scientists such as those running Folding@home have to deal with massive amounts of data, which must be stored and backed up, and this is very expensive.[2] Traditionally, methods such as storing the data on RAID servers are used, but these become impractical for research budgets at this scale.[3] The Pande Group already has to deal with reliably storing hundreds of terabytes of scientific data, and this continually growing.[2] Adam Beberg and Vijay Pande took experience from Folding@home and began work on Storage@home.[3] The project is designed based on the Cosm Distributed File System (Cosm FS), and the workload and analysis needed for Folding@home results.[3] While Folding@home volunteers can easily participate in Storage@home, much more disk space is needed from the user than Folding@home, to create a robust network. Volunteers each donate 10 GB of storage space, which would hold encrypted files.[3] These users gain points as a reward for reliable storage. Each file saved on the system is replicated four times, each spread across 10 geographically distant hosts.[3][4] Redundancy also occurs over different operating systems and across time zones. If the servers detect the disappearance of an individual contributor, the data blocks held by that user would then be automatically duplicated to other hosts. Ideally, users would participate for a minimum of six months, and would alert the Storage@home servers before certain changes on their end such as a planned move of a machine or a bandwidth downgrade. Data stored on Storage@home is maintained through redundancy and monitoring, with repairs done as needed.[3] Through careful application of redundancy, encryption, digital signatures, automated monitoring and correction, large quantities of data could be reliably and easily retrieved.[2][3] This ensures a robust network that will lose the least possible data.[4]
Storage Resource Broker is the closest storage project to Storage@home.[3]
Status
Storage@home was first launched on September 15, 2009 in a testing phase. It first monitored availability data and other basic statistics on the user's machine, which would be used to create a robust and capable storage system for storing massive amounts of scientific data.[5] However, in the same year it became inactive, despite initial plans for more to come.[6] On April 11, 2011 Professor Vijay Pande stated that "We currently do not have any active plans with Storage@home. We're concentrating on other areas at the moment."[7]
See also
External links
References
- ↑ "Storage@home Installation". 2009-09-12. Retrieved 2011-09-17.
- 1 2 3 4 "General Information about Storage@home". 2009. Retrieved 2011-09-17.
- 1 2 3 4 5 6 7 8 9 Adam L. Beberg and Vijay S. Pande (2007). "Storage@home: Petascale Distributed Storage". Parallel and Distributed Processing Symposium, IEEE International: 1–6. doi:10.1109/IPDPS.2007.370672.
- 1 2 "The plan for splitting up data in Storage@home". 2009. Retrieved 2011-09-17.
- ↑ Vijay Pande (2009-09-15). "First stage of Storage@home roll out". Retrieved 2011-12-14.
- ↑ "Storage@home FAQ". 2009. Retrieved 2011-09-17.
- ↑ Vijay Pande (2011-04-11). "Re: Storage@Home". Retrieved 2011-09-17.