BigTable
BigTable is a compressed, high performance, and proprietary data storage system built on Google File System, Chubby Lock Service, SSTable (log-structured storage like LevelDB) and a few other Google technologies. It is not distributed outside Google, although Google offers access to it as part of its Google App Engine.
History
BigTable development began in 2004[1] and is now used by a number of Google applications, such as web indexing,[2] MapReduce, which is often used for generating and modifying data stored in BigTable,[3] Google Maps,[4] Google Book Search, "My Search History", Google Earth, Blogger.com, Google Code hosting, Orkut,[4] YouTube,[5] and Gmail.[6] Google's reasons for developing its own database include scalability and better control of performance characteristics.[7]
Google's Spanner RDBMS is layered on an implementation of BigTable with a Paxos group for two-phase commits to each tablet. Google F1 was built using Spanner to replace an implementation based on MySQL.[8]
Design
BigTable maps two arbitrary string values (row key and column key) and timestamp (hence three dimensional mapping) into an associated arbitrary byte array. It is not a relational database and can be better defined as a sparse, distributed multi-dimensional sorted map.[9] BigTable is designed to scale into the petabyte range across "hundreds or thousands of machines, and to make it easy to add more machines [to] the system and automatically start taking advantage of those resources without any reconfiguration".[10]
Each table has multiple dimensions (one of which is a field for time, allowing for versioning and garbage collection). Tables are optimized for Google File System (GFS) by being split into multiple tablets – segments of the table are split along a row chosen such that the tablet will be ~200 megabytes in size. When sizes threaten to grow beyond a specified limit, the tablets are compressed using the algorithm BMDiff[11][12] and the Zippy compression algorithm[13] publicly known and open-sourced as Snappy,[14] which is a less space-optimal variation of LZ77 but more efficient in terms of computing time. The locations in the GFS of tablets are recorded as database entries in multiple special tablets, which are called "META1" tablets. META1 tablets are found by querying the single "META0" tablet, which typically resides on a server of its own since it is often queried by clients as to the location of the "META1" tablet which itself has the answer to the question of where the actual data is located. Like GFS's master server, the META0 server is not generally a bottleneck since the processor time and bandwidth necessary to discover and transmit META1 locations is minimal and clients aggressively cache locations to minimize queries.
Other similar software
- Apache Accumulo — built on top of Hadoop, ZooKeeper, and Thrift. Has cell-level access labels and a server-side programming mechanism. Written in Java.
- Apache Cassandra — brings together Dynamo's fully distributed design and BigTable's data model. Written in Java.
- Apache HBase — Provides BigTable-like support on the Hadoop Core.[15] Written in Java.
- Hypertable — Hypertable is designed to manage the storage and processing of information on a large cluster of commodity servers.[16] Written in C++.
- "KDI", Bluefish, GitHub — Kosmix attempt to make a BigTable clone. Written in C++.
- LevelDB — Google's embedded key/value store that uses similar design concepts as the BigTable Tablet.[17]
See also
- Amazon SimpleDB
- Big data
- Distributed data stores, an overview
- Dynamo (storage system)
- Column-oriented DBMS
- Hadoop
References
- ↑ Kumar, Aswini, Whitchcock, Andrew, ed., Google's BigTable, "First an overview. BigTable has been in development since early 2004 and has been in active use for about eight months (about February 2005).".
- ↑ Chang, Fay; Dean, Jeffrey; Ghemawat, Sanjay; Hsieh, Wilson C; Wallach, Deborah A; Burrows, Michael ‘Mike’; Chandra, Tushar; Fikes, Andrew; Gruber, Robert E (2006), "Bigtable: A Distributed Storage System for Structured Data" (PDF), Research, Google.
- ↑ Chang et al. 2006, p. 3: ‘Bigtable can be used with MapReduce, a framework for running large-scale parallel computations developed at Google. We have written a set of wrappers that allow a Bigtable to be used both as an input source and as an output target for MapReduce jobs’
- ↑ 4.0 4.1 Whitchcock, Andrew, Google's BigTable, "There are currently around 100 cells for services such as Print, Search History, Maps, and Orkut".
- ↑ Cordes, Kyle (2007-07-12), YouTube Scalability (talk), "Their new solution for thumbnails is to use Google’s BigTable, which provides high performance for a large number of rows, fault tolerance, caching, etc. This is a nice (and rare?) example of actual synergy in an acquisition.".
- ↑ "How Entities and Indexes are Stored", Google App Engine, Google Code.
- ↑ Chang et al. 2006, Conclusion: ‘We have described Bigtable, a distributed system for storing structured data at Google... Our users like the performance and high availability provided by the Bigtable implementation, and that they can scale the capacity of their clusters by simply adding more machines to the system as their resource demands change over time... Finally, we have found that there are significant advantages to building our own storage solution at Google. We have gotten a substantial amount of flexibility from designing our own data model for Bigtable.’
- ↑ Shute, Jeffrey ‘Jeff’; Oancea, Mircea; Ellner, Stephan; Handy, Benjamin ‘Ben’; Rollins, Eric; Samwel, Bart; Vingralek, Radek; Whipkey, Chad; Chen, Xin; Jegerlehner, Beat; Littlefield, Kyle; Tong, Phoenix (2012), "Summary; F1 — the Fault-Tolerant Distributed RDBMS Supporting Google's Ad Business" (presentation), Research, Sigmod: Google, p. 19, "We've moved a large and critical application suite from MySQL to F1".
- ↑ Chang et al. 2006.
- ↑ "Google File System and BigTable" (World Wide Web log), Radar, Database War Stories (7), O’Reilly, 5 2006 .
- ↑ "Google Bigtable, Compression, Zippy and BMDiff" (Web log), Feed, 2008-10-12.
- ↑ McIlroy, Bentley, Data compression using long common strings, "DCC ‘99", Explore (IEEE).
- ↑ "Google's Bigtable" (Weblog), Outer court, 2005-10-23.
- ↑ "Snappy" (project), Code, Google.
- ↑ "Background; HBase" (wiki), Hadoop Core, Apache.
- ↑ "About", Hyper table.
- ↑ "Leveldb file layout and compactions", Code, Google.
Bibliography
- Chang, Fay; Dean, Jeffrey; Ghemawat, Sanjay; Hsieh, Wilson C; Wallach, Deborah A; Burrows, Michael ‘Mike’; Chandra, Tushar; Fikes, Andrew; Gruber, Robert E (2006), "Bigtable: A Distributed Storage System for Structured Data" (PDF), Research, Google.
External links
- BigTable: A Distributed Structured Storage System, Washington. Video, Google.
- UWTV (video).
- Witchcock, Andrew, Google's BigTable (notes on the official presentation).
- Carr, David F (2006-07-06), "How Google Works", Baseline.
- "Is the Relational Database Doomed?", Read-write web.
|