Fractal tree index

From HandWiki
Short description: Tree data structure
Fractal tree index
Typetree
Invented2007
Invented byMichael A. Bender, Martin Farach-Colton, Bradley C. Kuszmaul
Time complexity in big O notation
Algorithm Average Worst case
Space O(N/B) O(N/B)
Search O(logB N) O(logB N)
Insert O(logB N/Bε) O(logB N/Bε)
Delete O(logB N/Bε) O(logB N/Bε)

In computer science, a fractal tree index is a tree data structure that keeps data sorted and allows searches and sequential access in the same time as a B-tree but with insertions and deletions that are asymptotically faster than a B-tree. Like a B-tree, a fractal tree index is a generalization of a binary search tree in that a node can have more than two children. Furthermore, unlike a B-tree, a fractal tree index has buffers at each node, which allow insertions, deletions and other changes to be stored in intermediate locations. The goal of the buffers is to schedule disk writes so that each write performs a large amount of useful work, thereby avoiding the worst-case performance of B-trees, in which each disk write may change a small amount of data on disk. Like a B-tree, fractal tree indexes are optimized for systems that read and write large blocks of data. The fractal tree index has been commercialized in databases by Tokutek. Originally, it was implemented as a cache-oblivious lookahead array,[1] but the current implementation is an extension of the Bε tree.[2] The Bε is related to the Buffered Repository Tree.[3] The Buffered Repository Tree has degree 2, whereas the Bε tree has degree Bε. The fractal tree index has also been used in a prototype filesystem.[4][5] An open source implementation of the fractal tree index is available,[6] which demonstrates the implementation details outlined below.

Overview

In fractal tree indexes, internal (non-leaf) nodes can have a variable number of child nodes within some pre-defined range. When data is inserted or removed from a node, its number of child nodes changes. In order to maintain the pre-defined range, internal nodes may be joined or split. Each internal node of a B-tree will contain a number of keys that is one less than its branching factor. The keys act as separation values which divide its subtrees. Keys in subtrees are stored in search tree order, that is, all keys in a subtree are between the two bracketing values. In this regard, they are just like B-trees.

Fractal tree indexes and B-trees both exploit the fact that when a node is fetched from storage, a block of memory, whose size is denoted by [math]\displaystyle{ B }[/math], is fetched. Thus, nodes are tuned to be of size approximately [math]\displaystyle{ B }[/math]. Since access to storage can dominate the running time of a data structure, the time-complexity of external memory algorithms is dominated by the number of read/writes a data structure induces. (See, e.g.,[7] for the following analyses.)

In a B-tree, this means that the number of keys in a node is targeted to be enough to fill the node, with some variability for node splits and merges. For the purposes of theoretical analysis, if [math]\displaystyle{ O(B) }[/math] keys fit in a node, then the tree has depth [math]\displaystyle{ O(\log_B N) }[/math], and this is the I/O complexity of both searches and insertions.

Fractal trees nodes use a smaller branching factor, say, of [math]\displaystyle{ \sqrt{B} }[/math]. The depth of the tree is then [math]\displaystyle{ O(\log_\sqrt{B} N) = O(\log_B N) }[/math], thereby matching the B-tree asymptotically. The remaining space in each node is used to buffer insertions, deletion and updates, which we refer to in aggregate as messages. When a buffer is full, it is flushed to the children in bulk. There are several choices for how the buffers are flushed, all leading to similar I/O complexity. Each message in a node buffer will be flushed to a particular child, as determined by its key. Suppose, for concreteness, that messages are flushed that are heading to the same child, and that among the [math]\displaystyle{ \sqrt{B}+1 }[/math] children, we pick the one with the most messages. Then there are at least [math]\displaystyle{ \frac B {\sqrt{B}+1} \approx \sqrt{B} }[/math] messages that can be flushed to the child. Each flush requires [math]\displaystyle{ O(1) }[/math] flushes, and therefore the per-message cost of a flush is [math]\displaystyle{ O\left(\frac 1 \sqrt{B}\right) }[/math].

Consider the cost of an insertion. Each message gets flushed [math]\displaystyle{ O(\log_B N) }[/math] times, and the cost of a flush is [math]\displaystyle{ O\left(\frac 1 \sqrt{B}\right) }[/math]. Therefore, the cost of an insertion is [math]\displaystyle{ O\left(\frac {\log_B N} \sqrt{B}\right) }[/math]. Finally, note that the branching factor can vary, but for any branching factor [math]\displaystyle{ B^{\varepsilon} }[/math], the cost of a flush is [math]\displaystyle{ O\left(\frac 1 {B^{1-\varepsilon}}\right) }[/math], thereby providing a smooth tradeoff between search cost, which depends on the depth of the search tree, and therefore the branching factor, versus the insertion time, which depends on the depth of the tree but more sensitively on the size of the buffer flushes.

Comparisons with other external-memory indexes

This section compares fractal tree indexes with other external memory indexing data structures. The theoretical literature on this topic is very large, so this discussion is limited to a comparison with popular data structures that are in use in databases and file systems.

B-trees

The search time of a B-tree is asymptotically the same as that of a fractal tree index. However, a fractal tree index has deeper trees than a B-tree, and if each node were to require an I/O, say if the cache is cold, then a fractal tree index would induce more IO. However, for many workloads most or all internal nodes of both B-trees and fractal tree indexes are already cached in RAM. In this case, the cost of a search is dominated by the cost of fetching the leaf, which is the same in both cases. Thus, for many workloads, fractal tree indexes can match B-trees in terms of search time.

Where they differ is on insertions, deletions and updates. An insertion in a fractal tree index takes [math]\displaystyle{ O\left(\frac {\log_B N} \sqrt{B}\right) }[/math] whereas B-trees require [math]\displaystyle{ O(\log_B N) }[/math]. Thus, fractal tree indexes are faster than B-trees by a factor of [math]\displaystyle{ O(\sqrt{B}) }[/math]. Since [math]\displaystyle{ B }[/math] can be quite large, this yields a potential two-order-of-magnitude improvement in worst-case insertion times, which is observed in practice. Both B-trees and fractal tree indexes can perform insertions faster in the best case. For example, if keys are inserted in sequential order, both data structures achieve a [math]\displaystyle{ O\left(\frac 1 B\right) }[/math] I/Os per insertion. Thus, because the best and worst cases of B-trees differ so widely, whereas fractal tree indexes are always near their best case, the actual speedup that fractal tree indexes achieve over B-trees depends on the details of the workload.

Log-structured merge-trees

Log-structured merge-trees (LSMs) refer to a class of data structures which consists of two or more index structures of exponentially growing capacities. When a tree at some level reaches its capacity, it is merged into the next bigger level. The IO-complexity of an LSM depends on parameters such as the growth factor between levels and the data structure chosen at each level, so in order to analyze the complexity of LSMs, we need to pick a specific version. For comparison purposes, we select the version of LSMs that match fractal tree indexes on insertion performance.

Suppose an LSM is implemented via [math]\displaystyle{ O(\log_B N) }[/math] B-trees, each of which has a capacity that is [math]\displaystyle{ \sqrt{B} }[/math] larger than its predecessor. The merge time depends on three facts: The sorted order of keys in an [math]\displaystyle{ N }[/math]-item B-tree can be produced in [math]\displaystyle{ O\left(\frac N B\right) }[/math] IOs; Two sorted lists of [math]\displaystyle{ N }[/math] and [math]\displaystyle{ M }[/math] items can be merged into a sorted list in [math]\displaystyle{ O\left(\frac{N+M} B\right) }[/math] IOs; and a B-tree of a sorted list of [math]\displaystyle{ N }[/math] items can be built in [math]\displaystyle{ O\left(\frac N B\right) }[/math] IOs. When a tree overflows, it is merged into a tree whose size is [math]\displaystyle{ O(\sqrt{B}) }[/math] larger, therefore a level that holds [math]\displaystyle{ k }[/math] items requires [math]\displaystyle{ O\left(\frac k \sqrt{B}\right) }[/math] IOs to merge. An item may be merged once per level, giving a total time of [math]\displaystyle{ O\left(\frac{\log_B N} \sqrt{B}\right) }[/math], which matches the fractal tree index.

The query time is simply the B-tree query time at each level. The query time into the [math]\displaystyle{ i }[/math]th level is [math]\displaystyle{ O(\log_B B^{\frac i 2}) = O(i) }[/math], since the [math]\displaystyle{ i }[/math]th level has capacity [math]\displaystyle{ B^{\frac i 2} }[/math]. The total time is therefore [math]\displaystyle{ O(\log^2_B N) }[/math]. This is larger than both the B-tree and fractal tree indexes by a logarithmic factor. In fact, although B-trees and fractal tree indexes are both on the optimal tradeoff curve between insertions and queries, LSMs are not. They are incomparable with B-trees and are dominated by fractal tree indexes.

A few notes about LSMs: there are ways to make the queries faster. For example, if only membership queries are required and no successor/predecessor/range queries are, then Bloom filters can be used to speed up queries. Also, the growth factor between levels can be set to some other value, giving a range of insertion/query tradeoffs. However, for every choice of insertion rate, the corresponding fractal tree index has faster queries.

Bε trees

The fractal tree index is a refinement of the Bε tree. Like a Bε tree, it consists of nodes with keys and buffers and realizes the optimal insertion/query tradeoff. The fractal tree index differs in including performance optimization and in extending the functionality. Examples of improved functionality include ACID semantics. B-tree implementations of ACID semantics typically involve locking rows that are involved in an active transactions. Such a scheme works well in a B-tree because both insertions and queries involve fetching the same leaf into memory. Thus, locking an inserted row does not incur an IO penalty. However, in fractal tree indexes, insertions are messages, and a row may reside in more than one node at the same time. Fractal tree indexes therefore require a separate locking structure that is IO-efficient or resides in memory in order to implement the locking involved in implementing ACID semantics.

Fractal tree indexes also have several performance optimizations. First, buffers are themselves indexed in order to speed up searches. Second, leaves are much larger than in B-trees, which allows for greater compression. In fact, the leaves are chosen to be large enough that their access time is dominated by the bandwidth time, and therefore amortizes away the seek and rotational latency. Large leaves are an advantage with large range queries but slow down point queries, which require accessing a small portion of the leaf. The solution implemented in fractal tree indexes is to have large leaves that can be fetched as a whole for fast range queries but are broken into smaller pieces call basement nodes which can be fetched individually. Accessing a basement node is faster than accessing a leaf, because of the reduced bandwidth time. Thus the substructure of leaves in fractal tree indexes, as compared to Bε trees allows both range and point queries to be fast.

Messaging and fractal tree indexes

Insertions, deletions and updates are inserted as message into buffers that make their way towards the leaves. The messaging infrastructure can be exploited to implement a variety of other operations, some of which are discussed below.

Upserts

An upsert is a statement that inserts a row if it does not exist and updates it if it does. In a B-tree, an upsert is implemented by first searching for the row and then implementing an insertion or an update, depending on the result of the search. This requires fetching the row into memory if it is not already cached. A fractal tree index can implement an upsert by inserting a special upsert message. Such a message can, in theory, implement arbitrary pieces of code during the update. In practice, four update operations are supported:

  1. [math]\displaystyle{ x := \text{constant} }[/math]
  2. [math]\displaystyle{ x := x + \text{constant} }[/math] (a generalized increment)
  3. [math]\displaystyle{ x := x - \text{constant} }[/math] (a generalized decrement)
  4. [math]\displaystyle{ x := \begin{cases} 0 & \text{if } x=0 \\ x-1 & \text{if } x \ne 0\end{cases} }[/math] (a decrement with a floor at 0)

These correspond to the update operations used in LinkBench,[8] a benchmark proposed by Facebook. By avoiding the initial search, upsert messages can improve the speed of upserts by orders of magnitude.

Schema changes

So far, all message types have modified single rows. However, broadcast messages, which are copied to all outgoing buffers, can modify all rows in a database. For example, broadcast messages can be used to change the format of all rows in a database. Although the total work required to change all rows is unchanged over the brute-force method of traversing the table, the latency is improved, since, once the message is injected into the root buffer, all subsequent queries will be able to apply the schema modification to any rows they encounter. The schema change is immediate and the work is deferred to such a time when buffers overflow and leaves would have gotten updated anyway.

Implementations

The fractal tree index has been implemented and commercialized by Tokutek. It is available as TokuDB as a storage engine for MySQL and MariaDB, and as TokuMX, a more complete integration with MongoDB. Fractal tree indexes have also been used in prototype filesystems, TokuFS[4] and BetrFS.[5]

References

  1. Bender, M. A.; Farach-Colton, M.; Fineman, J.; Fogel, Y.; Kuszmaul, B.; Nelson, J. (June 2007). "Cache-Oblivious streaming B-trees". Proceedings of the 19th Annual ACM Symposium on Parallelism in Algorithms and Architectures (CA: ACM Press): 81–92. http://supertech.csail.mit.edu/cacheObliviousBTree.html. 
  2. Brodal, G.; Fagerberg, R. (Jan 2003). "Lower Bounds for External Memory Dictionaries". Proceedings of the Fourteenth Annual ACM-SIAM Symposium on Discrete Algorithms (New York City: ACM Press): 546–554. http://www.madalgo.au.dk/~gerth/pub/soda03.html. 
  3. Buchsbaum, A.; Goldwasswer, M.; Venkatasubramanian, S.; Westbrook, J. (Jan 2000). "On External Memory Graph Traversal". Proceedings of the Eleventh Annual ACM-SIAM Symposium on Discrete Algorithms: 859–860. 
  4. 4.0 4.1 Esmet, J.; Bender, M.; Farach-Colton, M.; Kuszmaul, B. (June 2012). "The TokuFS Streaming File System". MA: USENIX Association. pp. 14. http://www.cs.sunysb.edu/~bender/newpub/2012-EsmetBeFa-HotStorage.pdf. 
  5. 5.0 5.1 Jannen, William; Yuan, Jun; Zhan, Yang; Akshintala, Amogh; Esmet, John; Jiao, Yizheng; Mittal, Ankur; Pandey, Prashant et al. (February 2015). "BetrFS: A Right-Optimized Write-Optimized File System". Santa Clara, California. https://www.usenix.org/system/files/conference/fast15/fast15-paper-jannen_william.pdf. 
  6. Github Repository
  7. Cormen, T.; Leiserson, C.E.; Rivest, R.; Stein, C. (2001). Introduction to Algorithms (2nd ed.). MIT Press and McGraw-Hill. ISBN 0-262-03293-7.