AVL tree
In computer science, an AVL tree is a self-balancing binary search tree. It was the first such data structure to be invented. In an AVL tree, the heights of the two child subtrees of any node differ by at most one; if at any time they differ by more than one, rebalancing is done to restore this property. Lookup, insertion, and deletion all take time in both the average and worst cases, where is the number of nodes in the tree prior to the operation. Insertions and deletions may require the tree to be rebalanced by one or more tree rotations.
The AVL tree is named after its two Soviet inventors, Georgy Adelson-Velsky and Evgenii Landis, who published it in their 1962 paper "An algorithm for the organization of information".
AVL trees are often compared with red–black trees because both support the same set of operations and take time for the basic operations. For lookup-intensive applications, AVL trees are faster than red–black trees because they are more strictly balanced. Similar to red–black trees, AVL trees are height-balanced. Both are, in general, neither weight-balanced nor -balanced for any ; that is, sibling nodes can have hugely differing numbers of descendants.
Definition
Balance factor
In a binary tree the [|balance factor] of a is defined to be the height differenceof its two child sub-trees. A binary tree is defined to be an AVL tree if the invariant
holds for every in the tree.
A with is called "left-heavy", one with is called "right-heavy", and one with is sometimes simply called "balanced".
;Remark
In what follows, because there is a one-to-one correspondence between nodes and the sub-trees rooted by them, the name of an object is sometimes used to refer to the node and sometimes used to refer to the sub-tree.
Properties
Balance factors can be kept up-to-date by knowing the previous balance factors and the change in height – it is not necessary to know the absolute height. For holding the AVL balance information in the traditional way, two bits per node are sufficient. However, later research showed if the AVL tree is implemented as a rank balanced tree with delta ranks allowed of 1 or 2—with meaning "when going upward there is an additional increment in height of one or two", this can be done with one bit.The of an AVL tree with lies in the interval:
with the golden ratio √, , and .
This is because an AVL tree of contains at least nodes where is the Fibonacci sequence with the seed values,.
Operations
Read-only operations of an AVL tree involve carrying out the same actions as would be carried out on an unbalanced binary search tree, but modifications have to observe and restore the height balance of the sub-trees.Searching
Searching for a specific key in an AVL tree can be done the same way as that of any balanced or unbalanced binary search tree. In order for search to work effectively it has to employ a comparison function which establishes a total order on the set of keys. The number of comparisons required for successful search is limited by the height and for unsuccessful search is very close to, so both are in.Traversal
Once a node has been found in an AVL tree, the next or previous node can be accessed in amortized constant time. Some instances of exploring these "nearby" nodes require traversing up to links. However, exploring all nodes of the tree in this manner would visit each link exactly twice: one downward visit to enter the subtree rooted by that node, another visit upward to leave that node's subtree after having explored it. And since there are links in any tree, the amortized cost is, or approximately 2.Insert
When inserting a node into an AVL tree, you initially follow the same process as inserting into a Binary Search Tree. If the tree is empty, then the node is inserted as the root of the tree. In case the tree has not been empty then we go down the root, and recursively go down the tree searching for the location to insert the new node. This traversal is guided by the comparison function. In this case, the node always replaces a NULL reference of an external node in the tree i.e., the node is either made a left-child or a right-child of the external node.After this insertion if a tree becomes unbalanced, only ancestors of the newly inserted node are unbalanced. This is because only those nodes have their sub-trees altered So it is necessary to check each of the node's ancestors for consistency with the invariants of AVL trees: this is called "retracing". This is achieved by considering the balance factor of each node.
Since with a single insertion the height of an AVL subtree cannot increase by more than one, the temporary balance factor of a node after an insertion will be in the range For each node checked, if the temporary balance factor remains in the range from –1 to +1 then only an update of the balance factor and no rotation is necessary. However, if the temporary balance factor becomes less than –1 or greater than +1, the subtree rooted at this node is AVL unbalanced, and a rotation is needed. With insertion as the code below shows, the adequate rotation immediately perfectly [|rebalances] the tree.
In figure 1, by inserting the new node Z as a child of node X the height of that subtree Z increases from 0 to 1.
;Invariant of the retracing loop for an insertion
The height of the subtree rooted by Z has increased by 1. It is already in AVL shape.
In order to update the balance factors of all nodes, first observe that all nodes requiring correction lie from child to parent along the path of the inserted leaf. If the above procedure is applied to nodes along this path, starting from the leaf, then every node in the tree will again have a balance factor of −1, 0, or 1.
The retracing can stop if the balance factor becomes 0 implying that the height of that subtree remains unchanged.
If the balance factor becomes ±1 then the height of the subtree increases by one and the retracing needs to continue.
If the balance factor temporarily becomes ±2, this has to be repaired by an appropriate rotation after which the subtree has the same height as before.
The time required is for lookup, plus a maximum of retracing levels on the way back to the root, so the operation can be completed in time.
Delete
The preliminary steps for deleting a node are described in section Binary search tree#Deletion.There, the effective deletion of the subject node or the replacement node decreases the height of the corresponding child tree either from 1 to 0 or from 2 to 1, if that node had a child.
Starting at this subtree, it is necessary to check each of the ancestors for consistency with the invariants of AVL trees. This is called "retracing".
Since with a single deletion the height of an AVL subtree cannot decrease by more than one, the temporary balance factor of a node will be in the range from −2 to +2.
If the balance factor remains in the range from −1 to +1 it can be adjusted in accord with the AVL rules. If it becomes ±2 then the subtree is unbalanced and needs to be rotated. ≠ 0 The various cases of rotations are described in section [|Rebalancing].
;Invariant of the retracing loop for a deletion
The height of the subtree rooted by N has decreased by 1. It is already in AVL shape.
The retracing can stop if the balance factor becomes ±1 meaning that the height of that subtree remains unchanged.
If the balance factor becomes 0 then the height of the subtree decreases by one and the retracing needs to continue.
If the balance factor temporarily becomes ±2, this has to be repaired by an appropriate rotation. It depends on the balance factor of the sibling Z whether the height of the subtree decreases by one –and the retracing needs to continue– or does not change and the whole tree is in AVL-shape.
The time required is for lookup, plus a maximum of retracing levels on the way back to the root, so the operation can be completed in time.
Set operations and bulk operations
In addition to the single-element insert, delete and lookup operations, several set operations have been defined on AVL trees: union, intersection and set difference. Then fast bulk operations on insertions or deletions can be implemented based on these set functions. These set operations rely on two helper operations, Split and Join. With the new operations, the implementation of AVL trees can be more efficient and highly-parallelizable.- Join: The function Join is on two AVL trees and and a key will return a tree containing all elements in, as well as. It requires to be greater than all keys in and smaller than all keys in. If the two trees differ by height at most one, Join simply create a new node with left subtree, root and right subtree. Otherwise, suppose that is higher than for more than one. Join follows the right spine of until a node which is balanced with. At this point a new node with left child, root and right child is created to replace c. The new node satisfies the AVL invariant, and its height is one greater than. The increase in height can increase the height of its ancestors, possibly invalidating the AVL invariant of those nodes. This can be fixed either with a double rotation if invalid at the parent or a single left rotation if invalid higher in the tree, in both cases restoring the height for any further ancestor nodes. Join will therefore require at most two rotations. The cost of this function is the difference of the heights between the two input trees.
=expose
if <=h
T'=Node
if <=h then return Node
else return rotateLeft
else
T'=joinRightAVL
T=Node
if <=h return T
else return rotateLeft
function joinLeftAVL
/* symmetric to joinRightAVL */
function join
if >h return joinRightAVL
if >h return joinLeftAVL
return Node
Here of a node the height of. expose= means to extract a tree node 's left child, the key of the node, and the right child. Node means to create a node of left child, key, and right child.
- Split: To split an AVL tree into two smaller trees, those smaller than key x, and those larger than key x, first draw a path from the root by inserting x into the AVL. After this insertion, all values less than x will be found on the left of the path, and all values greater than x will be found on the right. By applying Join, all the subtrees on the left side are merged bottom-up using keys on the path as intermediate nodes from bottom to top to form the left tree, and the right part is asymmetric. The cost of Split is, order of the height of the tree.
if return
=expose
if return
if
=split
return
if
=split
return )
The union of two AVLs and representing sets and, is an AVL that represents.
function union:
if t1 = nil:
return t2
if t2 = nil:
return t1
t<, t> ← split t2 on t1.root
return join
Here, Split is presumed to return two trees: one holding the keys less its input key, one holding the greater keys.
The algorithm for intersection or difference is similar, but requires the Join2 helper routine that is the same as Join but without the middle key. Based on the new functions for union, intersection or difference, either one key or multiple keys can be inserted to or deleted from the AVL tree. Since Split calls Join but does not deal with the balancing criteria of AVL trees directly, such an implementation is usually called the "join-based" implementation.
The complexity of each of union, intersection and difference is for AVLs of sizes and. More importantly, since the recursive calls to union, intersection or difference are independent of each other, they can be executed in parallel with a parallel depth. When, the join-based implementation has the same computational DAG as single-element insertion and deletion.
Rebalancing
If during a modifying operation a height difference of more than one arises between two child subtrees, the parent subtree has to be "rebalanced". The given repair tools are the so-called tree rotations, because they move the keys only "vertically", so that the in-order sequence of the keys is fully preserved.Let X be the node that has a balance factor of −2 or +2. Its left or right subtree was modified. Let Z be the higher child. Note that Z is in AVL shape by induction hypothesis.
In case of insertion this insertion has happened to one of Z's children in a way that Z's height has increased.
In case of deletion this deletion has happened to the sibling t1 of Z in a way so that t1's height being already lower has decreased.
There are four situations that might arise. We will describe them as Dir1 Dir2, where Dir1 comes from the set and Dir2 as a balance factor comes from the set.
Situation Dir1 Dir2 denotes:
i.e.
The balance violation of case Dir1 Dir2 is repaired by a simple rotation rotate_.
The case Dir1 != Dir2 is repaired by a double rotation rotate_ rotate_Dir1Dir2.
The cost of a rotation, both simple and double, is constant.
Simple rotation
Figure 4 shows a Right Right situation. In its upper half, node X has two child trees with a balance factor of +2. Moreover, the inner child t23 of Z is not higher than its sibling t4. This can happen by a height increase of subtree t4 or by a height decrease of subtree t1. In the latter case, also the pale situation where t23 has the same height as t4 may occur.The result of the left rotation is shown in the lower half of the figure. Three links and two balance factors are to be updated.
As the figure shows, before an insertion, the leaf layer was at level h+1, temporarily at level h+2 and after the rotation again at level h+1. In case of a deletion, the leaf layer was at level h+2, where it is again, when t23 and t4 were of same height. Otherwise the leaf layer reaches level h+1, so that the height of the rotated tree decreases.
;Code snippet of a simple left rotation
Double rotation
Figure 5 shows a Right Left situation. In its upper third, node X has two child trees with a balance factor of +2. But unlike figure 4, the inner child Y of Z is higher than its sibling t4. This can happen by the insertion of Y itself or a height increase of one of its subtrees t2 or t3 or by a height decrease of subtree t1. In the latter case, it may also occur that t2 and t3 are of same height.The result of the first, the right, rotation is shown in the middle third of the figure. The result of the final left rotation is shown in the lower third of the figure. Five links and three balance factors are to be updated.
As the figure shows, before an insertion, the leaf layer was at level h+1, temporarily at level h+2 and after the double rotation again at level h+1. In case of a deletion, the leaf layer was at level h+2 and after the double rotation it is at level h+1, so that the height of the rotated tree decreases.
;Code snippet of a right-left double rotation
Comparison to other structures
Both AVL trees and red–black trees are self-balancing binary search trees and they are related mathematically. Indeed, every AVL tree can be colored red–black, but there are RB trees which are not AVL balanced. For maintaining the AVL resp. RB tree's invariants, rotations play an important role. In the worst case, even without rotations, AVL or RB insertions or deletions require inspections and/or updates to AVL balance factors resp. RB colors. RB insertions and deletions and AVL insertions require from zero to three tail-recursive rotations and run in amortized time, thus equally constant on average. AVL deletions requiring rotations in the worst case are also on average. RB trees require storing one bit of information in each node, while AVL trees mostly use two bits for the balance factor, although, when stored at the children, one bit with meaning «lower than sibling» suffices. The bigger difference between the two data structures is their height limit.For a tree of size
- an AVL tree's height is at most
- :
- an RB tree's height is at most
- : .