Linked lists and arrays store information in a linear arrangement. In this chapter, I introduce the binary tree recursive data structure, one of the most important concepts in the field of computer science. In Chapter 5, you learned about the concept of recursion, where a function calls itself. In this chapter, you will learn that the binary tree is a recursive data structure—that is, it refers to other binary tree structures. To introduce the concept of a recursive data structure, let’s revisit the linked list data structure you have already seen.
A linked list is an example of a recursive data structure because each node
has a next
reference to the first node of a sublist. Linked lists improve
upon fixed-length arrays by supporting dynamic growth and reduction of a
collection of N values. The sum_list()
recursive
function, shown in Listing 6-1, processes a linked list to
return its sum. Compare its implementation against a traditional iterative
solution.
class
Node
:
def
__init__
(
self
,
val
,
rest
=
None
)
:
self
.
value
=
val
self
.
next
=
rest
def
sum_iterative
(
n
)
:
total
=
0
while
n
:
total
+
=
n
.
value
n
=
n
.
next
return
total
def
sum_list
(
n
)
:
if
n
is
None
:
return
0
return
n
.
value
+
sum_list
(
n
.
next
)
Initialize total
to 0 to prepare for computation.
For each node, n
, in linked list, add its value to total
.
Advance to the next node in the linked list.
Base case: the sum of a nonexistent list is 0.
Recursive case: the sum of a linked list, n
, is the sum of its value added to the sum of the rest of the list.
The while
loop visits each node in the linked list and accumulates a
running total
of all values stored by nodes in the list. In contrast,
sum_list()
is a recursive function with a base case that terminates the
recursion and a recursive case that composes results of smaller
problem instances together. In the base case, the sum of a nonexistent list is
0. If a list, n
, has at least one node, then the recursive case
computes the sum of the rest of the list (that is, the list starting with
n.next
) and adds that result to n.value
to determine the total sum.
A linked list of N nodes is recursively decomposed into a first node, n
,
and the rest, a sublist containing N – 1 nodes. This is a recursive
decomposition—since, by definition, the rest is a sublist—but it only
subdivides the problem of size N (i.e., find the sum of the values
contained in N nodes) into a smaller problem of size N – 1 (i.e., find the
sum of the values contained in N – 1 nodes). To envision a recursive data
structure that has more productive subdivisions, consider representing
basic mathematical expressions using binary operations, such as
multiplication. A valid expression is either a value or combines two
sub-expressions using a binary operation:
3 — any numeric value is valid
(3 + 2) — add a left value 3 with a right value 2
(((1 + 5) ∗ 9) – (2 ∗ 6)) — subtract a right expression (2 ∗ 6) from a left expression ((1 + 5) ∗ 9)
Expressions can combine and grow to be as large as desired—the expression in Figure 6-1 has seven mathematical operations and eight numeric values. Linked lists cannot model this non-linear expression. If you’ve ever tried to visualize genealogies using family trees, then you can see why the diagram is called an expression tree.
The top multiply node has two child nodes, ultimately leading to four grandchild nodes (one of which is the value 4), six great-grandchild nodes, and two great-great grandchild nodes.
The expression in Figure 6-1 represents the multiplication of two expressions, which demonstrates its recursive structure. To compute the value of this expression, first recursively compute the left expression to produce the result 1. In similar recursive fashion, the right expression evaluates to 42, so the overall result of the original expression is 1 ∗ 42 = 42.
Figure 6-1 visualizes the recursive substructures of the original expression. At the top is a box representing a multiplication expression, with a left arrow and a right arrow to its left and right sub-expressions. Each circle represents a Value
node containing a numeric value, providing the base cases that stop recursion. The Expression
data structure in Listing 6-2 models expressions using left
and right
sub-expressions.
class
Value
:
def
__init__
(
self
,
e
)
:
self
.
value
=
e
def
__str__
(
self
)
:
return
str
(
self
.
value
)
def
eval
(
self
)
:
return
self
.
value
class
Expression
:
def
__init__
(
self
,
func
,
left
,
right
)
:
self
.
func
=
func
self
.
left
=
left
self
.
right
=
right
def
__str__
(
self
)
:
return
'
({} {} {})
'
.
format
(
self
.
left
,
self
.
func
.
__doc__
,
self
.
right
)
def
eval
(
self
)
:
return
self
.
func
(
self
.
left
.
eval
(
)
,
self
.
right
.
eval
(
)
)
def
add
(
left
,
right
)
:
"""+"""
return
left
+
right
A Value
stores a numeric value. It can return its value and string representation.
An Expression
stores a function, func
, and left
and right
sub-expressions.
Provides built-in __str()__
method to recursively produce strings with
parentheses around expressions.
Evaluate an Expression
by evaluating left
and right
children and
passing those values to func
.
Function to perform addition; mult()
for multiplication is similar. The docString
__doc__
for the function contains the operator symbol.
Evaluating an Expression
is a recursive process that will eventually
terminate at Value
objects. Using the same technique I introduced in
Chapter 5, Figure 6-2 visualizes the recursive
evaluation of the expression m
= ((1 + 5) ∗ 9):
>>>
a
=
Expression
(
add
,
Value
(
1
),
Value
(
5
))
>>>
m
=
Expression
(
mult
,
a
,
Value
(
9
))
>>>
(
m
,
'='
,
m
.
eval
())
((
1
+
5
)
*
9
)
=
54
To evaluate m
, there could be up to two recursive calls, one on the
left
and right
sub-expressions, respectively. In this case, the left
sub-expression, a
= (1 + 5) is evaluated recursively, while the right
sub-expression, 9, is not. The final computation of 54 is returned as
the final result. This example demonstrates the usefulness of the
Expression
recursive binary tree data structure. It also shows that
recursive implementations are brief and elegant.
It is critical that recursive data structures have no structural defects. For example:
>>>
n
=
Node
(
3
)
>>>
n
.
next
=
n
# This is dangerous!
>>>
(
sum_list
(
n
))
RecursionError
:
maximum
recursion
depth
exceeded
This linked list is defined by a node, n
, whose next
node in the linked list is itself! There is nothing wrong with the sum_list()
function—the linked list has a structural defect and so the base case for sum_list(n)
never occurs. A similar situation could occur in an Expression
. These defects are programming mistakes and you can avoid them by carefully testing your code.
Binary trees are the godfather of all recursive data structures. A binary search tree can store collections of values with efficient search, insert, and remove operations.
Storing values in a sorted array is necessary for Binary Array Search to
provide
O(log
N) performance. There are countless other reasons to
produce information in sorted order to make it easier for users to view
information. From a practical standpoint, very large fixed arrays are
challenging because they require contiguous memory that must be allocated
by the underlying operating system. In addition, changing the size of the
array is problematic:
To add a new value to an array, a new and larger array is created, and values from the old array are copied into the new array—while making room for the new value. Finally, memory for the old array is released.
To remove a value from the array, all values to the right of the removed value must shift one index location to the left. The code must remember that there are “unused” index locations at the end of the array.
Python programs avoid these difficulties because the built-in list
structure can already grow and shrink without programmer effort, but in the worst case, inserting a value into a Python list
remains O(N).
Table 6-1 measures the runtime performance both when prepending 1,000 values (one at a time) to the front of a list
of size N, and when appending 1,000 values (one at a time) to the end of a list
of size N.
N | Prepend | Append | Remove | Tree |
---|---|---|---|---|
1,024 |
|
|
|
|
2,048 |
|
|
|
|
4,096 |
|
|
|
|
8,192 |
|
|
|
|
16,384 |
|
|
|
|
32,768 |
|
|
|
|
65,536 |
|
|
|
|
131,072 |
|
|
|
|
262,144 |
|
|
|
|
524,288 |
|
|
|
|
1,048,576 |
|
|
|
|
As you can see in Table 6-1, the time to append values to the
end of a list is constant at 0.004; this can be considered the best
case for inserting values into a list
. The time to prepend 1,000
values to the front of a list
essentially doubles when the size of the
problem instance, N, doubles. This operation can be classified as
O(N). This table demonstrates the hidden cost of using a list
to maintain
a collection of sorted values. The column labeled “Remove” in Table 6-1
reveals that removing the first value from a list 1,000
times is O(N) too, since its runtime performance doubles as N doubles. Each
successive value in this column is almost exactly twice as large as the
value above it.
If the performance of inserting a value 1,000 times is O(N), you know that the performance of inserting a single value is also O(N), using the reasoning I introduced in Chapter 2. Inserting 10,000 values is O(N) using the same reasoning, since these behaviors are all just a multiplicative constant different from each other.
This empirical trial reveals O(N) performance when simply inserting or
removing a value; maintaining the array in sorted order will only slow the
program down. In contrast, the column labeled “Tree” in Table 6-1
reports the runtime performance when using a balanced binary search tree to maintain the collection of values while inserting 1,000 new values. As the problem size doubles, the runtime performance
appears to increase by a constant amount, which is characteristic of
O(log
N) performance. Even better, the binary search tree provides
efficient search, insert, and remove operations.
Figure 6-3 contains an example of a binary search tree placed side by side with a sorted array containing the same values. This is the same array from Figure 2-5, where I presented Binary Array Search. The top node in a binary tree is designated as the root of the tree, analogous to how the first node in a linked list is specially designated. In total there are seven nodes in this tree.
Each node in the binary search tree has the structure defined in
Listing 6-3. left
refers to a node that is the root of its own
subtree; the same is true of right
. A binary search tree adds two global
constraints for each node, n
, in the tree:
If a node, n
, has a left
subtree, all values in the subtree are
≤ n.value
.
If a node, n
, has a right
subtree, all values in the subtree are
≥ n.value
.
You can confirm these properties hold in
Figure 6-3. A leaf node is a node without a
left
or right
subtree; there are four leaf nodes in this tree,
containing the values 3, 15, 26, and 58. Many people have
commented that computer science trees are upside down, because the leaves
are at the bottom, while the root is at the top.
class
BinaryNode
:
def
__init__
(
self
,
val
)
:
self
.
value
=
val
self
.
left
=
None
self
.
right
=
None
Referring back to Figure 6-3, you can see how the left subtree of the root node is itself a tree whose root node, 14, has a left leaf node (representing 3
) and a right leaf node (representing 15). These are exactly the same values in the array depicted on the right that are smaller than or equal to 19, the middle index in the array. A binary tree grows top to bottom as values are inserted, one at a time, as
shown in Table 6-2.
![]() |
To insert 19, create a new subtree with root of 19. |
![]() |
To insert 14, 14 is smaller than or equal to 19, so insert 14 into the left subtree of 19, but there is no left subtree, so create a new subtree with root of 14. |
![]() |
To insert 15, 15 is smaller than or equal to 19, so insert 15 into the left subtree of 19 rooted at 14. Now 15 is larger than 14, so insert 15 into the right subtree of 14, but there is no right subtree, so create a new subtree with root of 15. |
![]() |
To insert 53, 53 is larger than 19, so insert 53 into the right subtree of 19, but there is no right subtree, so create a new subtree with root of 53. |
![]() |
To insert 58, 58 is larger than 19, so insert 58 into the right subtree of 19 rooted at 53. Now 58 is larger than 53, so insert 58 into the right subtree of 53, but there is no right subtree, so create a new subtree with root of 58. |
![]() |
To insert 3, 3 is smaller than or equal to 19, so insert 3 into the left subtree of 19 rooted at 14. Now 3 is smaller than or equal to 14, so insert 3 into the left subtree of 14, but there is no left subtree, so create a new subtree with root of 3. |
![]() |
To insert 26, 26 is larger than 19, so insert 26 into the right subtree of 19 rooted at 53. Now 26 is smaller than or equal to 53, so insert 26 into the left subtree of 53, but there is no left subtree, so create a new subtree with root of 26. |
It is convenient to have a BinaryTree
class to maintain the reference to
the root
node for a binary tree; over time in this chapter, additional
functions will be added to this class. Listing 6-4 contains the
code needed to insert values into a binary search tree.
BinaryTree
class to improve usability of binary search treeclass
BinaryTree
:
def
__init__
(
self
)
:
self
.
root
=
None
def
insert
(
self
,
val
)
:
self
.
root
=
self
.
_insert
(
self
.
root
,
val
)
def
_insert
(
self
,
node
,
val
)
:
if
node
is
None
:
return
BinaryNode
(
val
)
if
val
<
=
node
.
value
:
node
.
left
=
self
.
_insert
(
node
.
left
,
val
)
else
:
node
.
right
=
self
.
_insert
(
node
.
right
,
val
)
return
node
self.root
is the root node of the BinaryTree
(or None
if empty).
Use _insert()
helper function to insert val
into tree rooted at self.root
.
Base case: to add val
to an empty subtree, return a new BinaryNode
.
If val
is smaller than or equal to node
’s value, set node.left
to be the subtree that results when inserting val
into subtree node.left
.
If val
is larger than node
value, set node.right
to be the subtree that results when inserting val
into subtree node.right
.
This method must return node
to uphold its contract that it returns the root of the subtree into which val
was inserted.
The insert(val)
function in BinaryTree
invokes the recursive _insert(node, val)
helper function to set
self.root
to be the subtree that results when inserting val
into the subtree rooted at self.root
.1
The casual and elegant one-line implementation of insert()
is a feature
of programs using recursive data structures. The _insert()
function both
inserts val
and returns the root of the resulting subtree.
While all of the values being inserted are unique in this example, in
general, binary search trees can contain duplicate values, which is why the
_insert()
function checks if val
≤ node.value
.
In _insert(node, val)
, the base case occurs when node
is None
, which
occurs whenever val
is requested to be inserted into a nonexistent
subtree; it just returns a new subtree rooted by the newly created
BinaryNode
. For the recursive case, val
is inserted into either the
left subtree, node.left
, or the right subtree, node.right
. At the end
of the recursive case, it is critical that _insert()
returns node
to
fulfill its obligation of returning the root of the subtree that results
from adding val
to the subtree rooted at node
.
_insert(node, val)
maintains the binary search tree property such that
all values in the left subtree for node
are smaller than or equal to
node.value
, and values in the right subtree are larger than or equal to
node.value
.
A node, n
, in a binary tree can have a left
and a right
child. This
makes n
the parent node for left
and right
. The descendants of
n
are the nodes in its left
and right
subtrees. Each node, other than
the root, has at least one ancestor from which it descends.
Try inserting 29 into the binary search tree from Figure 6-3. 29 is larger than the root, so it must be inserted into the right subtree rooted by 53. 29 is smaller than 53, so insert into its left subtree rooted at 26. Finally, 29 is larger than 26, so it forms the new right subtree for 26, as shown in Figure 6-4.
The order in which values are inserted determines the structure of the final binary tree, as you can see in Figure 6-5. For the binary search tree on the left, you know that 5 was the first value inserted, since it is the root of the tree. In addition, every node must have been inserted after its ancestor.
The binary search tree on the right was formed by inserting the seven values in increasing order, which reveals the worst case for inserting values into a binary search tree; if you rotate the image counterclockwise about 45 degrees, it looks like a linked list, in which case it loses its efficiency. Toward the end of this chapter, I present a strategy to maintain more balanced tree structures in the face of insertions and removals.
The _insert()
method recursively finds the appropriate location to insert
a new leaf node containing the value to be added. This same recursive approach could
simply check whether a value is contained in a binary search tree; in
practice, however, the code in Listing 6-5 offers a
simpler, non-recursive solution using a while
loop.
BinaryTree
contains a valueclass
BinaryTree
:
def
__contains__
(
self
,
target
)
:
node
=
self
.
root
while
node
:
if
target
==
node
.
value
:
return
True
if
target
<
node
.
value
:
node
=
node
.
left
else
:
node
=
node
.
right
return
False
Start the search at the root
.
If target
value is same as node
’s value
, return True
for success.
If target
is smaller than node
’s value
, set node
to its left
subtree to continue search in that subtree.
If target
had been larger than node
’s value, continue search in right
subtree.
If the search runs out of nodes to inspect, the value does not exist in the tree, so return False
.
This __contains()__
function is added to the BinaryTree
class.2 Its structure is similar to searching for a value within a linked list; the
difference is that the next node to be searched could either be left
or
right
based on the relative value of target
.
Removing a value from a linked list was relatively straightforward—as discussed in Chapter 3—but removing a value from a binary search tree is more challenging. To start with, if the value to be removed is contained in the root node, how do you “stitch together” its orphaned left and right subtrees? Also, there should be a least-effort consistent strategy that works every time. Let’s try to find an intuitive solution to remove the value contained in the root node of a binary search tree. Figure 6-6 offers two possible binary search trees after removing the root value of 19.
You can confirm that both of these options remain binary search trees: the values in each left subtree remain smaller than or equal to its root, and the values in each right subtree remain larger than or equal to its root. The effort involved appears to be minimal for both options:
Option #1: Find and remove the maximum value in the left subtree, and use that value for the root.
Option #2: Find and remove the minimum value in the right subtree, and use that value for the root.
Each of these options is acceptable, and I choose to implement the second one. The resulting binary tree is a valid binary search tree because the new root value of 26 is the smallest value in the original right subtree—which means by definition it is smaller than or equal to all values in the revised subtree shaded in Figure 6-6. In addition, it is larger than or equal to all values in the original left subtree because it is larger than or equal to the original root value of 19, which was already larger than or equal to the values in the original left subtree.
Let’s start by solving the sub-problem that removes the minimum value in a given subtree. If you think about it, the minimum value in a subtree cannot have a left child—since otherwise a smaller value would exist. Given the binary search tree in Figure 6-7, the minimum value in the right subtree rooted at 53 is 26, and as you can see, it has no left child. Removing this value only requires “lifting up” its right subtree, rooted at 29, to become the new left subtree to 53. Setting the left child of 53 to be the tree rooted at 29 will always work, since 26 has no left subtree and no values will be lost.
Listing 6-6 contains the helper function in BinaryTree
,
_remove_min(node)
, that removes the minimum value in the subtree rooted
at node
; this function is never called when node
is None
. When
invoking _remove_min()
on rc
—the right child of the tree in
Figure 6-7—the recursive case is invoked, namely to remove
the minimum value in the left subtree rooted at 26. This leads to the
base case, since 26 has no left subtree, and in its place the function
“lifts up” and returns its right subtree rooted at 29 to become the new
left subtree to 53.
def
_remove_min
(
self
,
node
)
:
if
node
.
left
is
None
:
return
node
.
right
node
.
left
=
self
.
_remove_min
(
node
.
left
)
return
node
Base case: if node
has no left
subtree, then it is the smallest
value in the subtree rooted at node
; to remove it, just “lift up” and return its right
subtree (which could be None
).
Recursive case: remove the minimum value from left
subtree, and the returned subtree becomes new left
subtree for node
.
_remove_min()
completes the recursive case by returning the node whose left subtree may have been updated.
Once again, this code is brief and elegant. As with the other recursive
functions discussed earlier, _remove_min()
returns the root node of the
modified subtree. With this helper function, I can now complete the
implementation of remove()
that removes a value from a binary search
tree. To visualize what the code must do, Table 6-3 shows an
example of the changes to a binary tree when removing the value, 19,
contained in its root node.
![]() |
Since the root node
contains the value to be removed, set |
![]() |
Once the |
![]() |
After removing the
minimum value from the subtree rooted at |
![]() |
To complete the
update, |
The implementation of remove()
is shown in Listing 6-7.
BinaryTree
def
remove
(
self
,
val
)
:
self
.
root
=
self
.
_remove
(
self
.
root
,
val
)
def
_remove
(
self
,
node
,
val
)
:
if
node
is
None
:
return
None
if
val
<
node
.
value
:
node
.
left
=
self
.
_remove
(
node
.
left
,
val
)
elif
val
>
node
.
value
:
node
.
right
=
self
.
_remove
(
node
.
right
,
val
)
else
:
if
node
.
left
is
None
:
return
node
.
right
if
node
.
right
is
None
:
return
node
.
left
original
=
node
node
=
node
.
right
while
node
.
left
:
node
=
node
.
left
node
.
right
=
self
.
_remove_min
(
original
.
right
)
node
.
left
=
original
.
left
return
node
Use _remove()
helper function to remove val
from tree rooted at self.root
.
Base case: attempting to remove val
from nonexistent tree returns None
.
Recursive case #1: if value to be removed is smaller than node.value
, set node.left
to be the subtree that results from removing val
from node.left
.
Recursive case #2: if value to be removed is larger than node.value
, set node.right
to be the subtree that results from removing val
from node.right
.
Recursive case #3: it may be that node
is root of subtree and contains value to be removed, so there’s work to be done.
Handle easy cases first. If node
is a leaf, then None
is returned. If it has just one child, then return that child node.
Remember original reference to node
, since we don’t want to lose track of node
’s original left
and right
subtrees, both of which must exist.
Start with node = node.right
to find the smallest value in the subtree rooted at node.right
: as long as node
has a left subtree, then it does not contain the smallest value, so iteratively locate the node
with no left
subtree—this is the smallest value in the right
subtree of original
.
node
will become the new root to the left
and right
children of original
. Here I set node.right
to the subtree that results from removing the minimum value from original.right
. You might notice that this recursive method essentially repeats the process of the while
loop, but this code is much easier to understand than trying to do everything in just one pass.
Stitch the subtree rooted at node
back together.
The final capability supported by a binary search tree is returning the values in ascending order. In computer science this is known as a traversal.
To process each element in a linked list, start at the first node and use a
while
loop to follow next
references until all nodes are visited. This
linear approach can’t work with a binary search tree because there are
left
and right
references to follow. Given the recursive nature of a
binary tree data structure, there needs to be a recursive solution.
Listing 6-8 contains an elegant recursive solution that uses Python generators.
class
BinaryTree
:
def
__iter__
(
self
)
:
for
v
in
self
.
_inorder
(
self
.
root
)
:
yield
v
def
_inorder
(
self
,
node
)
:
if
node
is
None
:
return
for
v
in
self
.
_inorder
(
node
.
left
)
:
yield
v
yield
node
.
value
for
v
in
self
.
_inorder
(
node
.
right
)
:
yield
v
Yield all values that result from the in order traversal of binary search tree rooted at self.root
Base case: nothing to generate for a nonexistent subtree.
To generate all values in order, first generate all values in order from the subtree rooted at node.left
.
Now it is node
’s turn to yield its value.
Finally, generate all values in order from the subtree rooted at node.right
.
The __iter()__
function repeatedly yields the values provided by the
recursive helper function, _inorder()
, using a common idiom provided by
Python. For the base case of the recursion, when asked to yield the values
for a nonexistent binary search tree rooted at node
, _inorder()
returns
and does nothing. For the recursive case, this function relies on the
binary search tree property that all values in the subtree rooted at
node.left
are smaller than or equal to node.value
and that all values
in the subtree rooted at node.right
are larger than or equal to
node.value
. It recursively yields all values in node.left
before
yielding its own value, and subsequently yielding the values in
node.right
. The process is visualized in Figure 6-8 with
a binary search tree, T
, containing five values.
You can also choose to traverse a binary tree in two other traversal strategies. Use a preorder traversal to copy a binary tree. A postorder traversal visits all children before the parent, so use it to evaluate the value of an expression tree, such as shown in Figure 6-1.
I have now shown how to search, insert, and remove values from a binary search tree; in addition, you can retrieve these values in ascending order. It’s time to collectively analyze the performance of these fundamental operations.
The determining factor for search, insert, and remove operations is the
height of the tree, which is defined as the height of its root node. The
height of a node is the number of left
or right
references you need to
get from that node to its most distant descendant leaf node. This means the
height of a leaf node is 0.
The height of a nonexistent binary node cannot be 0, since leaf nodes
have a height of 0. The height of None
—that is, a nonexistent binary
node—is defined as –1 to make computations consistent.
In the worst case, the number of nodes visited during a search is based
on the height of the root of the binary search tree. Given N nodes in a
binary search tree, what is its height? It depends entirely on the order in
which the values had been inserted. A complete binary tree represents
the best case since it efficiently stores N = 2k
– 1 nodes in a tree
whose height is k
– 1. The binary tree in Figure 6-9,
for example, has N = 63 nodes, and the height of the root node is 5.
Searching for a target
value will involve no more than 6 comparisons
(since with 5 left
or right
references you will visit 6
nodes). Since 26 – 1 = 63, this means that the time to search for a
value is proportional to log
(N + 1). But in the worst case, all values were
inserted in ascending (or descending) order, and the binary search tree is
just a long linear chain, as shown in Figure 6-5. In
general, the runtime performance of search is O(h
), where h
is the
height of the binary search tree.
Inserting a value has the same time complexity as searching for a value—the only difference is that a new leaf node is inserted once the search
ends with a nonexistent left
or right
subtree; so inserting a value is
O(h
) as well.
Removing a value from a binary search tree requires three steps:
Locate the node containing the value to be removed
Locate the minimum value in the right subtree of the node containing the value to be removed
Remove that value from the right subtree
Each of these substeps in the worst case can be directly proportional to the height.3 At worst, then, the time to remove a value is proportional to 3 × h
, where h
is the height of the binary search tree. Based on the results from Chapter 2, since 3 is just a multiplicative constant, this means that the time to remove a value remains O(h
).
The structure of the binary search tree is based entirely on the order in
which the values are inserted and removed. Since the binary search tree
cannot control how it is used, it needs a mechanism to detect when its
structure is performing poorly. In Chapter 3, I explained how hashtables
resized themselves—rehashing all its entries—once a threshold size
was hit. It was acceptable for hashtables to do this, because this costly
O(N) operation would become ever more infrequently requested using the
geometric resizing strategy. As you may recall, doing so enabled the
average O(1
) runtime performance for get()
.
The geometric resizing strategy will not work here because there is no simple threshold computation based on N that can determine when to resize, and you can’t ensure that resize events become ever more infrequent: all it takes is a small sequence of awkward insertions to unbalance a tree, as visualized in Figure 6-10. Each node is color-coded by its height.
The complete binary tree on the left in Figure 6-10 has a perfectly balanced structure; each leaf node has a height of 0, and the root node has a height of 2. When 29 is inserted—as shown in the middle tree—a new leaf node is created in the proper location. Note that all ancestor nodes of 29 increase their height by 1, and they are shaded accordingly. After 27 is inserted—as shown in the right tree—the tree has lost its balance: its left subtree at 14 has a height of 1, but its right subtree at 53 has a height of 3. Other nodes—such as 26 and 53—are similarly out of balance. In the next section, I explain a strategy for detecting and rebalancing binary search trees.
The first known self-balancing binary tree data structure, the AVL tree, was invented in 1962.4 The premise is that as values are inserted into, or removed from, a binary search tree, weaknesses in the structure of the resulting tree are detected and repaired. An AVL tree guarantees that the height difference of any node—defined as the height of the node’s left subtree minus the height of the node’s right subtree—is –1, 0, or 1.
As shown in Listing 6-9, each BinaryNode
must store its height
in the binary search
tree. Whenever a node is inserted into the binary search tree, the height
of the affected nodes must be computed so an unbalanced tree node can be
detected immediately.
class
BinaryNode
:
def
__init__
(
self
,
val
)
:
self
.
value
=
val
self
.
left
=
None
self
.
right
=
None
self
.
height
=
0
def
height_difference
(
self
)
:
left_height
=
self
.
left
.
height
if
self
.
left
else
-
1
right_height
=
self
.
right
.
height
if
self
.
right
else
-
1
return
left_height
-
right_height
def
compute_height
(
self
)
:
left_height
=
self
.
left
.
height
if
self
.
left
else
-
1
right_height
=
self
.
right
.
height
if
self
.
right
else
-
1
self
.
height
=
1
+
max
(
left_height
,
right_height
)
Structure of a BinaryNode
is essentially the same as a binary search tree.
Record the height for each BinaryNode
.
Helper function that computes the height difference between left and right subtree.
Set left_height
to —1
for nonexistent left
subtree, or its proper height.
Return height difference, which must be left_height
subtracting right_height
.
Helper function that updates the height
for a node assuming that the height
of its respective left
and right
subtrees (if they exist) have accurate height
values.
Listing 6-10 shows that the node returned by _insert()
has its height
properly
computed.
_insert()
to compute height properly
def
_insert
(
self
,
node
,
val
)
:
if
node
is
None
:
return
BinaryNode
(
val
)
if
val
<
=
node
.
value
:
node
.
left
=
self
.
_insert
(
node
.
left
,
val
)
else
:
node
.
right
=
self
.
_insert
(
node
.
right
,
val
)
node
.
compute_height
(
)
return
node
During the invocation of insert(27)
, a new leaf node for 27 is added to
the binary search tree at the end of a sequence of recursive invocations,
depicted in Figure 6-11. The final invocation to
_insert()
involves the base case where a new leaf node containing 27 is
returned. This figure captures the brief moment when both the new
leaf node (for 27) and the original leaf node (for 29) have a height of
0. With just one additional statement to compute the node’s height at the
end of _insert()
, as the recursion unwinds, the height for each ancestor
node (highlighted in Figure 6-11) are recomputed—note
that these are the only nodes in the binary search tree whose heights need
to be adjusted. The compute_height()
function captures the logical
definition for the height of a node, namely, that it is one greater than
the larger of the heights of its children subtrees.
As the recursive invocations unwind, each ancestor node to 27 has its
height recomputed. Because each node has an accurate height
in the
binary search tree, _insert()
can detect whenever a node has become
unbalanced—that is, when the height of the left and right subtrees for
that node differ by more than 1.
In an AVL tree, the height difference for any node is -1, 0, or 1. The
height difference is computed as the height of the node’s left subtree
minus the height of the node’s right subtree. If a subtree doesn’t exist,
then use -1
for its height.
The node containing 26 leans to the right because its height difference is –1 – 1 = –2; the node containing 53 leans to the left because its height difference is 2 – 0 = 2; lastly the root node
leans to the right because its height difference is 1 – 3 = –2. Once these nodes are identified, there needs to be a strategy to adjust the tree
in some way to bring it back into balance. In the same way that the height
is computed as the recursion unwinds, the _insert()
function can
immediately detect when the insertion of a new node has unbalanced the
tree. It detects the imbalance as the recursion unwinds, which means the
first unbalanced node detected is 26.
The designers of the AVL tree invented the concept of a node rotation,
which is best described visually in Figure 6-12. The three
nodes—containing the values 10, 30, and 50—are shaded to
present their height. The root node, containing 50, has height h
. The
gray triangles are subtrees whose values conform to the binary search tree
property. The left subtree of the node containing 10, for example, is labeled
10L, and it contains values that are all smaller than or equal to
10. All you need to know is that the height of this subtree (and the
other three shaded subtrees, 10R, 30R, and 50R) is h
– 3.
The tree leans to the left: its left subtree has a height of h
– 1, while
its right subtree has a smaller height of h
– 3, meaning the height
difference is +2. An AVL tree rebalances itself by detecting this
imbalance and rotating nodes to reconfigure the tree, as shown on the right in
Figure 6-12. After the rotation, the resulting binary search
tree has a height of h
– 1, and the node containing 30 has become the new
root. This particular rotation is a rotate right, which you can visualize
as placing your hand on the original node containing 30 and rotating your
hand to the right, which “lifts up” the node containing 30 while
“dropping down” the node containing 50.
There are four possible unbalanced scenarios in a binary search tree containing just three values, as shown in Figure 6-13. Scenario Left-left represents a simplified version of the example in Figure 6-12, which only needs a rotate right to balance the tree; similarly, scenario Right-right represents its mirror image, needing only a rotate left to bring the tree back into balance. These scenarios are named for the relative positioning of each descendant node to the root. These rotate operations result in a balanced tree whose root node contains 30, with a left child containing 10 and a right child containing 50.
Scenario Left-right presents a more complicated unbalanced tree that can be rebalanced in two steps. First, rotate left the left subtree rooted at 10, which “drops down” the 10 node and “lifts up” the 30 node, resulting in a tree that matches scenario Left-left. Second, perform a rotate right to balance the tree. This two-step composite operation is called rotate left-right. Scenario Right-left represents the mirror image of scenario Left-right, resulting in a rotate right-left operation that rebalances the tree. The repository contains an optimized implementation for these composite operations.
Two new helper functions resolve situations when a node is unbalanced to the left (or to the right), as shown in Listing 6-11.
def
resolve_left_leaning
(
node
)
:
if
node
.
height_difference
(
)
==
2
:
if
node
.
left
.
height_difference
(
)
>
=
0
:
node
=
rotate_right
(
node
)
else
:
node
=
rotate_left_right
(
node
)
return
node
def
resolve_right_leaning
(
node
)
:
if
node
.
height_difference
(
)
==
-
2
:
if
node
.
right
.
height_difference
(
)
<
=
0
:
node
=
rotate_left
(
node
)
else
:
node
=
rotate_right_left
(
node
)
return
node
A node leans to the left when height difference is +2.
Detects the rotate_right
case by confirming that node
’s left
subtree is partially leaning left.
Otherwise, node
’s left
subtree is partially leaning right, meaning a rotate_left_right
is in order.
A node leans to the right when height difference is –2.
Detects the rotate_left
case by confirming that node
’s right
subtree is partially leaning right.
Otherwise, node
’s right
subtree is partially leaning left, meaning a rotate_right_left
is in order.
Be sure to remember to return node of (potentially rebalanced) subtree.
The strategy is to immediately resolve an unbalanced node once this
situation has been detected. The final implementation of _insert()
is
shown in Listing 6-12, which takes immediate advantage of
these resolution helper functions. Adding a value to a left subtree of a
node can never make that node right-leaning; similarly, adding a value to a
right subtree of a node can never make that node left-leaning.
def
_insert
(
self
,
node
,
val
)
:
if
node
is
None
:
return
BinaryNode
(
val
)
if
val
<
=
node
.
value
:
node
.
left
=
self
.
_insert
(
node
.
left
,
val
)
node
=
resolve_left_leaning
(
node
)
else
:
node
.
right
=
self
.
_insert
(
node
.
right
,
val
)
node
=
resolve_right_leaning
(
node
)
node
.
compute_height
(
)
return
node
The implementations for these rotation functions can be found in the code
repository. Table 6-4 describes the rotate_left_right
case, showing the code and the rebalanced tree. At the top, the new_root
and the other affected nodes and subtrees are identified. Below, the tree
is rebalanced, and, importantly, new heights are computed for child
and
node
.
Pay attention to how rotate_left_right()
returns the new root node for the balanced binary tree, since this unbalanced node could exist within a larger binary tree. There is no need to recompute the height of new_root
since the calling function—in _insert()
or _remove()
—will do that. You can confirm visually that the resulting binary tree still conforms to the binary search tree property: for example, all of the values in 30L are greater than or equal to 10 and less than or equal to 30, so this
subtree can be a right subtree of the node containing 10. A similar
argument explains why the subtree labeled 30R can be a left subtree to
the node containing 50.
![]() |
|
![]() |
|
The revised _insert()
method now rebalances the binary search tree as
needed. A similar change to _remove()
and _remove_min()
is
straightforward with these helper functions, as shown in Listing 6-13. The code is modified to
include four targeted interventions whenever the tree’s structure is
changed.
_remove()
to maintain AVL property
def
_remove_min
(
self
,
node
)
:
if
node
.
left
is
None
:
return
node
.
right
node
.
left
=
self
.
_remove_min
(
node
.
left
)
node
=
resolve_right_leaning
(
node
)
node
.
compute_height
(
)
return
node
def
_remove
(
self
,
node
,
val
)
:
if
node
is
None
:
return
None
if
val
<
node
.
value
:
node
.
left
=
self
.
_remove
(
node
.
left
,
val
)
node
=
resolve_right_leaning
(
node
)
elif
val
>
node
.
value
:
node
.
right
=
self
.
_remove
(
node
.
right
,
val
)
node
=
resolve_left_leaning
(
node
)
else
:
if
node
.
left
is
None
:
return
node
.
right
if
node
.
right
is
None
:
return
node
.
left
original
=
node
node
=
node
.
right
while
node
.
left
:
node
=
node
.
left
node
.
right
=
self
.
_remove_min
(
original
.
right
)
node
.
left
=
original
.
left
node
=
resolve_left_leaning
(
node
)
node
.
compute_height
(
)
return
node
Removing the minimum value from a subtree rooted at node.left
could make node
right-leaning; rotate to rebalance as needed.
Removing a value from the left
subtree of node
could make node
right-leaning; rotate to rebalance as needed.
Removing a value from the right
subtree of node
could make node
left-leaning; rotate to rebalance as needed.
After the minimum has been removed from the subtree returned to be node.right
, node
could be left-leaning; rotate to rebalance as needed.
The AVL implementation now properly rebalances whenever a new value is
inserted into the tree or a value is removed. Each rebalancing contains a
fixed number of operations and executes in O(1
) constant time. Because an
AVL tree is still a binary search tree, the search and traversal functions
do not need to change.
The compute_height()
helper function and the different node rotation
methods all perform in a constant amount of time—there are no further
recursive calls or loops in any of these functions. These tree maintenance
functions are invoked only when a node is detected to be unbalanced. It
turns out that when inserting a value into an AVL tree, there will never be
more than one node rotation required. When removing a value, it is
theoretically possible there will be multiple node rotations (see the
challenge exercise at the end of this chapter that investigates this
behavior). In the worst case, there will never be more log
(N)
rotations, which means that the runtime performance for search, insert, and
remove are all O(log
N).
Using the information from this chapter, you are now prepared to further investigate any recursive data structures. To close this chapter, I now reconsider the symbol table and priority queue data types to consider whether a binary tree can provide a more efficient implementation.
The same binary search tree structure can be used to implement the symbol table data type introduced in Chapter 3, as shown in Figure 6-14.
To do this, you have to modify the BinaryNode
structure to store both key
and value
, as shown in Listing 6-14.
BinaryNode
when using binary tree to store symbol tableclass
BinaryNode
:
def
__init__
(
self
,
k
,
v
)
:
self
.
key
=
k
self
.
value
=
v
self
.
left
=
None
self
.
right
=
None
self
.
height
=
0
BinaryTree
now needs put(k,v)
and get(k)
functions, instead of
insert()
and __contains()__
, to support the expected interface for a
symbol table. The changes are minimal, and only incidental changes are
required, so the code is not reproduced here; find it in the associated
code repository. When
navigating through the binary search tree, that is, whether to go left
or
right
, the decision is based on node.key
.
Using a binary search tree provides the added benefit that the keys can be retrieved from the symbol table in ascending order using the __iter()__
traversal function.
Chapter 3 described how open addressing and separate chaining can implement
the symbol table data type. It’s worth comparing the runtime performance of
a binary search tree against the results of open addressing and separate
chaining hashtables, as shown in Table 3-4. This
trial inserts N = 321,129 words from the English dictionary into a symbol
table. What is the smallest height for a binary tree that stores all of
these words? Recall that this height is computed by the formula
log
(N + 1) – 1, which is 17.293. After inserting all of these words in
ascending order from the English dictionary, the height of the resulting
AVL binary search tree is 18, which further demonstrates the efficiency
of AVL trees in storing information.
The hashtable implementations from Chapter 3 significantly outperform binary search trees, as shown in Table 6-5. If you ever need the keys for a symbol table in ascending order, then I recommend retrieving the keys from the symbol table and then sorting them separately.
Type | Open addressing | Separate chaining | AVL trees |
---|---|---|---|
Build time |
|
|
|
Access time |
|
|
|
Given that the heap data structure described in Chapter 4 is based on a binary tree structure, it is only natural to compare the runtime performance of a priority queue implemented using an AVL binary search tree where the priority
is used to navigate the structure of the binary search tree, as shown in Figure 6-15.
There are two benefits to using a binary search tree to implement a priority queue:
An array-based heap must create storage for a fixed number of values in advance. Using a binary search tree, the structure can grow to be as large as needed.
In the heap structure, there is no way to provide an iterator of the entries in the priority queue in priority order without dequeuing the values.5 With a binary search tree structure, this capability now exists using the traversal logic.
To get started, BinaryNode
now stores both value
and priority
, as shown in Listing 6-15. The priority
field will be used to navigate the binary tree for the search, insert, and remove operations.
BinaryNode
when using binary tree to store priority queueclass
BinaryNode
:
def
__init__
(
self
,
v
,
p
)
:
self
.
value
=
v
self
.
priority
=
p
self
.
left
=
None
self
.
right
=
None
self
.
height
=
0
In a max binary heap, the entry with highest priority is in storage[1]
, and it can be located in O(1
) constant time. This is not the case when using a binary search tree to store a priority queue. The BinaryNode
with highest priority is the rightmost node in the binary tree. To locate this value requires O(log
N) runtime performance, if the underlying binary tree is balanced using the techniques described in this chapter.
The biggest change, however, is that when a priority queue uses a binary search tree for storage, the only value to remove is the one with highest priority; this means the general purpose remove()
function is not needed. In its place, a _remove_max()
helper function is added to PQ
, as shown in Listing 6-16. The other helper functions are part of the standard priority queue interface. Note that the count, N
, of pairs is stored and managed by the PQ
class.
enqueue()
and dequeue()
functionsclass
PQ
:
def
__init__
(
self
)
:
self
.
tree
=
BinaryTree
(
)
self
.
N
=
0
def
__len__
(
self
)
:
return
self
.
N
def
is_empty
(
self
)
:
return
self
.
N
==
0
def
is_full
(
self
)
:
return
False
def
enqueue
(
self
,
v
,
p
)
:
self
.
tree
.
insert
(
v
,
p
)
self
.
N
+
=
1
def
_remove_max
(
self
,
node
)
:
if
node
.
right
is
None
:
return
(
node
.
value
,
node
.
left
)
(
value
,
node
.
right
)
=
self
.
_remove_max
(
node
.
right
)
node
=
resolve_left_leaning
(
node
)
node
.
compute_height
(
)
return
(
value
,
node
)
def
dequeue
(
self
)
:
(
value
,
self
.
tree
.
root
)
=
self
.
_remove_max
(
self
.
tree
.
root
)
self
.
N
-
=
1
return
value
Use a balanced binary search tree for storage.
To enqueue a (v
, p
) pair, insert that pair into the binary search tree and increment N
count.
The _remove_max()
helper method both removes the node with maximum priority from the subtree rooted at node
and returns its value and the node of the resulting subtree as a tuple.
Base case: with no right subtree, this node has maximum priority; return both the value in the node being deleted and the left subtree that will eventually take its place.
Recursive case: retrieve removed value
and root of updated subtree.
If node
is out of balance (it could now lean left), fix with rotations.
Compute node
height before returning it along with value that was removed.
The dequeue()
method removes node with maximum priority from the binary search tree and returns its value.
After decrementing count, N
, return the value
that had been associated with highest priority.
The runtime performance of this priority queue implementation still offers
O(log
N) behavior, although in absolute terms it is twice as slow as the
heap-based priority queue implementation from Chapter 4. The reason is that
maintaining the AVL binary search tree structure is more work than actually
needed for a priority queue. Still, if you need the ability to iterate over
its (value
, priority
) pairs in the order in which they would be
removed, this is an efficient alternative.
Binary trees are a dynamic, recursive data structure that organizes its values into left and right substructures, offering the potential to evenly subdivide a collection of N values into two structures, each containing (more or less) N/2 values. Binary trees form the basis for countless other recursive data structures that lead to efficient implementations, including:
Red-black trees, which offer a more efficient approach for balancing binary search trees, although the implementation is more complicated than with AVL trees.
B-trees and B+ trees, used for databases and file systems.
R-trees and R∗ trees, used for spatial information processing.
k-d trees, Quadtrees, and Octrees for space-partitioning structures.
To summarize:
Because trees are recursive data structures, it is natural to write recursive functions to manipulate their structure.
The most common technique for traversing a binary search tree is inorder
traversal, which returns all values in ascending order. The Expression
recursive structure includes a postorder traversal function that
produces the values in postfix order, which conforms to the postfix
notation, which is used by some handheld calculators.
Binary search trees must rebalance their structure to ensure they can
achieve O(log
N) runtime performance for its key operations. The AVL
technique is able to balance a tree by enforcing the AVL property, that
the height difference of any node is –1, 0, or 1. To make this
work efficiently, each binary node also stores its height in the tree.
A priority queue can be implemented using a balanced binary search tree
to store the (value
, priority
) pairs, using priority
when comparing
nodes. One benefit of this structure is you can use inorder traversal to
return the pairs stored by the priority queue in priority order, without
affecting the structure of the priority queue.
A symbol table can be implemented using a balanced binary search tree by enforcing the restriction that each key is unique in the binary search tree. However, the performance will not be as efficient as the hashtable implementations described in Chapter 3.
Write a recursive count(n, target)
function that returns the number of
times that target
exists within the linked list whose first node is n
.
Sketch the structure of a binary search tree with N nodes that requires
O(N) time to find the two largest values. Next, sketch the structure of a binary search tree with N nodes that requires O(1
) time to find the two largest values.
What if you wanted to find the kth smallest key in a binary search
tree? An inefficient approach would be to traverse the entire tree until k
nodes have been visited. Instead, add a function, select(k)
, to BinaryTree
that returns the
kth smallest key for k
from 0 to N – 1. For an efficient implementation,
you will need to augment the BinaryNode
class to store an additional
field, N
, that records the number of nodes in the subtree rooted at that
node (including that node itself). A leaf has an N
value of 1, for example.
Also add the companion method, rank(key)
, to BinaryTree
that returns an integer from 0 to
N – 1 that reflects the rank of key
in sorted order (or in other words, the
number of keys in the tree strictly less-than key
).
Given the values [3,14,15,19,26,53,58]
, there are 7! = 5,040 ways to
insert these seven values into an empty binary search tree. Compute the
number of different ways that the resulting tree is perfectly balanced,
with a height of 2, such as the binary search tree shown in Figure 6-3.
Can you generalize your result to an arbitrary collection of 2
k-1
values and present a recursive formula c
(k
) that computes this for any k
?
Write a contains(val)
method for BinaryTree
that invokes a recursive
contains(val)
method in BinaryNode
.
As described in this chapter, AVL trees are self-balancing. For a given N, can you compute the maximum height of an AVL tree containing N values, since they cannot all be so perfectly compact as a complete binary tree? Generate 10,000 random AVL trees of size N, and record the maximum observed height for each N.
Create a table that records whenever this maximum observed height increases. Predict the values of N such that an AVL tree of N nodes can have a tree height that is one greater than any AVL tree with N – 1 nodes.
Complete the SpeakingBinaryTree
in Listing 6-17 whose insert(val)
operation
produces English descriptions of the actions as they are
performed. Table 6-2 contains the desired output for each of
the corresponding operations. This recursive operation is different from
others in this chapter because it processes “top-down,” whereas most
recursive functions process “bottom-up” from base cases.
_insert()
method to return description of what happenedclass
BinaryNode
:
def
__init__
(
self
,
val
):
self
.
value
=
val
self
.
left
=
None
self
.
right
=
None
class
SpeakingBinaryTree
:
def
__init__
(
self
):
self
.
root
=
None
def
insert
(
self
,
val
):
(
self
.
root
,
explanation
)
=
self
.
_insert
(
self
.
root
,
val
,
'To insert `{}`, '
.
format
(
val
))
return
explanation
def
_insert
(
self
,
node
,
val
,
sofar
):
"""
Return (node,explanation) resulting from inserting val into subtree
rooted at node.
"""
Modify the _insert()
function to return a
tuple (node
, explanation
), where node
is the resulting node
and
explanation
contains the growing explanation for the actions.
Write a method check_avl_property(n)
that validates the subtree rooted at n
to ensure that (a) that each descendant
node’s computed height
is correct, and (b) that each descendant node satisfies the AVL
tree property.
Write a tree_structure(n)
function that produces a string with
parentheses in prefix order to capture the structure of the binary tree
rooted at n
. In prefix order, the value for the node is printed first,
before the left representation and right representation. This string should
use commas and parentheses to separate information, so it can be parsed
later. For the complete binary tree in
Figure 6-3, the resulting string should be
'(19,(14,(3,,),(15,,)),(53,(26,,),(58,,)))'
, while for the binary tree in
the left side of Figure 6-5, the resulting string should be
'(5,(4,(2,(1,,),(3,,)),),(6,,(7,,)))'
.
Write the companion recreate_tree(expr)
function that takes in an expr
tree structure string using parentheses and returns the root node of a
binary tree.
If you count rotate left-right and rotate right-left as single rotations (in addition to rotate left and rotate right), then you will never need more than a single rotation when inserting a value into an AVL binary search tree. However, when removing a value from an AVL tree, you might need multiple rotations.
What is the smallest AVL binary search tree that requires multiple node
rotations when a single value is removed? Such a tree would have to have
at least four nodes. To answer this question, you will have to instrument
the rotation methods to increase a count for the number of times a rotation
is used. Also, use the results of tree_structure()
in the previous
exercise to record the tree so you can recover its structure after you
have detected the multiple rotations. Write a function that (a) generates
10,000 random AVL trees containing between 4 and 40 nodes, and (b) for each of
these trees, select one of its random values to remove. You should be able to compute the size of the AVL trees that require up to three rotations for a remove request. As a hint, you should be able to generate an AVL tree with 4 nodes that requires 1 rotation for a given remove request, and an AVL tree with 12 nodes that requires 2 rotations for a given remove request. What is the smallest AVL tree that you can find which requires 3 rotations for a given remove request?
A complete binary tree with N = 2k
– 1 nodes is the most compact
representation for storing N nodes. This question asks what is the “least
compact” AVL tree you can construct. A Fibonacci tree is an AVL tree
such that in every node, the height of its left subtree is bigger (by just
1) than the height of its right subtree. Think of this as an AVL tree
that is one insert away from rebalancing. Write a recursive function,
fibonacci_avl(N)
, for N > 0 that returns a BinaryNode
representing the
root of a Fibonacci tree. It is simpler to do this without involving
any BinaryTree
objects. The root node returned contains the value
FN. For example,
fibonacci_avl(6)
would return the root node for the
binary tree depicted in Figure 6-16.
1 All recursive helper functions start with an underscore (_
) to declare that these functions are not intended to be part of the public interface for BinaryTree
.
2 Implementing this function means a program can use the Python in
operator to determine if a value is contained in a BinaryTree
object.
3 See a challenge exercise at the end of the chapter on this point.
4 Named after the inventors Adelson-Velsky and Landis.
5 See challenge exercise at end of this chapter showing how to do this.