This week's lab is a bit of a potpourri of topics:
Quicksort is the most common sorting algorithm in use today. In light of its technological impact, it's good to know how it works.
Fundamentally, Quicksort is akin to merge sort, in that it seeks to partition the set into two pieces and then recursively sort each piece. Mergesort simply whacks the set into two equal-sized pieces, which has the side effect of requiring a non-trivial merging process to re-join the sorted sub-pieces.
Rather than split the set (array) into two arbitrary and equal-sized pieces, Quicksort divides the set into subsets where one subset contains values that are all smaller than a given "pivot value" while the other subset are all values that are greater than the pivot. Some implementations put the pivot in the set of smaller numbers and some put it in the set of larger numbers and still others keep it completely separate.
Joining the sorted subsets is trivial because we already know that one set is completely smaller than the other set.
Which value to use as the pivot? Consider this: on average, for any randomly chosen element in the set (array), the value of that element is the mean of all the values. Thus, on average, half of the values in the set are smaller than any randomly chosen value and half are thus larger. What does this all mean? It means that it doesn't matter which element we choose to be our pivot because on average, half of the values will be larger than our pivot and half will be smaller, which is what we want. So, why not pick the first element? It's statistically no different than any other elenent and its real easy to find!
On average, quicksort will be faster than mergesort because the partitioning process of breaking the set into "smaller" and a "larger" subsets is less computationally expensive than mergesort's generalized merge operation. On the other hand, if the set is pathological, e.g. it's already sorted, Quicksort will degenerate into a O(N*N) behavior while mergesort will always be O(N*log(N)).
Here are some links to some animations of the Quicksort algorithm (all imperative/procedural code):
Here's a very basic implementation of Quicksort in an imperative coding style:
It is possible to optimize this code further to make it even faster.
From an OO sorting perspective, Quicksort is a hard-split, easy join example of Merritt's sorting theorem.
The split operations is just the partitioning of the array into two parts, one smaller than the pivot and one larger than the pivot. The split operation corresponds to the "partition" function that many Quicksort implementations identify. The join operation is simply a no-op because the two sorted subsets are disjoint and contiguous, so they are already in order.
Also download the lecture notes on this subject that includes an interesting decomposition of the Quicksort problem into an architecture that can be efficiently run in parallel on multi-core processors.
Variable Argument Lists
We discussed this before in lab and lecture, but it's worth revisiting to make sure everyone understands it.
A parameter defined as "P... params" is a vararg input parameter of type P. A vararg input parameter allows you to call the method with zero, one, two, or however many values as you want.
For instance, suppose I have a method define as such:
(Note: there are more efficient ways of writing this method, but I wanted to show you several features at once)
A vararg parameter comes in as an array, so one can figure out how many parameters were supplied (params.length) and one can access each supplied value using regular array syntax, e.g. params[i].
The following are all valid calls to the above method:
sum()--> returns zero.
sum(42)--> returns 0+42 = 42.
sum(10, 20)--> returns 0+10+20 = 30.
sum(100, 1000, 10000)--> returns 11100 etc
Varargs and Visitors
Vararg input parameters are very useful when you might want to call a method, such as the
accept method of a host (and hence a case on the visitor), with varying numbers of input parameters, or perhaps, none at all.
If you don't need to use the input parameter, I suggest using the "
Void" object type and calling your parameter something like "
nu" (for "not used"):
When you call the visitor, simply don't supply an input parameter.
If you need an input parameter, just set the
P (the generic type of the input parameter) to the type you need and call the
accept method with as many values as you need. Note that the one restriction of a vararg input parameter is that all the input values have to be of the same type.
The visitor system that HW11 uses allows you to give parameters to the visitor at the time that you use the visitor rather than having to imbed additional parameters inside the visitor as was done in many examples in class. This means that the same visitor instance can be used in multiple situations with different input parameters. To note however, when using anonymous inner classes, input parameter values are often not needed because the anonymous inner class already has access, via its closure, to most of the values it needs. Input parameters are particularly useful for things such as accumulators however, which cannot be curried in.
A Tale of Two Visitors
You were shown two different styles of writing visitors, with and without input parameters. Why two different techniques?
The main reason: because different people view the world in different manners and everyone's viewpoint has their strengths and weaknesses.
Here's a more technical breakdown:
Dr. Cartwright's No-input Parameter Visitor Style
Any necessary parameters for a visitor are handed to the constructor of the visitor and are saved in field(s) of the visitor to be used as the visitor executes.
- Parameters of multiple types are easily handled.
- The host's
acceptmethod is simplified and maintains exactly the same call structure for any visitor.
- When using anonymous inner class implementations of visitors, most parameters are curried in via the anonymous inner class's closure, so parameter storage fields are often not necessary.
- If the parameter value changes, e.g. for accululator algorithms, either one has to instantiate an entirely new instance of the visitor or take the risk of mutating the parameter storage field and hope that no one else is using the visitor.
- Anonymous inner class implementations treat parameters that are invariant over the lifetime of the visitor equivalently to those that are varying from call to call.
- The design does not cleanly separate the variant nature of the parameters from the invariant nature of the processing algorithm that the vistior represents.
Dr. Wong's Vararg Input Parameter Visitor Style
Parameters for a visitor are handed to the vistior at execution time (when accepted by the host) via a variable input argument list (vararg).
- The variant input parameters are cleanly separated from the invariant processing algorithm the vistior represents. This enables many more visitors to be written as singletons.
- Input parameters with varying values, e.g. accumulator algorithms are easily handled without instantiating a new visitor instance or worrying about multiple usages of the visitor.
- The same visitor instance can be used simultaneously, e.g. in parallel taskings and in branching recursive algorithms.
- Varargs allows varied numbers of parameter values to be handed to the same vistior, including none at all.
- Using anonymous inner class implementations allow the separation of parameters that are invariant over the lifetime of the visitor (curried-in values) vs. parameters that vary from call to call (input parameters).
- Parameters of different types are difficult to handle.
- Since the host's
acceptmethod allows any number of input parameters to be provided, it is impossible to tell if the correct number of input parameters has been passed.
- No parameter visitors are still required to declare an input parameter type (e.g. "
Who's "right"? That depends on where you stand and what's important to you. Perhaps neither. You choose what's right for what you're trying to do at that moment. There is no gospel, only defensible but non-absolute positions.
Binary Search in a Sorted Array
The recursive technique used to find a value in a sorted array is very useful in many situations.
Consider the following sorted array of values:
[1, 2, 4, 7, 10, 11, 12, 13, 20, 25, 27, 28, 30]
Supposed you're give a value,
x, what's the fastest way to either find that value in the array or determine if it is not present?
The trick is to use the same technique many people use to find a name in a phonebook, a technique called a "binary search" because it is continually making the searched space smaller by a factor of two:
- Open the book in the middle
mid = (max+min)/2(min, max and mid are the page numbers here)
- If the name you want is on that page, you're done.
- If the name is before the names where you opened the book, re-open the book halfway on the smaller side and repeat the process.
- If the name is after the names where you opened the book, re-open the book halfway on the larger side and repeat the process.
For the above array, we would look for the number 11 by doing the following:
- min = 0, max = 13, so mid = 6 ==> so look at the #12 (min, max and mid are indices into the array here)
- min= 0, max = 6, and 11 < 12, so look at mid = (6+0)/2 = 3, i.e. the #7.
- min= 3, max = 6, and 11 > 7, so look at the mid = (6+3)/2 = 4, i.e. the #10
- min = 4, max = 6 and 11>10 so look at the mid=(4+6)/2 = 5, i.e. the #11
- min = 5, max = 6, there's only one element left and since 11 = 11, we found it!
Here's some Java code to do the same:
Such as search is very fast, running in O(log N) time because you are reducing the size of the space by a constant multiplicative factor every time your recurse.
One way to look at what is going on here is to think of the sorted array as already being in the form of a binary search tree, where each mid-point is the top node of a sub-tree. All the above algorithm is doing is walking down from the top of the binary search tree, i.e the path 12 => 7 => 10 => 11.
It is a very useful notion to conceptually view an array as a tree. This enables one to leverage the power of recursive data structures with the speed and compactness of an array.
Insertion into a Doubly-linked List
Inserting a node into a doubly-linked list, ala
BiList, isn't difficult, but it does take care to make sure that one has the right values for the node references at any given moment. Here's a pictoral explanation of the process: