I have found a way to handle repeated elements more efficiently in quicksort and would like to know if anyone has seen this done before.
This method greatly reduces the overhead involved in checking for repeated elements which improves performance both with and without repeated elements. Typically, repeated elements are handled in a few different ways which I will first enumerate.
First, there is the Dutch National Flag method which sort the array like [ < pivot | == pivot | unsorted | > pivot]
.
Second, there is the method of putting the equal elements to the far left during the sort and then moving them to the center the sort is [ == pivot | < pivot | unsorted | > pivot]
and then after the sort the ==
elements are moved to the center.
Third, the Bentley-McIlroy partitioning puts the ==
elements to both sides so the sort is [ == pivot | < pivot | unsorted | > pivot | == pivot]
and then the ==
elements are moved to the middle.
The last two methods are done in an attempt to reduce the overhead.
Now, let me explain how my method improves the quicksort by reducing the number of comparisons. I use two quicksort functions together rather than just one.
The first function I will call q1
and it sorts an array as [ < pivot | unsorted | >= pivot]
.
The second function I will call q2
and it sorts the array as [ <= pivot | unsorted | > pivot]
.
Let's now look at the usage of these in tandem in order to improve the handling of repeated elements.
First of all, we call q1
to sort the whole array. It picks a pivot which we will further reference to as pivot1
and then sorts around pivot1
. Thus, our array is sorted to this point as [ < pivot1 | >= pivot1 ]
.
Then, for the [ < pivot1]
partition, we send it to q1
again, and that part is fairly normal so let's sort the other partition first.
For the [ >= pivot1]
partition, we send it to q2
. q2
choses a pivot, which we will reference to as pivot2
from within this sub-array and sorts it into [ <= pivot2 | > pivot2]
.
If we look now at the entire array, our sorting looks like [ < pivot1 | >= pivot1 and <= pivot2 | > pivot2]
. This looks very much like a dual-pivot quicksort.
Now, let's return to the subarray inside of q2
([ <= pivot2 | > pivot2]
).
For the [ > pivot2]
partition, we just send it back to q1
which is not very interesting.
For the [ <= pivot2]
partition, we first check if pivot1 == pivot2
. If they are equal, then this partition is already sorted because they are all equal elements! If the pivots aren't equal, then we just send this partition to q2
again which picks a pivot (further pivot3
), sorts, and if pivot3 == pivot1
, then it does not have to sort the [ <= pivot 3]
and so on.
Hopefully, you get the point by now. The improvement with this technique is that equal elements are handled without having to check if each element is also equal to the pivots. In other words, it uses less comparisons.
There is one other possible improvement that I have not tried yet which is to check in qs2
if the size of the [ <= pivot2]
partition is rather large (or the [> pivot2]
partition is very small) in comparison to the size of its total subarray and then to do a more standard check for repeated elements in that case (one of the methods listed above).
Here are two very simplified qs1
and qs2
functions. They use the Sedgewick converging pointers method of sorting. They can obviously can be very optimized (they choose pivots extremely poorly for instance), but this is just to show the idea. My own implementation is longer, faster and much harder to read so let's start with this:
// qs sorts into [ < p | >= p ] void qs1(int a[], long left, long right){ // Pick a pivot and set up some indicies int pivot = a[right], temp; long i = left - 1, j = right; // do the sort for(;;){ while(a[++i] < pivot); while(a[--j] >= pivot) if(i == j) break; if(i >= j) break; temp = a[i]; a[i] = a[j]; a[j] = temp; } // Put the pivot in the correct spot temp = a[i]; a[i] = a[right]; a[right] = temp; // send the [ < p ] partition to qs1 if(left < i - 1) qs1(a, left, i - 1); // send the [ >= p] partition to qs2 if( right > i + 1) qs2(a, i + 1, right); } void qs2(int a[], long left, long right){ // Pick a pivot and set up some indicies int pivot = a[left], temp; long i = left, j = right + 1; // do the sort for(;;){ while(a[--j] > pivot); while(a[++i] <= pivot) if(i == j) break; if(i >= j) break; temp = a[i]; a[i] = a[j]; a[j] = temp; } // Put the pivot in the correct spot temp = a[j]; a[j] = a[left]; a[left] = temp; // Send the [ > p ] partition to qs1 if( right > j + 1) qs1(a, j + 1, right); // Here is where we check the pivots. // a[left-1] is the other pivot we need to compare with. // This handles the repeated elements. if(pivot != a[left-1]) // since the pivots don't match, we pass [ <= p ] on to qs2 if(left < j - 1) qs2(a, left, j - 1); }
I know that this is a rather simple idea, but it gives a pretty significant improvement in runtime when I add in the standard quicksort improvements (median-of-3 pivot choosing, and insertion sort for small array for a start). If you are going to test using this code, only do it on random data because of the poor pivot choosing (or improve the pivot choice). To use this sort you would call:
qs1(array,0,indexofendofarray);
If you want to know just how fast it is, here is a little bit of data for starters. This uses my optimized version, not the one given above. However, the one given above is still much closer in time to the dual-pivot quicksort than the std::sort
time.
On highly random data with 2,000,000 elements, I get these times (from sorting several consecutive datasets):
std::sort - 1.609 seconds dual-pivot quicksort - 1.25 seconds qs1/qs2 - 1.172 seconds
Where std::sort
is the C++ Standard Library sort, the dual-pivot quicksort is one that came out several months ago by Vladimir Yaroslavskiy, and qs1/qs2
is my quicksort implementation.
On much less random data. with 2,000,000 elements and generated with rand() % 1000
(which means that each element has roughly 2000 copies) the times are:
std::sort - 0.468 seconds dual-pivot quicksort - 0.438 seconds qs1/qs2 - 0.407 seconds
There are some instances where the dual-pivot quicksort wins out and I do realize that the dual-pivot quicksort could be optimized more, but the same could be safely stated for my quicksort.
I know this is a long question/explanation, but have any of you seen this improvement before? If so, then why isn't it being used?
Quicksort partitions an array and then calls itself recursively twice to sort the two resulting subarrays. This algorithm is quite efficient for large-sized data sets as its average and worst-case complexity are O(n2), respectively.
The best case for the algorithm now occurs when all elements are equal (or are chosen from a small set of k ≪ n elements).
Vladimir Yaroslavskiy | 11 Sep 12:35 Replacement of Quicksort in java.util.Arrays with new Dual-Pivot Quicksort
Visit http://permalink.gmane.org/gmane.comp.java.openjdk.core-libs.devel/2628
To answer your question, no I have not seen this approach before. I'm not going to profile your code and do the other hard work, but perhaps the following are next steps/considerations in formally presenting your algorithm. In the real world, sorting algorithms are implemented to have:
Good scalability / complexity and Low overhead
Scaling and overhead are obvious and are easy to measure. When profiling sorting, in addition to time measure number of comparisons and swaps. Performance on large files will also be dependent on disk seek time. For example, merge sort works well on large files with a magnetic disk. ( see also Quick Sort Vs Merge Sort )
Wide range of inputs with good performance
There's lots of data that needs sorting. And applications are known to produce data in patterns, so it is important to make the sort is resilient against poor performance under certain patterns. Your algorithm optimizes for repeated numbers. What if all numbers are repeated but only once (i.e. seq 1000>file; seq 1000>>file; shuf file)? What if numbers are already sorted? sorted backwards? what about a pattern of 1,2,3,1,2,3,1,2,3,1,2,3? 1,2,3,4,5,6,7,6,5,4,3,2,1? 7,6,5,4,3,2,1,2,3,4,5,6,7? Poor performance in one of these common scenarios is a deal breaker! Before comparing against a published general-purpose algorithm it is wise to have this analysis prepared.
Low-risk of pathological performance
Of all the permutations of inputs, there is one that performs worse than the others. How much worse does that perform than average? And how many permutations will provide similar poor performance?
Good luck on your next steps!
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With