5 \chapter{Approaching Optimality}
9 Recently, Chazelle \cite{chazelle:ackermann} and Pettie \cite{pettie:ackermann}
10 have presented algorithms for the MST with worst-case time complexity
11 $\O(m\timesalpha(m,n))$. We will devote this chapter to their results
12 and especially to another algorithm by Pettie and Ramachandran \cite{pettie:optimal},
13 which is provably optimal.
15 At the very heart of all these algorithms lies the \df{soft heap} discovered by
16 Chazelle \cite{chazelle:softheap}. It is a~meldable priority queue, roughly
17 similar to the Vuillemin's binomial heaps \cite{vuillemin:binheap} or Fredman's
18 and Tarjan's Fibonacci heaps \cite{ft:fibonacci}. The soft heaps run faster at
19 the expense of \df{corrupting} a~fraction of the inserted elements by raising
20 their values (the values are however never decreased). This allows for
21 an~trade-off between accuracy and speed controlled by a~parameter~$\varepsilon$.
22 The heap operations take $\O(\log(1/\varepsilon))$ amortized time and at every
23 moment at most~$\varepsilon n$ elements of the~$n$ elements inserted can be
26 \defnn{Soft heap operations}%
27 The soft heap contains distinct elements from a~totally ordered universe and it
28 supports the following operations:
30 \:$\<Create>(\varepsilon)$ --- create an~empty soft heap with the given accuracy parameter,
31 \:$\<Insert>(H,x)$ --- insert a~new element~$x$ to the heap~$H$,
32 \:$\<Meld>(P,Q)$ --- merge two heaps into one, more precisely insert all elements of a~heap~$Q$
33 to the heap~$P$, destroying~$Q$ in the process (both heaps must have the same~$\varepsilon$),
34 \:$\<DeleteMin>(H)$ --- delete the minimum element of the heap~$H$ and return its value
35 (optionally signalling that the value has been corrupted).
38 \examplen{Linear-time selection}
39 We can use soft heaps to select the median (or generally the $k$-th smallest element)
40 of a~sequence. We insert all $n$~elements to a~soft heap with error rate $\varepsilon=1/3$.
41 Then we delete the minimum about $n/3$ times and we remember the current (possibly corrupted)
42 value~$x$ of the last element deleted. This~$x$ is greater or equal than the current
43 values of the previously deleted elements and thus also than their correct values.
44 On the other hand, the current values of the $2n/3$ elements remaining in the heap
45 are greater than~$x$ and at most $n/3$ of them are corrupted. Therefore at least $n/3$
46 elements are greater than~$x$ and at least $n/3$ ones are smaller. Depending on the value
47 of~$k$, we can always throw away one of these parts and recurse on the rest (similarly to
48 the Hoare's Quickselect algorithm \cite{hoare:qselect}). The total time complexity will
49 be $\O(n+(2/3)\cdot n+(2/3)^2\cdot n+\ldots) = \O(n)$. We have obtained a~nice alternative to the standard
50 linear-time selection algorithm by Blum \cite{blum:selection}. The same trick can be used
51 to select a~good pivot in Quicksort \cite{hoare:qsort}, leading to time complexity $\O(n\log n)$
55 The soft heap is built from \df{soft queues} (we will omit the adjective soft
56 in the rest of this section). Each queue has a~shape of a~binary tree.\foot{%
57 Actually, Chazelle defines the queues as binomial trees, but he transforms them in ways that are
58 somewhat counter-intuitive, albeit well-defined. We prefer describing the queues as binary
59 trees with a~special distribution of values. In fact, the original C~code in the Chazelle's
60 paper uses this representation internally.}
61 Each vertex~$v$ of the tree remembers a~doubly-linked list $\<list>(v)$ of values. The
62 item list in every left son will be used only temporarily and it will be kept
63 empty between operations. Only right sons and the root have their lists
64 permanently occupied. The former vertices will be called \df{white}, the latter
65 \df{black.} (See the picture.)
67 The first value in the list is called the \df{controlling key} of the vertex,
68 denoted by $\<ckey>(v)$. If the list is empty, we keep the most recently used
69 value or we set $\<ckey>(v)=+\infty$. The \<ckey>'s obey the standard \df{heap order}
70 --- a~\<ckey> of a~parent is always less than the \<ckey>'s of its sons.
72 Each vertex is also assigned its \df{rank,} which is a~non-negative integer.
73 The rank of leaves is always zero, the rank of every internal vertex must be strictly
74 greater than the ranks of its sons. We define the rank of a~queue to be equal to the
75 rank of its root vertex and similarly for its \<ckey>.
77 A~queue is called \df{complete} if every two vertices joined by an~edge have rank
78 difference exactly one. In other words, it is a~complete binary tree and the
79 ranks correspond to vertex heights.
81 \figure{softheap1.eps}{\epsfxsize}{\multicap{A~complete and a~partial soft queue tree\\
82 (black vertices contain items, numbers indicate ranks)}}
85 A~complete queue of rank~$k$ contains exactly~$2^{k+1}-1$ vertices, $2^k$~of which are
86 black (by induction). Any other queue can be trivially embedded to a~complete queue of the same
87 rank, which we will call the \df{master tree} of the queue. This embedding preserves vertex
88 ranks, colors and the ancestor relation.
90 The queues have a~nice recursive structure. We can construct a~queue of rank~$k$ by \df{joining}
91 two queues of rank~$k-1$ under a~new root. The root will inherit the item list of one
92 of the original roots and also its \<ckey>. To preserve the heap order, we will choose
93 the one whose \<ckey> is smaller.
95 Sometimes, we will need to split a~queue to smaller queues. We will call this operation
96 \df{dismantling} the queue and it will happen only in cases when the item list in the root
97 is empty. It suffices to remove the leftmost (all white) path going from the root. From
98 a~queue of rank~$k$, we get queues of ranks $0,1,\ldots,k-1$, some of which may be missing
99 if the original queue was not complete.
101 We will now combine the soft queues to the soft heap.
103 \figure{softheap2.eps}{\epsfxsize}{Joining and dismantling of soft queues}
105 \defn A~\df{soft heap} consists of:
107 \:a~doubly linked list of soft queues of distinct ranks (in increasing order of ranks),
108 we will call the first queue the \df{head}, the last queue will be its \df{tail};
109 \:\df{suffix minima:} each queue contains a~pointer to the queue with minimum \<ckey>
110 of those following it in the list;
111 \:and a~global parameter~$r\in{\bb N}$, to be set depending on~$\varepsilon$.
114 \>We will define the \df{rank} of a~heap as the highest of the ranks of its queues
115 (that is, the rank of the heap's tail).
117 The heap always keeps the \df{Rank invariant:} When a~root of any tree has rank~$k$,
118 its leftmost path has length at least~$k/2$.
121 \em{Melding} of two soft heaps involves merging of their lists of queues. We disassemble
122 the heap with the smaller maximum rank and we insert its queues to the other heap.
123 If two queues of the same rank~$k$ appear in both lists, we \em{join} them to
124 a~single queue of rank~$k+1$ as already described and we propagate the new queue as
125 a~\df{carry} to the next iteration, similarly to addition of binary numbers.
126 Finally we have to update the suffix minima by walking the new list backwards
127 from the last inserted item.
129 \em{Creation} of a~new soft heap is trivial, \em{insertion} is handled by creating
130 a~single-element heap and melding it to the destination heap.
132 \algn{Creating a~new soft heap}
134 \algin A~parameter~$\varepsilon$.
135 \:Allocate memory for a~new heap structure~$H$.
136 \:Initialize the list of queues in~$H$ as an~empty list.
137 \:Set the parameter~$r$ to~$2\lceil\log(1/\varepsilon)\rceil+2$ (to be justified later).
138 \algout A~newly minted soft heap~$H$.
141 \algn{Melding of two soft heaps}
143 \algin Two soft heaps~$P$ and~$Q$.
144 \:If the heap~$P$ has smaller rank than the heap~$Q$, exchange their item lists.
146 \brk\cmt{Whenever we run into an~end of a~list in this procedure, we assume that it contains
147 an~empty queue of infinite rank.}
148 \:While $Q$ still has some queues:
150 \::If $\<rank>(p) < \<rank>(q)$, then $p\=$ the successor of~$p$,
151 \::else if $\<rank>(p) > \<rank>(q)$, remove~$q$ from~$Q$ and insert it before~$p$,
152 \::otherwise (the ranks are equal, we need to propagate the carry):
154 \:::$p\=$ the successor of~$p$.
155 \:::While $\<rank>(q)=\<rank>(\<carry>)$:
156 \::::Remove~$q$ from~$Q$.
157 \::::$\<carry>\=\<join>(q, \<carry>)$.
158 \::::$q\=\<head>(Q)$.
159 \:::Insert~\<carry> before~$q$.
160 \:Update the suffix minima: Walk with~$p$ backwards to the head of~$P$:
161 \::$p'\=$ successor of~$p$.
162 \::If $\<ckey>(p) < \<ckey>(p')$, set $\<suffix\_min>(p)\=p$.
163 \::Otherwise set $\<suffix\_min>(p)\=\<suffix\_min>(p')$.
164 \:Destroy the heap~$Q$.
165 \algout The merged heap~$P$.
168 \algn{Insertion of an~element to a~soft heap}
170 \algin A~heap~$H$ and a~new element~$x$.
171 \:Create a~new heap~$H'$ with the same parameters as~$H$. Let~$H'$ contain a~sole queue of rank~0,
172 whose only vertex has the element~$x$ in its item list.
174 \algout An~updated heap~$H$.
178 So far, the mechanics of the soft heaps was almost identical to the binomial heaps
179 and the reader could rightfully yawn. The things are going to get interesting
180 now as we approach the parts where corruption of items takes place.
182 If all item lists contain at most one item equal to the \<ckey> of the particular
183 vertex, no information is lost and the heap order guarantees that the minimum item
184 of every queue stays in its root. We can however allow longer lists and let the items
185 stored in a~single list travel together between the vertices of the tree, still
186 represented by a~common \<ckey>. This data-structural analogue of car pooling will
187 allow the items to travel at a~faster rate, but as only a~single item can be equal
188 to the \<ckey>, all other items will be inevitably corrupted.
190 We of course have to be careful about the size of the lists, as we must avoid corrupting
191 too many items. We will control the growth according to the vertex ranks. Vertices
192 with rank at most~$r$ will always contain just a~single item. Above this level,
193 the higher is the rank, the longer list will we allow.
196 \em{Deleting of minimum} will follow this principle. The minimum is easy to locate
197 --- we follow the \<suffix\_min> of the head of the heap to the queue with the minimum
198 \<ckey> and we look inside the item list of the root of this queue. We remove the \em{last}
199 item from the list (we do not want the \<ckey> to change) and return it as the minimum.
200 (It is not necessarily the real minimum of all items, but always the minimum of those
203 If the list becomes empty, we \em{refill} it with items from the lower levels of the
204 same tree. This can be best described recursively: We ask the left son to refill itself
205 (remember that the left son is always white, so there are no items to lose). If the new
206 \<ckey> of the left son is smaller than of the right son, we immediately move the left
207 son's list to its parent. Otherwise, we exchange the sons and move the list from the
208 new left son to the parent. This way, we obey the heap order and the same time we keep
209 the white left son free of items.
211 Ocassionally, we repeat this process once again and we concatenate the resulting lists
212 (we append the latter list to the former, using the smaller of the two \<ckey>'s). This
213 makes the lists grow longer and we want to do that roughly on every other level of the
214 tree. The exact condition will be that either the rank of the current vertex is odd,
215 or the difference in ranks between this vertex and its right son is at least two.
217 If refilling of the left son fails because there are no more items in that subtree
218 (we report this by setting its \<ckey> to $+\infty$), the current vertex is no longer
219 needed, as the items would just travel through it unmodified. We therefore want to
220 remove it. Instead of deleting it directly, we will rather make it point to its former
221 grandsons and we will remove the (now orhpaned) original son. This helps us to ensure
222 that both sons always have the same rank, which will be useful for the analysis.
224 When all refilling is done, we update the suffix minima by walking from the current
225 queue to the head of the heap exactly as we did in the \<Meld> procedure.
227 Our only remaining worry is that the Rank invariant can be broken after the
228 refilling. When the leftmost path of the tree becomes too short, we just
229 \em{dismantle} the tree as already described and we meld the new trees back to
230 the heap. This is easier to handle when the item list at the root vertex is
231 empty. We will therefore move this check before the refilling of the root list.
232 It will turn out that we have enough time to always walk the leftmost path
233 completely, so no explicit counters are needed.
235 Let us translate these ideas to real (pseudo)code:
237 \algn{Deleting the minimum item of a~soft heap}
239 \algin A~soft heap~$H$.
240 \:Use \<suffix\_min> of the head queue of~$H$ to locate the queue~$q$ with the smallest \<ckey>.
241 \:Remove the last element~$x$ of the item list in the root of~$q$.
242 \:If the item list is empty:
243 \::Count the vertices on the leftmost path of~$q$.
244 \::If there are less than $\<rank>(q)$ of them:
245 \:::Remove~$q$ from the list of queues.
246 \:::Recalculate the suffix minima as in the \<Meld> procedure.
247 \:::Dismantle~$q$ and create a~heap~$H'$ holding the resulting trees.
248 \:::Meld them back: $\<Meld>(H,H')$.
250 \:::Call \<Refill> on the root of~$q$.
251 \:::If $\<ckey>(q)=+\infty$ (no items left), remove the tree~$q$.
252 \:::Recalculate the suffix minima.
253 \algout The deleted minimum item~$x$ (possibly corrupted).
256 \algn{Refilling the item list of a~vertex}
258 \algin A~soft queue and its vertex~$v$ with an~empty item list.
259 \:Handle trivial cases: If~$v$ has no children or both have $\<ckey>=+\infty$,
260 set $\<ckey>(v)$ to~$+\infty$ and return.
261 \:Let \<left> and~\<right> denote the sons of~$v$.
262 \:Recurse: call $\<Refill>(\<left>)$.
263 \:If $\<ckey>(\<left>) > \<ckey>(\<right>)$, swap the sons.
264 \:Move the item list from \<left> to~$v$ (implying $\<ckey>(v)=\<ckey>(\<left>)$).
265 \:If $\<rank>(v) > r$ and either $\<rank>(v)$ is odd or $\<rank>(v) > \<rank>(\<right>)+1$, recurse once more:
266 \::Repeat steps 3--4.
267 \::Append the item list from \<left> to the item list at~$v$.
268 \:Clean up. If $\<ckey>(\<right>) = +\infty$:
269 \::If $\<ckey>(\<left>) = +\infty$, unlink and discard both sons.
270 \::Otherwise relink the sons of \<left> to~$v$ and discard \<left>.
271 \algout A~modified soft queue.
274 \paran{Analysis of accuracy}
275 The description of the operations is complete, let us analyse their behavior
276 and verify that we have delivered what we promised --- first the accuracy of
277 the structure, then the time complexity of operations. In the whole analysis,
278 we will denote the total number of elements inserted in the history of the
279 structure by~$n$. We will also assume that the threshold~$r$ is even.
281 We start by bounding the size of the item lists.
284 For every vertex~$v$ of a~soft queue, the size $\ell(v)$ of its item list
286 $$\ell(v) \le \max(1, 2^{\lceil \<rank>(v)/2 \rceil - r/2}).$$
289 Initially, all item lists contain at most one item, so the ineqality trivially
290 holds. Let us continue by induction. Melds can affect it only in the favorable
291 direction (they sometimes move an~item list to a~vertex of a~higher rank)
292 and so do deletes (they only remove items from lists). The only potentially
293 dangerous place is the \<Refill> procedure.
295 Refilling sometimes just moves items upwards, which is safe, and sometimes it
296 joins two lists into one, which generally is not. When $\<rank>(v) \le r$,
297 no joining takes place and~$\ell(v)$ is still~1. Otherwise we join when either
298 $\<rank>(v)$ is odd or $\<rank>(w) < \<rank>(v)-1$ for any son~$w$ of~$v$ (remember
299 that both sons have the same rank). In both cases, $\lceil\<rank>(w)/2\rceil \le
300 \lceil\<rank>(v)/2\rceil - 1$. By the induction hypothesis, the size of each
301 of the two joined lists is at most $2^{\max(1,\lceil\<rank>(v)/2\rceil - 1 - r/2)}$,
302 so the result has at most $2^{\lceil\<rank>(v)/2\rceil - r/2}$ items. (The maximum
303 has disappeared since $\<rank>(v)>r$ and therefore the desired bound is at least~2.)
306 We will now sum the sizes of the lists over all vertices containing corrupted items.
308 \lemma\id{shcorrlemma}%
309 At any given time, the heap contains at most~$n/2^{r-2}$ corrupted items.
312 We first prove an~auxiliary claim: The master trees of all queues contain at most~$n$
313 black vertices. This follows by induction: If no deletions have taken place,
314 there are exactly~$n$ of them, because insertion adds one black vertex and
315 melding preserves their number. A~deletion affects the master trees only when
316 dismantling takes place and then it only removes a~black vertex.
318 An~obvious upper bound on the number of corrupted items is the total size of item
319 lists in all vertices of rank greater than~$r$. We already know from the previous
320 lemma that the list sizes are limited by a~function of the ranks. A~complete tree
321 is obviously the worst case, so we will prove that this lemma holds for the master
322 tree of every queue in the heap. The actual trees can be much sparser, but the
323 above claim guarantees that the total size of the master trees is bounded by the
324 number of insertions properly.
326 So let us consider a~complete tree of~rank~$k$. It has exactly $2^{k-i}$ vertices
327 of rank~$i$ and each such vertex contains a~list of at most~$2^{\lceil i/2\rceil - r/2}$
328 items by the previous lemma. Summing over all ranks greater than~$r$, we get that
329 the total number of corrupted items in this tree is at most:
331 \sum_{i=r+1}^k 2^{k-i}\cdot 2^{\lceil i/2\rceil - r/2}
332 = 2^{k-r/2} \cdot \sum_{i=r+1}^k 2^{\lceil i/2\rceil - i}
333 \le 2^{k-r/2} \cdot \sum_{i=r+1}^k 2^{-i/2}
334 \le 2^{k-r} \cdot \sum_{i=0}^\infty 2^{-i/2}.
336 The sum of a~geometric series with quotient $2^{-1/2}$ is less than four, so the
337 last formula is less than $2^{k-r+2}$. Since the tree contains $n_k=2^k$ black vertices,
338 this makes less than $n_k/2^{k-2}$ corrupted items as we asserted.
341 \paran{Analysis of time complexity}
342 Now we will analyse the amortized time complexity of the individual operations.
343 We will show that if we charge $\O(r)$ time against every element inserted, it suffices
344 to cover the cost of all other operations. We take a~look at the melds first.
347 The amortized cost of a~meld is $\O(1)$, except for melds induced by dismantling
348 which take $\O(\<rank>(q))$, where $q$~is the queue to be dismantled.
351 The real cost of a~meld of heaps $P$ and~$Q$ is the smaller of their ranks plus
352 the time spent on carry propagation. The latter is easy to dispose of: since
353 every time there is a~carry, the total number of trees in all heaps decreases
354 by one, it suffices to charge $\O(1)$ against creation of a~tree. An~insert creates
355 one tree, dismantling creates at most $\<rank>(q)$ trees, all other operations
356 alter only the internal structure of trees.
358 As for the $\O(\min(\<rank>(P),\<rank>(Q)))$ part, let us assume for a~while that
359 no dismantling ever takes place and consider the \df{meld forest.} It is a~forest
360 whose leaves correspond to the $n$~single-element heaps constructed by \<Insert>
361 and each internal vertex represents a~heap arisen from melding its sons. The left
362 son will be the one with the greater (or equal) rank. We therefore want to bound
363 the sum of ranks of all right sons.
365 For every right son, we will distribute the change for its rank~$k$ on all leaves
366 in its subtree. There are at least $2^k$ such leaves. No leaf ever receives the same
367 rank twice, because the ranks of right sons on every path from the root of the
368 tree to a~leaf are strictly decreasing. (This holds because melding two heaps
369 of the same rank always produces a~heap of higher rank.) Hence at most~$n/2^k$
370 right sons have rank~$k$ and the total time charged against the leaves is bounded by:
372 \sum_{k=0}^{\rm max. rank}k\cdot {n\over 2^k} \le n\cdot\sum_{k=0}^\infty {k\over 2^k} = \O(n).
375 Let us return dismantling to the game. When a~queue is dismantled, melding the parts
376 back to the heap takes $\O(\<rank>(q))$ time. We can therefore let the dismantling pay for it
377 and omit such induced melds from the meld forest. As the rank of the heap is never increased
378 by induced melds, the above calculation is still a~proper upper bound on the cost
379 of the regular melds.
382 To estimate the time spent on deletions, we analyse the refills first.
385 Every call of the \<Refill> procedure spends time $\O(1)$ amortized.
388 When \<Refill> is called from the \<DeleteMin> operation, it recurses on a~subtree of the
389 queue. This subtree can be split to the ``lossless'' lower part (rank~$r$ and below)
390 and the upper part where list concatenation and thus also corruption takes place. Whenever
391 we visit the lower part during the recursion, we spend at worst $\O(r)$ time there.
392 We will prove that the total time spent in the upper parts during the whole life of the
393 data structure is $\O(n)$. Since each upper vertex can perform at most two calls to the lower
394 part, the total time spent in the lower parts is $\O(rn)$. All this can be pre-paid by the
397 Let us focus on the upper part. There are three possibilities of what can happen
398 when we visit a~vertex:
402 \:We delete it: Every vertex deleted has to have been created at some time in the past.
403 New vertices are created only during inserts and melds (when joining two trees) and
404 we have already shown that these operations have constant amortized complexity. So the
405 same must hold for deletions.
407 \:We recurse twice and concatenate the lists: The lists are disassembled only when
408 they reach the root of the tree, otherwise they are only concatenated. We can easily
409 model the situation by a~binary tree forest similar to the meld forest. There are~$n$
410 leaves and every internal vertex has outdegree two, so the total number of concatenations
411 is at most~$n$. Each of them can be performed in constant time as the list is doubly linked.
413 \:We recurse only once: This occurs only if the rank is even and the gap between the
414 rank of this vertex and its sons is small. It therefore cannot happen twice in a~row,
415 so it is clearly dominated by the other possibilities.
418 \>The total cost of all operations on the upper part is therefore $\O(n)$.
421 It remains to examine the rest of the \<DeleteMin> operation.
424 Every \<DeleteMin> takes $\O(1)$ time amortized.
427 Aside from refilling, which is $\O(1)$ by the previous lemma, the \<DeleteMin>
428 takes care of the Rank invariant. This happens by checking the length of the leftmost
429 path and dismantling the tree if the length is too far from the tree's rank~$k$.
430 The leftmost path is however always visited by the call to \<Refill>, so we can
431 account the check on the refilling. The same holds for disassembling. We then have
432 to pay $\O(k)$ for melding the trees back to the heap, but since there are at most
433 $k/2$ trees, a~subtree of rank at least $k/2$ must have been deleted. This tree
434 contained at least $2^{k/2}$ vertices which are now permanently gone from the
435 data structure, so we can charge the cost of the meld against these vertices.
437 We must not forget that \<DeleteMin> also has to recalculate the suffix minima.
438 In the worst case, it requires touching $k$~trees. Because of the Rank invariant,
439 this is linear in the size of the
440 leftmost path and therefore it can be also paid for by \<Refill>. (Incidentally,
441 this was the only place where we needed the invariant.)
444 Now we can put the bits together and laurel our effort with the following theorem:
446 \thmn{Performance of soft heaps, Chazelle \cite{chazelle:softheap}}
447 A~soft heap with error rate~$\varepsilon$ ($0<\varepsilon\le 1/2$) processes
448 a~sequence of operations starting with an~empty heap and containing $n$~\<Insert>s
449 in time $\O(n\log(1/\varepsilon))$. At every moment, the heap contains at most
450 $\varepsilon n$ corrupted items.
453 We set the parameter~$r$ to~$2+2\lceil\log (1/\varepsilon)\rceil$. The rest follows
454 from the analysis above. By Lemma \ref{shcorrlemma}, there are always at most $n/2^{r-2}
455 \le \varepsilon n$ corrupted items in the heap. By Lemma \ref{shmeld}--\ref{shdelmin},
456 the time spent on all operations in the sequence can be paid for by charging $\O(r)$ time
457 against each \<Insert>, which yields the time bound.
460 \FIXME{Remark on optimality.}
462 \FIXME{Example of use: pivots.}