Le séminaire a lieu le mardi à 11 h 45 (sauf modification exceptionnelle), au campus Côte de Nacre, bâtiment Sciences 3, salle S3 351, 3ème étage.
Résumé du séminaire du Mardi 26 Mars 2013
A general framework for the realistic analysis of sorting and searching algorithms. Application for some popular algorithms
par Thu Hien Ngyuen Thi (GREYC, Caen)
As there are some algorithms which deal with keys (viewed as a whole structure) and some others which deal with symbols, it is not easy to compare their efficiency. Furthermore, since in most real-life situations a key is complex and cannot be reduced to a single machine word, it is not realistic to consider the comparison between two keys as a unitary comparison. That is why Sedgewick proposed in 1998 to unify the analyses of basic sorting and searching algorithms by studying on the average the number of symbol comparisons between keys (seen as words). First results are due to Fill and Janson, Fill and Nakama who analyzed QuickSort and QuickSelect, but only on data composed of random uniform bits. Then, Clément, Fill, Flajolet and Vallée extend the previous analyses when data are words emitted by a general source.
In the present work, we draw a general framework for such realistic analyses: we provide an "automatic" method which takes as an input the probability of comparing two random words for a given algorithm and computes the exact formula for the average number of symbol comparisons. This method is applied to three popular algorithms: Insertion Sort, Bubble Sort and Minimum Selection. We also recover the results of QuickSort and QuickSelect. The results are the following: with respect to key comparisons, the average-case complexity of QuickSort is asymptotic to 2n log n, Insertion Sort to 2n, BubbleSort to n2/2, QuickMin to 2n, Selection Min to n-1. With respect to the number of symbol comparisons, the complexity becomes of respective order Θ(n log2n), Θ(n2), Θ(n2 log n) for sorting algorithms and remains Θ(n) for Selection algorithms. We further discuss the change in the complexity measure of the algorithms, from the number of key comparisons to the number of symbol comparisons and also the different constants arising in the analysis reflecting the mixing between algorithms and the way data are produced.
Joint work with Julien Clément and Brigitte Vallée.