Problem Set 9 appeared!

Oral exams will be between 2nd and 4th of August. Please enroll in BASIS to register for the exam and then schedule the date and time with Christiane Andrade in room II.62!

The Lecture Notes cover the lecture except for one possible new chapter. Part I is largely based on the following two books:

- [MR95] Rajeev Motwani and Prabhakar Raghavan. Randomized Algorithms. ISBN: 978-0521474658, Cambridge University Press, 1995.
- [MU05] Michael Mitzenmacher and Eli Upfal. Probability and Computing. ISBN: 978-0521835404, Cambridge University Press, 2005.

Date | Contents |
---|---|

April 20 | 1 Discrete Event Spaces and Probabilities 1.1 Discrete Probability Spaces 1.2 Independent Events & Conditional Probability |

April 24 | 1.2 (continued) Conditional Probability 1.3 Applications 1.3.1 Minimum Cut, Karger's Contract Algorithm |

April 27 | 1.3.1 (continued) Karger's Contract algorithm 1.3.1 FastCut |

May 1 | no lecture |

May 4 | 1.3.1 (continued) FastCut 1.3.1 Reservoir Sampling |

May 8 | 2.1 Random Variables and Expected Values 2.1.1 Integer Random Variables |

May 11 | 2.1.2 Conditional Expectation 2.2 Binomial and Geometric Distribution 2.3 Applications 2.3.1 Randomized QuickSort |

May 15 | 2.3.2 Randomized Approximation Algorithms 3 Concentration Bounds: Markov's Inequality |

May 18 | 3.1 Chebyshev's Inequality 3.3 Applications 3.3.1 A sublinear algorithm |

May 22 | 3.3.1 A sublinear algorithm |

May 25 | no lecture |

May 29 | 3.2 Chernoff/Rubin bounds 4 Random Walks 4.1 Useful math |

June 1 | 4.1 (continued) Useful math 4.2 Applications 4.2.1 An Algorithm for 2-SAT |

June 5 | no lecture |

June 8 | no lecture |

June 12 | 4.2.2 Algorithms for 3-SAT |

June 15 | no lecture |

June 19 | 6 Knapsack Problem and Multiobjective Optimization 6.1 Nemhauser-Ullmann Algorithm |

June 22 | 6.2 Number of Pareto-optimal Solutions 6.2.1 Upper Bound |

June 26 | 6.2.1 Upper Bound (continued) 6.3 Multiobjective Optimization |

June 29 | 6.2.2 Lower Bound |

- Problem Set 1 (hand in until May 2nd, to be discussed May 4th)
- Problem Set 2 (hand in until May 8th, to be discussed May 11th)
- Problem Set 3 (hand in until May 15th, to be discussed May 18th)
- Problem Set 4 (hand in until May 29th, to be discussed June 1st)
- Problem Set 5 (hand in until May 29th, to be discussed June 1st)
- Problem Set 6 (hand in until June 19th, to be discussed June 22nd)
- Problem Set 7 (hand in until June 19th, to be discussed June 22nd)
- Problem Set 8 (hand in until June 26th, to be discussed June 29th)
- Problem Set 9 (hand in until July 3rd, to be discussed July 6th)

The lecture has two parts. First, we consider the design and analysis of randomized algorithms. Many algorithmic problems can be solved more efficiently when allowing randomized decisions. Additionally, randomized algorithms are often easier to design and analyze than their (known) deterministic counterparts. For example, we will see an elegant algorithm for the minimum cut problem. Randomized algorithms can also be more robust on average, like randomized Quicksort.

The analysis of randomized algorithms builds on a set of powerful tools. We will get to know basic tools from probabily theory, very useful tail inequalities and techniques to analyze random walks and Markov chains. We apply these techniques to develop and analyze algorithms for important algorithmic problems like sorting and k-SAT.

Statements on randomized algorithms are either proven to hold on expectation or with high probability over the random choices. This deviates from the classical algorithm analysis but is still a worst-case analysis in its core. In the second part of the lecture, we learn about probabilistic analysis of algorithms. There are a number of important problems and algorithms for which worst-case analysis does not provide useful or empirically accurate results. One prominent example is the simplex method for linear programming whose worst-case running time is exponential while in fact it runs in near-linear time on almost all inputs of interest. Another example is the knapsack problem. While this problem is NP-hard, it is a very easy optimization problem in practice and even very large instances with millions of items can be solved efficiently. The reason for this discrepancy between worst-case analysis and empirical observations is that for many algorithms worst-case instances have an artificial structure and hardly ever occur in practical applications.

In smoothed analysis, one does not study the worst-case behavior of an algorithm but its (expected) behavior on random or randomly perturbed inputs. We will prove, for example, that there are algorithms for the knapsack problem whose expected running time is polynomial if the profits or weights are slightly perturbed at random. This shows that instances on which these algorithms require exponential running time are fragile with respect to random perturbations and even a small amount of randomness suffices to rule out such instances with high probability. Hence, it can be seen as an explanation for why these algorithms work well in practice. We will also apply smoothed analysis to the simplex method, clustering problems, the traveling salesman problem, etc.