Lecture 2 Summary

This lecture was recorded with a very small audience - only four people were present, two of them being students in the class, and the other two were graduate students. There was also a technical problem at the start of the lecture that required the first part of the lecture to be repeated.

This lecture presented more advanced material on stable matchings. The goal of the first lecture was to show a basic algorithm along with its analysis and correctness proof. In contrast, the goal of the second lecture is to show an advanced topic, and give the flavor of the types of things that can be studied in an advanced course. There are two main results presented in this lecture. The first is that the proposal algorithm is "Optimal" for the men, but worst case for the women. The technical details of the proof are not presented - they are in the text, and can be read by a motivated student. The second result considers the average case analysis assuming that the preference lists are random. This might give some insight into how the algorithm would behave in practice. I think this is an interesting result for the analysis of algorithms, and draws on some probability theory. The result should be viewed strictly as "enrichment" - if students don't understand the result, it will not harm their performance later in the course. I hope that the result will inspire some of the students to be more interested in the study of algorithms.

In terms of spending time in class - more time should be spent on lecture one, than on lecture two.

Introduction

Slide 1, 00:00-01:48
The initial question is raised - is the algorithm better for M's or W's, this could be discussed.


Example

Slide 3, 01:48-04:35
Have the students submit their answers, identifying the different matchings. This is to show some matchings are good for the M's, and some are good for the W's.


M-Optimal Theorem

Slides 4-6, 04:35-12:34
The main theorem is discussed. The goal of the discussion is to provide a framework for someone to later read the proof in the text book, and not to present the actual proof. There may not be much discussion in this section.

A question is asked near the end of the discussion on slide six - how to construct an optimal algorithm for W. You should stop the video for this question.




Example

Another example is generated which shows what is good for one side can be bad for the other. This is done over a pair of slides, although in class, it turned out easiest to use just a single slide.



M-Rank, W-Rank

Following the summary slide, a pair of slides were used to define M-Rank. These can be done as student submissions, and then discussed with the class.




Average Case Analysis

Slides 12-16, 23:13-39:26
The question is now asked, what happens if the preferences lists are random. The approach taken in class was to have students try to predict what the M and W ranks would be. On the recorded version of the class, it is hard to hear the student answer - which was that they would both be close to n squared. This question is set up as a student submission - but students might be hestitant to guess. You could either do this as a student submission, or try to get students to guess. However, if you can't get any student answers - just run the lecture. The answer is that the M-rank is O(n log n), and the W-rank is O(n2/log n), or to put it a different way, the average m goes log n steps down his preference list, while the average w is n/log n steps down her preference list.