270 likes | 517 Views
Mutual Information Scheduling for Ranking. Hamza Aftab Nevin Raj Paul Cuff Sanjeev Kulkarni Adam Finkelstein. Applications of Ranking. Pair-wise Comparisons. Query: A > B ? Ask a voter whether candidate I is better than candidate J Observe the outcome of a match. Scheduling.
E N D
Mutual Information Scheduling for Ranking HamzaAftab Nevin Raj Paul Cuff SanjeevKulkarni Adam Finkelstein
Pair-wise Comparisons • Query: • A > B ? • Ask a voter whether candidate I is better than candidate J • Observe the outcome of a match
Scheduling • Design queries dynamically, based on past observations.
Select Informative Matches • Assume matches are expensive but computation is cheap • Previous Work (Finkelstein) • Use Ranking Algorithm to make better use of information • Select matches by giving priority based on two criterion • Lack of information: Has a team been in a lot of matches already? • Comparability of the match: Are the two teams roughly equal in strength? • Our innovation • Select matches based on Shannon’s mutual information
Related Work • Sensor Management (tracking) • Information-Driven • [Manyika, Durrant-Whyte 1994] • [Zhao et. al. 2002] – Bayesian filtering • [Aoki et. al. 2011] – This session • Learning Network Topology • [Hayek, Spuckler 2010] • Noisy Sort
Ranking Algorithms – Linear Model • Each player has a skill level µ • The probability that Player I beats Player J is a function of the difference µi - µj • Transitive • Use Maximum Likelihood • Thurstone-Mosteller Model • Q function • Performance has Gaussian distribution about the mean µ • Bradley-Terry Model • Logistic function
Examples • Elo’s chess ranking system • Based on Bradley-Terry model • Sagarin’s sports rankings
Mutual Information • Mutual Information: • Conditional Mutual information
Entropy • Entropy: • Conditional Entropy High entropy Low entropy
Mutual Information Scheduling • Let R be the information we wish to learn • (i.e. ranking or skill levels) • Let Ok be the outcome of the kth match • At time k, scheduler chooses the pair (ik+1, jk+1):
Why use Mutual Information? • Additive Property • Fano’s Inequality • Related entropy to probability of error • For small error: • Continuous distributions: MSE bounds differential entropy
Greedy is Not Optimal • Consider Huffman codes---Greedy is not optimal
Evaluating Goodness-of-Fit 1 4 1 2 3 4 • Ranking: • Inversions • Skill Level Estimates: • Mean squared error (MSE) • Kullback-Leibler (KL) divergence (relative entropy) • Others • Betting risk • Sampling inconsistency 3 2
Numerical Techniques • Calculate mutual information • Importance sampling • Convex Optimization (tracking of ML estimate)
Summary of Main Idea • Get the most out of measurements for estimating a ranking • Schedule each match to maximize • (Greedy, to make the computation tractable) • Flexible • S is any parameter of interest, discrete or continuous • (skill levels; best candidate; etc.) • Simple design---competes well with other heuristics
Ranking Based on Pair-wise Comparisons • Bradley Terry Model: • Examples: • A hockey team scores Poisson- goals in a game • Two cities compete to have the tallest person • is the population
Computing Mutual Information • Importance Sampling: • Multidimensional integral • Probability distributions • Skill level estimates Skill level of player 2 Skill level of player 1 • Why is it good for estimating skill levels? • Faster than convex optimization • Efficient memory use
Visualizing the Algorithm Outcomes Scheduling A B C D ?