Approximation Algorithms for Correlated Knapsacks and Non-Martingale Bandits

Subscribers:
344,000
Published on ● Video Link: https://www.youtube.com/watch?v=g_huEDAA71E



Duration: 1:03:22
389 views
4


In the stochastic knapsack problem, we are given a knapsack with size B, and a set of jobs whose sizes and rewards are drawn from a known probability distribution. However, the only way to know the actual size and reward is to schedule the job---when it completes, we get to know these values. How should we schedule jobs to maximize the expected total reward? We know constant-factor approximations for this problem when we assume that rewards and sizes are independent random variables, and that we cannot prematurely cancel jobs after we schedule them. What can we say when either or both of these assumptions are dropped? Not only is the stochastic knapsack problem of interest in its own right, but techniques developed for it are applicable to other stochastic packing problems. Indeed, ideas for this problem have been useful for budgeted learning problems, where one is given several arms which evolve in a specified stochastic fashion with each pull, and the goal is to pull the arms a total of B times to maximize the reward obtained. Much recent work on this problem focus on the case when the evolution of the arms follows a martingale, i.e., when the expected reward from the future is the same as the reward at the current state. However, what can we say when the rewards do not form a martingale? We give constant-factor approximation algorithms for the stochastic knapsack problem with correlations and cancelations, and also for some budgeted learning problems where the martingale condition is not satisfied, using similar ideas. Indeed, we can show that previously proposed linear programming relaxations for these problems have large integrality gaps. We propose new time-indexed LP relaxations; using a decomposition and ``shifting'' approach, we convert these fractional solutions to distributions over strategies, and then use the LP values and the time ordering information from these strategies to devise a randomized scheduling algorithm. We hope our LP formulation and decomposition methods may provide a new way to address other correlated bandit problems with more general contexts. This is joint work with Anupam Gupta, Ravishankar Krishnaswamy and Marco Molinaro, all from CMU. The paper is available at http://arxiv.org/abs/1102.3749




Other Videos By Microsoft Research


2016-08-16Strategies for scaling up data mining algorithms
2016-08-16Applications of Grammatical Inference in Security and Program Analysis
2016-08-16User-Tailored Privacy for Internationally Operating Personalized Websites
2016-08-16Theories, Solvers and Static Analysis by Abstract Interpretation
2016-08-16Opinion Dynamics and Influence in Social Networks
2016-08-16Understanding Visual Scenes: Where are We?
2016-08-16Energy Functionals: Choices and Consequences For Medical Image Segmentation
2016-08-16Abelian Sandpiles and the Harmonic Model
2016-08-16Full-rank Gaussian Modeling of Convolutive Audio Mixtures Applied to Source Separation
2016-08-16MADDER and Self-Tuning Data Analytics on Hadoop with Starfish
2016-08-16Approximation Algorithms for Correlated Knapsacks and Non-Martingale Bandits
2016-08-16ChatArt: Interactive Pictographic Chat
2016-08-16Nonnegative k-sums, fractional covers, and probability of small deviations
2016-08-16Injective Tensor Norms: Hardness and Reductions
2016-08-16Monitoring Untrusted Modern Applications with Collective Record and Replay
2016-08-16Practical Boogie (on the example of VCC)
2016-08-16Coherent Depth in Stereo Vision
2016-08-16Multi-microphone Dereverberation and Intelligibility Estimation in Speech Processing
2016-08-16From Personalized Retinal Image Mapping to Large Scale Parallel Image Processing
2016-08-16Coding4Fun XAPfest!
2016-08-16Your Abstractions are Worth Powerless! Non-Volatile Storage and Computation on Embedded Devices



Tags:
microsoft research