English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Conference Paper

Random Knapsack in Expected Polynomial Time

MPS-Authors
/persons/resource/persons44108

Beier,  Rene
Algorithms and Complexity, MPI for Informatics, Max Planck Society;

/persons/resource/persons45673

Vöcking,  Berthold
Algorithms and Complexity, MPI for Informatics, Max Planck Society;

External Resource
No external resources are shared
Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)
There are no public fulltexts stored in PuRe
Supplementary Material (public)
There is no public supplementary material available
Citation

Beier, R., & Vöcking, B. (2003). Random Knapsack in Expected Polynomial Time. In Proceedings of the 35th Annual ACM Symposium on Theory of Computing (STOC-03) (pp. 232-241). New York, USA: ACM.


Cite as: https://hdl.handle.net/11858/00-001M-0000-000F-2DDC-9
Abstract
In this paper, we present the first average-case analysis proving an expected polynomial running time for an exact algorithm for the 0/1 knapsack problem. In particular, we prove, for various input distributions, that the number of {\em dominating solutions\/} (i.e., Pareto-optimal knapsack fillings) to this problem is polynomially bounded in the number of available items. An algorithm by Nemhauser and Ullmann can enumerate these solutions very efficiently so that a polynomial upper bound on the number of dominating solutions implies an algorithm with expected polynomial running time. The random input model underlying our analysis is very general and not restricted to a particular input distribution. We assume adversarial weights and randomly drawn profits (or vice versa). Our analysis covers general probability distributions with finite mean, and, in its most general form, can even handle different probability distributions for the profits of different items. This feature enables us to study the effects of correlations between profits and weights. Our analysis confirms and explains practical studies showing that so-called {\em strongly correlated\/} instances are harder to solve than {\em weakly correlated\/} ones.