Fast Semidifferential-based Submodular Function Optimization: Extended Version
نویسندگان
چکیده
We present a practical and powerful new framework for both unconstrained and constrained submodular function optimization based on discrete semidifferentials (suband super-differentials). The resulting algorithms, which repeatedly compute and then efficiently optimize submodular semigradients, offer new and generalize many old methods for submodular optimization. Our approach, moreover, takes steps towards providing a unifying paradigm applicable to both submodular minimization and maximization, problems that historically have been treated quite distinctly. The practicality of our algorithms is important since interest in submodularity, owing to its natural and wide applicability, has recently been in ascendance within machine learning. We analyze theoretical properties of our algorithms for minimization and maximization, and show that many state-of-the-art maximization algorithms are special cases. Lastly, we complement our theoretical analyses with supporting empirical experiments.
منابع مشابه
Fast Semidifferential-based Submodular Function Optimization
We present a practical and powerful new framework for both unconstrained and constrained submodular function optimization based on discrete semidifferentials (suband super-differentials). The resulting algorithms, which repeatedly compute and then efficiently optimize submodular semigradients, offer new and generalize many old methods for submodular optimization. Our approach, moreover, takes s...
متن کاملSubmodular-Bregman and the Lovász-Bregman Divergences with Applications: Extended Version
We introduce a class of discrete divergences on sets (equivalently binary vectors)that we call the submodular-Bregman divergences. We consider two kinds ofsubmodular Bregman divergence, defined either from tight modular upper or tightmodular lower bounds of a submodular function. We show that the properties ofthese divergences are analogous to the (standard continuous) Bregman d...
متن کاملFast Multi-Stage Submodular Maximization: Extended version
Motivated by extremely large-scale machine learning problems, we introduce a new multistage algorithmic framework for submodular maximization (called MultGreed), where at each stage we apply an approximate greedy procedure to maximize surrogate submodular functions. The surrogates serve as proxies for a target submodular function but require less memory and are easy to evaluate. We theoreticall...
متن کاملOn fast approximate submodular minimization: Extended Version
We are motivated by an application to extract a representative subset of machine learning training data and by the poor empirical performance we observe of the popular minimum norm algorithm. In fact, for our application, minimum norm can have a running time of about O(n) (O(n) oracle calls). We therefore propose a fast approximate method to minimize arbitrary submodular functions. For a large ...
متن کاملMonotone Closure of Relaxed Constraints in Submodular Optimization: Connections Between Minimization and Maximization: Extended Version
It is becoming increasingly evident that many machine learning problems may be reduced to some form of submodular optimization. Previous work addresses generic discrete approaches and specific relaxations. In this work, we take a generic view from a relaxation perspective. We show a relaxation formulation and simple rounding strategy that, based on the monotone closure of relaxed constraints, r...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2013