Skip to yearly menu bar Skip to main content


Parameter Learning for Log-supermodular Distributions

Tatiana Shpakova · Francis Bach

Area 5+6+7+8 #67

Keywords: [ Structured Prediction ] [ Variational Inference ] [ Stochastic Methods ] [ (Other) Probabilistic Models and Methods ] [ (Application) Computer Vision ] [ Combinatorial Optimization ]


We consider log-supermodular models on binary variables, which are probabilistic models with negative log-densities which are submodular. These models provide probabilistic interpretations of common combinatorial optimization tasks such as image segmentation. In this paper, we focus primarily on parameter estimation in the models from known upper-bounds on the intractable log-partition function. We show that the bound based on separable optimization on the base polytope of the submodular function is always inferior to a bound based on ``perturb-and-MAP'' ideas. Then, to learn parameters, given that our approximation of the log-partition function is an expectation (over our own randomization), we use a stochastic subgradient technique to maximize a lower-bound on the log-likelihood. This can also be extended to conditional maximum likelihood. We illustrate our new results in a set of experiments in binary image denoising, where we highlight the flexibility of a probabilistic model to learn with missing data.

Live content is unavailable. Log in and register to view live content