Block Stability for MAP Inference

Hunter Lang, David Sontag, Aravindan Vijayaraghavan

Research output: Contribution to journalArticlepeer-review


To understand the empirical success of approximate MAP inference, recent work (Lang et al., 2018) has shown that some popular approximation algorithms perform very well when the input instance is stable. The simplest stability condition assumes that the MAP solution does not change at all when some of the pairwise potentials are (adversarially) perturbed. Unfortunately, this strong condition does not seem to be satisfied in practice. In this paper, we introduce a significantly more relaxed condition that only requires blocks (portions) of an input instance to be stable. Under this block stability condition, we prove that the pairwise LP relaxation is persistent on the stable blocks. We complement our theoretical results with an empirical evaluation of real-world MAP inference instances from computer vision. We design an algorithm to find stable blocks, and find that these real instances have large stable regions. Our work gives a theoretical explanation for the widespread empirical phenomenon of persistency for this LP relaxation.

Original languageEnglish (US)
JournalUnknown Journal
StatePublished - Oct 11 2018

ASJC Scopus subject areas

  • General

Fingerprint Dive into the research topics of 'Block Stability for MAP Inference'. Together they form a unique fingerprint.

Cite this