Spatial Bayesian Nonparametrics for Natural Image Segmentation Erik - - PowerPoint PPT Presentation

spatial bayesian nonparametrics for natural image
SMART_READER_LITE
LIVE PREVIEW

Spatial Bayesian Nonparametrics for Natural Image Segmentation Erik - - PowerPoint PPT Presentation

Spatial Bayesian Nonparametrics for Natural Image Segmentation Erik Sudderth Brown University Joint work with Soumya Ghosh Michael Jordan University of California Brown University Parsing Visual Scenes dome sky skyscraper sky buildings


slide-1
SLIDE 1

Spatial Bayesian Nonparametrics for Natural Image Segmentation

Erik Sudderth

Brown University

Joint work with

Michael Jordan

University of California

Soumya Ghosh

Brown University

slide-2
SLIDE 2

Parsing Visual Scenes

trees skyscraper sky bell dome temple buildings sky

slide-3
SLIDE 3

Region Classification with Markov Field Aspect Models

Local: 74% MRF: 78% Verbeek & Triggs, CVPR 2007

slide-4
SLIDE 4

Human Image Segmentation

slide-5
SLIDE 5

Berkeley Segmentation Database & Boundary Detection Benchmark

slide-6
SLIDE 6

BNP Image Segmentation

  • ! How many regions does this image contain?
  • ! What are the sizes of these regions?

Segmentation as Partitioning

  • ! Huge variability in segmentations across images
  • ! Want multiple interpretations, ranked by probability

Why Bayesian Nonparametrics?

slide-7
SLIDE 7

The Infinite Hype

  • ! Infinite Gaussian Mixture Models
  • ! Infinite Hidden Markov Models
  • ! Infinite Mixtures of Gaussian Process Experts
  • ! Infinite Latent Feature Models
  • ! Infinite Independent Components Analysis
  • ! Infinite Hidden Markov Trees
  • ! Infinite Markov Models
  • ! Infinite Switching Linear Dynamical Systems
  • ! Infinite Factorial Hidden Markov Models
  • ! Infinite Probabilistic Context Free Grammars
  • ! Infinite Hierarchical Hidden Markov Models
  • ! Infinite Partially Observable Markov Decision Processes
  • ! !
slide-8
SLIDE 8

Some Hope: BNP Segmentation

Inference !! Stochastic search & expectation propagation Model !! Dependent Pitman-Yor processes !! Spatial coupling via Gaussian processes Results !! Multiple segmentations of natural images Learning !! Conditional covariance calibration

slide-9
SLIDE 9

Pitman-Yor Processes

The Pitman-Yor process defines a distribution on infinite discrete measures, or partitions

Dirichlet process:

1

slide-10
SLIDE 10

Pitman-Yor Stick-Breaking

slide-11
SLIDE 11

Human Image Segmentations

Labels for more than 29,000 segments in 2,688 images of natural scenes

slide-12
SLIDE 12

Statistics of Human Segments

How many objects are in this image?

Many Small Objects Some Large Objects

Object sizes follow a power law

Labels for more than 29,000 segments in 2,688 images of natural scenes

slide-13
SLIDE 13

Why Pitman-Yor?

Jim Pitman Marc Yor

Generalizing the Dirichlet Process !! Distribution on partitions leads to a generalized Chinese restaurant process !! Special cases of interest in probability: Markov chains, Brownian motion, ! Power Law Distributions DP PY

Number of unique clusters in N

  • bservations

Heaps Law:

Size of sorted cluster weight k

Goldwater, Griffiths, & Johnson, 2005 Teh, 2006

Natural Language Statistics

Zipfs Law:

slide-14
SLIDE 14

Feature Extraction

  • ! Partition image into ~1,000 superpixels
  • ! Compute texture and color features:

Texton Histograms (VQ 13-channel filter bank) Hue-Saturation-Value (HSV) Color Histograms

  • ! Around 100 bins for each histogram
slide-15
SLIDE 15

Pitman-Yor Mixture Model

Observed features (color & texture) Assign features to segments PY segment size prior Visual segment appearance model Color: Texture:

π z1 z2 z3 z4 x1 x2 x3 x4

xc

i ∼ Mult(θc zi)

xs

i ∼ Mult(θs zi)

zi ∼ Mult(π)

πk = vk

k−1

  • ℓ=1

(1 − vℓ) vk ∼ Beta(1 − a, b + ka)

slide-16
SLIDE 16

Dependent DP&PY Mixtures

Observed features (color & texture) Visual segment appearance model Color: Texture:

z1 z2 z3 z4 x1 x2 x3 x4

xc

i ∼ Mult(θc zi)

xs

i ∼ Mult(θs zi)

π1 π2 π3 π4

Assign features to segments

zi ∼ Mult(πi)

Some dependent prior with DP/PY “like” marginals Kernel/logistic/probit stick-breaking process,

  • rder-based DDP,

!

slide-17
SLIDE 17

Example: Logistic of Gaussians

  • ! Pass set of Gaussian processes through softmax to get

probabilities of independent segment assignments

  • ! Nonparametric analogs have similar properties

Figueiredo et. al., 2005, 2007 Fernandez & Green, 2002 Woolrich & Behrens, 2006 Blei & Lafferty, 2006

slide-18
SLIDE 18

Discrete Markov Random Fields

Ising and Potts Models

  • ! Interactive foreground segmentation
  • ! Supervised training for known categories

Previous Applications

!but learning is challenging, and little success at unsupervised segmentation.

GrabCut: Rother, Kolmogorov, & Blake 2004 Verbeek & Triggs, 2007

slide-19
SLIDE 19

Phase Transitions in Action

Potts samples, 10 states sorted by size: largest in blue, smallest in red

slide-20
SLIDE 20

Product of Potts and DP?

Orbanz & Buhmann 2006 Potts Potentials DP Bias:

slide-21
SLIDE 21

Spatially Dependent Pitman-Yor

  • ! Cut random surfaces

(samples from a GP) with thresholds

(as in Level Set Methods)

  • ! Assign each pixel to

the first surface which exceeds threshold

(as in Layered Models)

Duan, Guindani, & Gelfand, Generalized Spatial DP, 2007

π z1 z2 z3 z4 x1 x2 x3 x4

slide-22
SLIDE 22

Spatially Dependent Pitman-Yor

  • ! Cut random surfaces

(samples from a GP) with thresholds

(as in Level Set Methods)

  • ! Assign each pixel to

the first surface which exceeds threshold

(as in Layered Models)

Duan, Guindani, & Gelfand, Generalized Spatial DP, 2007

slide-23
SLIDE 23

Spatially Dependent Pitman-Yor

  • ! Cut random surfaces

(samples from a GP) with thresholds

(as in Level Set Methods)

  • ! Assign each pixel to

the first surface which exceeds threshold

(as in Layered Models)

  • ! Retains Pitman-Yor

marginals while jointly modeling rich spatial dependencies

(as in Copula Models)

slide-24
SLIDE 24

Spatially Dependent Pitman-Yor

Non-Markov Gaussian Processes: PY prior: Segment size Feature Assignments

Normal CDF

slide-25
SLIDE 25

Samples from PY Spatial Prior

Comparison: Potts Markov Random Field

slide-26
SLIDE 26

Outline

Inference !! Stochastic search & expectation propagation Model !! Dependent Pitman-Yor processes !! Spatial coupling via Gaussian processes Results !! Multiple segmentations of natural images Learning !! Conditional covariance calibration

slide-27
SLIDE 27

Mean Field for Dependent PY

K K

Factorized Gaussian Posteriors Sufficient Statistics

Allows closed form update of via

slide-28
SLIDE 28

Robustness and Initialization

Log-likelihood bounds versus iteration, for many random initializations of mean field variational inference on a single image.

slide-29
SLIDE 29

Alternative: Inference by Search

Consider hard assignments of superpixels to layers (partitions) Integrate likelihood parameters analytically (conjugacy) Marginalize layer support functions via expectation propagation (EP): approximate but very accurate

No need for a finite, conservative model truncation!

slide-30
SLIDE 30

Discrete Search Moves

!! Merge: Combine a pair of regions into a single region !! Split: Break a single region into a pair of regions (for diversity, a few proposals) !! Shift: Sequentially move single superpixels to the most probable region !! Permute: Swap the position

  • f two layers in the order

Stochastic proposals, accepted if and only if they improve our EP estimate of marginal likelihood: Marginalization of continuous variables simplifies these moves!

slide-31
SLIDE 31

Inference Across Initializations

Mean Field Variational EP Stochastic Search Best Worst Best Worst

slide-32
SLIDE 32

BSDS: Spatial PY Inference

Spatial PY (EP) Spatial PY (MF)

slide-33
SLIDE 33

Outline

Inference !! Stochastic search & expectation propagation Model !! Dependent Pitman-Yor processes !! Spatial coupling via Gaussian processes Results !! Multiple segmentations of natural images Learning !! Conditional covariance calibration

slide-34
SLIDE 34

Covariance Kernels

  • ! Thresholds determine segment size: Pitman-Yor
  • ! Covariance determines segment shape:

Roughly Independent Image Cues:

Berkeley Pb (probability of boundary) detector

probability that features at locations are in the same segment

!! Color and texture histograms within each region: Model generatively via multinomial likelihood (Dirichlet prior) !! Pixel locations and intervening contour cues: Model conditionally via GP covariance function

slide-35
SLIDE 35

Learning from Human Segments

!! Data unavailable to learn models of all the categories we’re interested in: We want to discover new categories! !! Use logistic regression, and basis expansion of image cues, to learn binary “are we in the same segment” predictors:

!! Generative: Distance only !! Conditional: Distance, intervening contours, !

slide-36
SLIDE 36

From Probability to Correlation

There is an injective mapping between covariance and the probability that two superpixels are in the same segment.

slide-37
SLIDE 37

Low-Rank Covariance Projection

!! The pseudo-covariance constructed by considering each superpixel pair independently may not be positive definite !! Projected gradient method finds low rank (factor analysis), unit diagonal covariance close to target estimates

slide-38
SLIDE 38

Prediction of Test Partitions

Heuristic versus Learned Image Partition Probabilities Learned Probability versus Rand index measure

  • f partition overlap
slide-39
SLIDE 39

Comparing Spatial PY Models

Image PY Learned PY Heuristic

slide-40
SLIDE 40

Outline

Inference !! Stochastic search & expectation propagation Model !! Dependent Pitman-Yor processes !! Spatial coupling via Gaussian processes Results !! Multiple segmentations of natural images Learning !! Conditional covariance calibration

slide-41
SLIDE 41

Other Segmentation Methods

FH Graph Mean Shift NCuts gPb+UCM Spatial PY

slide-42
SLIDE 42

Quantitative Comparisons

Berkeley Segmentation LabelMe Scenes !! On BSDS, similar or better than all methods except gPb !! On LabelMe, performance of Spatial PY is better than gPb !! Implementation efficiency and search run-time !! Histogram likelihoods discard too much information !! Most probable segmentation does not minimize Bayes risk Room for Improvement:

slide-43
SLIDE 43

Multiple Spatial PY Modes

Most Probable

slide-44
SLIDE 44

Multiple Spatial PY Modes

Most Probable

slide-45
SLIDE 45

Spatial PY Segmentations

slide-46
SLIDE 46

Conclusions

Successful BNP modeling requires! !! careful study of how model assumptions match data statistics & model comparisons !! reliable, consistent (general-purpose?) inference algorithms, carefully validated !! methods for learning hyperparameters from data, often with partial supervision