Do Sparsity Promoting Hierarchical Prior Models Plausibly Model Empirical Image Data

Yash Dave Co-Author
University of California, Berkeley
 
Brandon Marks Co-Author
University of California, Berkeley
 
Zixun Wang Co-Author
University of California, Berkeley
 
Hannah Chung Co-Author
University of California, Berkeley
 
Alexander Strang First Author
University of California, Berkeley
 
Yash Dave Presenting Author
University of California, Berkeley
 
Tuesday, Aug 5: 10:30 AM - 12:20 PM
1443 
Contributed Posters 
Music City Center 
Scale mixtures of normal distributions are a popular family of hierarchical Bayesian models that compromise interpretability, flexibility, and tractability. Recent work has demonstrated that generalized gamma mixtures of normals admit efficient algorithms that allow inference, uncertainty quantification, and hyper-parameter tuning in large, sparse, ill-determined inverse problems. We demonstrate parameter choices that produce popular prior families (Gaussian, Laplace, Student-t), and relate the distinguishability of priors to level sets of low-order moments in the parameter space using the KL and KS distances for power and significance respectively. We test the empirical validity of the hierarchical priors in a series of large imaging data sets. As case studies, we consider benchmark remote sensing, MRI, and image segmentation data sets. We report plausible ranges of parameters under standard representations (Fourier, wavelet, etc.). We find that, relative to the data sets tested, previous computational work has focused on an overly narrow subset of the space of available priors.

Keywords

Bayesian hierarchical models

Sparse inference

Image data

Scale mixture models

Compressed sensing

Model validation 

Main Sponsor

Section on Bayesian Statistical Science