Abstract
Nonparametric Bayesian methods are considered for recovery of imagery based upon compressive, incomplete, and/or noisy measurements. A truncated beta-Bernoulli process is employed to infer an appropriate dictionary for the data under test and also for image recovery. In the context of compressive sensing, significant improvements in image recovery are manifested using learned dictionaries, relative to using standard orthonormal image expansions. The compressive-measurement projections are also optimized for the learned dictionary. Additionally, we consider simpler (incomplete) measurements, defined by measuring a subset of image pixels, uniformly selected at random. Spatial interrelationships within imagery are exploited through use of the Dirichlet and probit stick-breaking processes. Several example results are presented, with comparisons to other methods in the literature.
Index Terms: Bayesian nonparametrics, compressive sensing, dictionary learning, factor analysis, image denoising, image interpolation, sparse coding
I. Introduction
A. Sparseness and Dictionary Learning
Recently, there has been significant interest in sparse image representations, in the context of denoising and interpolation [1], [13], [24]–[26], [27], [28], [31], compressive sensing (CS) [5], [12], and classification [40]. All of these applications exploit the fact that images may be sparsely represented in an appropriate dictionary. Most of the denoising, interpolation, and CS literature assume “off-the-shelf” wavelet and DCT bases/dictionaries [21], but recent research has demonstrated the significant utility of learning an often overcomplete dictionary matched to the signals of interest (e.g., images) [1], [4], [12], [13], [24]–[26], [27], [28], [30], [31], [41].
Many of the existing methods for learning dictionaries are based on solving an optimization problem [1], [13], [24]–[26], [27], [28], in which one seeks to match the dictionary to the imagery of interest, while simultaneously encouraging a sparse representation. These methods have demonstrated state-of-the-art performance for denoising, superresolution, interpolation, and inpainting. However, many existing algorithms for implementing such ideas also have some restrictions. For example, one must often assume access to the noise/residual variance, the size of the dictionary is set a priori or fixed via cross-validation-type techniques, and a single (“point”) estimate is learned.
To mitigate the aforementioned limitations, dictionary learning has recently been cast as a factor-analysis problem, with the factor loading corresponding to the dictionary elements (atoms). Utilizing nonparametric Bayesian methods like the beta process (BP) [29], [37], [42] and the Indian buffet process (IBP) [18], [22], one may, for example, infer the number of factors (dictionary elements) needed to fit the data. Furthermore, one may place a prior on the noise or residual variance, with this inferred from the data [42]. An approximation to the full posterior may be manifested via Gibbs sampling, yielding an ensemble of dictionary representations. Recent research has demonstrated that an ensemble of representations can be better than a single expansion [14], with such an ensemble naturally manifested by statistical models of the type described here.
B. Exploiting Structure and Compressive Measurements
In image analysis, there is often additional information that may be exploited when learning dictionaries, with this well suited for Bayesian priors. For example, most natural images may be segmented, and it is probable that dictionary usage will be similar for regions within a particular segment class. To address this idea, we extend the model by employing a probit stick-breaking process (PSBP), with this generalization of the Dirichlet process (DP) stick-breaking representation [34]. Related clustering techniques have proven successful in image processing [26]. The model clusters the image patches, with each cluster corresponding to a segment type; the PSBP encourages proximate and similar patches to be included within the same segment type, thereby performing image segmentation and dictionary learning simultaneously.
The principal focus of this paper is on applying hierarchical Bayesian algorithms to new compressive measurement techniques that have been recently developed. Specifically, we consider dictionary learning in the context of CS [5], [9], in which the measurements correspond to projections of typical image pixels. We consider dictionary learning performed “offline” based on representative (training) images, with the learned dictionary applied within CS image recovery. We also consider the case for which the underlying dictionary is simultaneously learned with inversion (reconstruction), with this related to “blind” CS [17]. Finally, we design the CS projection matrix to be matched to the learned dictionary (when this is done offline), and demonstrate, as in [12], that in practice, this yields performance gains relative to conventional random CS projection matrices.
While CS is of interest for its potential to reduce the number of required measurements, it has the disadvantage of requiring the development of new classes of cameras. Such cameras are revolutionary and interesting [11], [35], but there have been decades of previous research performed on development of pixel-based cameras, and it would be desirable if such cameras could be modified simply to perform compressive measurements. We demonstrate that one may perform compressive measurements of natural images by simply sampling pixels measured by a conventional camera, with the pixels uniformly selected at random. This is closely related to recent research on matrix completion [6], [23], [33], but here, we move beyond simple low-rank assumptions associated with such previous research.
C. Contributions
This paper develops several hierarchical Bayesian models for learning dictionaries for analysis of imagery, with applications in denoising, interpolation, and CS. The inference is performed based on a Gibbs sampler, as is increasingly common in modern image processing [16]. Here, we demonstrate how generalizations of the beta-Bernoulli process allow one to infer the dictionary elements directly based on the underlying degraded image, without any a priori training data, while simultaneously inferring the noise statics and tolerating significant missingness in the imagery. This is achieved by exploiting the low-dimensional structure of most natural images, which implies that image patches may be represented in terms of a low-dimensional set of learned dictionary elements. Excellent results are realized in these applications, including for hyperspectral imagery, which has not been widely considered in these settings previously. We also demonstrate how the learned dictionary may be easily employed to define CS projection matrices that yield markedly improved image recovery, as compared with the much more typical random construction of such matrices.
The basic hierarchical Bayesian architecture developed here serves as a foundation that may be flexibly extended to incorporate additional prior information. As examples, we show here how dictionary learning may be readily coupled with clustering, through the use of a DP [15]. We also incorporate spatial information within the image via a PSBP [32], with these extensions yielding significant advantages for the CS application. The basic modeling architecture may also exploit additional information manifested in the form of general covariates, as considered in a recent paper [43].
The remainder of the paper is organized as follows: In Section II, we review the classes of problems being considered. The beta-Bernoulli process is discussed in Section III, with relationships made with previous work in this area, including those based on the IBP. The Dirichlet and PSBPs are discussed in Section IV, and several example results are presented in Section V. Conclusions and a discussion of future work are provided in Section VI, and details of the inference equations are summarized in the Appendix.
II. Problems Under Study
We consider data samples that may be expressed in the following form:
(1) |
where xi ∈ ℝP, εi ∈ ℝP, and wi ∈ ℝK. The columns of matrix D ∈ ℝP×K represent the K components of a dictionary with which xi is expanded. For our problem, the xi will correspond to B×B (overlapped) pixel patches in an image [1], [13], [24], [25], [27], [42]. The set of vectors{xi}i=1,N may be extracted from an image(s) of interest.
For the denoising problem, vectors εi may represent sensor noise, in addition to (ideally small) residual from representation of the underlying signal as Dwi. To perform denoising, we place restrictions on vectors wi, such that Dwi by itself does not exactly represent xi. A popular such restriction is that wi should be sparse, motivated by the idea that any particular xi may often be represented in terms of a small subset of representative dictionary elements, from the full dictionary defined by the columns of D. There are several methods that have been recently developed to impose such a sparse representation, including ℓ1-based relaxation algorithms [24], [25], iterative algorithms [1], [13], and Bayesian methods [42]. One advantage of a Bayesian approach is that the noise/residual statistics may be nonstationary (with unknown noise statistics). Specifically, in addition to placing a sparseness-promoting prior on wi, we may also impose a prior on the components of εi. From the estimated posterior density function on model parameters, each component of εi, corresponding to the ith B×B image patch, has its own variance. Given{xi}i=1,N, our goal may be to simultaneously infer D and {wi}i=1,N (and implicitly εi), and then, the denoised version of xi is represented as Dwi.
In many applications, the total number of pixels N · P may be large. However, it is well known that compression algorithms may be used on {xi}i=1,N after the measurements have been performed, to significantly reduce the quantity of data that need be stored or communicated. This compression indicates that while the data dimensionality N · P may be large, the underlying information content may be relatively low. This has motivated the field of CS [5], [9], [11], [35], in which the total number of measurements performed may be much less than N · P. Toward this end, researchers have proposed projection measurements of the following form:
(2) |
where Σ ∈ ℝn×P and yi ∈ ℝn, ideally with n ≪ P. The projection matrix Σ has traditionally been randomly constituted [5], [9], with a binary or real alphabet (and Σ may be also a function of the specific patch, and generalized as Σi). It is desirable that matrices Σ and D be as incoherent as possible.
The recovery of xi from yi is an ill-posed problem, unless restrictions are placed on xi. We may exploit the same class of restrictions used in the denoising problem; specifically, the observed data satisfy yi = Φwi + νi, with Φ = ΣD and νi = Σεi, and with sparse wi. Note that the sparseness constraint implies that {wi}i=1,N (and hence, {xi}i=1,N) occupy distinct subspaces of ℝP, selected from the overall linear subspace defined by the columns of Φ.
In most applications of CS, D is assumed known, corresponding to an orthonormal basis (e.g., wavelets or a DCT) [5], [9], [21]. However, such bases are not necessarily well matched to natural imagery, and it is desirable to consider design of dictionaries D for this purpose [12]. One may even consider recovering {xi}i=1,N from {yi}i=1,N while simultaneously inferring D. Thus, we again have a dictionary-learning problem, which may be coupled with optimization of CS matrix Σ, such that it is matched to D (defined by a low coherence between the rows of Σ and columns of D [5], [9], [12], [21]).
III. Sparse Factor Analysis With the Beta-Bernoulli Process
When presenting example results, we will consider three problems. For denoising, it is assumed that we measure xi = Dwi + εi, where εi represents measurement noise and model error; for the compressive-sensing application, we observe yi = Σ(Dwi + εi) = Φwi + νi, with Φ = ΣD and νi = Σεi; and finally, for the interpolation problem, we observe Qφ(Dwi+ εi), where Qφ(xi) is a vector of elements from xi contained within set φ. For all three problems, our objective is to infer underlying signal Dwi, with wi assumed sparse; we generally wish to simultaneously infer D and {wi}i=1,N. To address each of these problems, we consider a statistical model for xi = Dwi + εi, placing Bayesian priors on D, wi, and εi; the way the model is used is slightly modified for each specific application. For example, when considering interpolation, only the observed Qφ(xi) are used within the model likelihood function.
A. Beta-Bernoulli Process for Active-Set Selection
Let binary vector zi ∈ {0,1}K denote which of the K columns of D are used for representation of xi (active set); if a particular component of zi is equal to one, then the corresponding column of D is used in the representation of xi. Hence, for data {xi}i=1,N, there is an associated set of latent binary vectors {zi}i=1,N, and the beta-Bernoulli process provides a convenient prior for these vectors [29], [37], [42]. Specifically, consider the following model:
(3) |
where πk is the kth component of π, and a and b are model parameters; the impact of these parameters on the model are discussed below. Note that the use of product notation in (3) is meant to denote that each component of zi and π are independently drawn from distributions of the same form.
Considering limit K → ∞, and after integrating out π, the draws of {zi}i=1,N may be constituted as follows. For each zi, draw ci ~ Poisson(a/b + i − 1) and define , with C0 = 0. Let zik represent the kth component of zi, and zik = 0 for k > Ci. For k = 1,…,Ci−1, zik ~ Bernoulli(nik/b + i−1), where (nik represents the total number of times the kth component of {zj}j=1,i−1 is one). For k = Ci−1 + 1,…,Ci, we set zik = 1. Note that as a/(b + i−1) becomes small, with increasing i, it is probable that ci will be small. Hence, with increasing i, the number of new nonzero components of zi diminishes. Furthermore, as a consequence of Bernoulli(nik/b+i−1), when a particular component of vectors {zj}j=1,i−1 is frequently one, it is more probable that it will be one for subsequent zj, j≥i. When b = 1, this construction for {zi}i=1,N corresponds to the IBP [18].
Since zi defines which columns of D are used to represent xi, (3) imposes that it is probable that some columns of D are repeatedly used among set {xi}i=1,N, whereas other columns of D may be more specialized to particular xi. As demonstrated below, this has been found to be a good model when {xi}i=1,N are patches of pixels extracted from natural images.
B. Full Hierarchical Model
The hierarchical form of the model may now be expressed as
(4) |
where dk represents the kth component (atom) of D, ⊙ represents the elementwise or Hadamard vector product, IP (IK) represents a P×P (K×K) identity matrix, and {zi}i=1,N are drawn as in (3). Conjugate hyperpriors γs ~ Gamma(c,d) and γε ~ Gamma(e,f) are also imposed. The construction in (4), and with the prior in (3) for {zi}i=1,N, is henceforth referred to as the BP factor analysis (BPFA) model. This model was first developed in [22], with a focus on general factor analysis; here, we apply and extend this construction for image-processing applications.
Note that we impose independent Gaussian priors for dk and si and εi for modeling convenience (conjugacy of consecutive terms in the hierarchical model). However, the inferred posterior for these terms is generally not independent or Gaussian. The independent priors essentially impose prior information about the marginals of the posterior of each component, whereas the inferred posterior accounts for statistical dependence as reflected in the data.
To make connections of this model to more typical optimization-based approaches [24], [25], note that the negative logarithm of the posterior density function is
(5) |
where Θ represents all unknown model parameters, = {xi}i=1,N, represents the beta-Bernoulli process prior in (3), and represents model hyperparameters (i.e., a, b, c, d, e, and f). Therefore, the typical ℓ2 constraints [24], [25] on the dictionary elements dk and on the nonzero weights si correspond here to the Gaussian priors employed in (4). However, rather than employing an ℓ1 (Laplacian prior) constraint [24], [25] to impose sparseness on wi, we employ the beta-Bernoulli process and wi = si ⊙ zi. The beta-Bernoulli process imposes that binary zi should be sparse and that there should be a relatively consistent (re)use of dictionary elements across the image, thereby also imposing self-similarity. Furthermore, and perhaps most importantly, we do not constitute a point estimate, as one would do if a single Θ was sought to minimize (5). We rather estimate the full posterior density p(Θ| ), implemented via Gibbs sampling. A significant advantage of the hierarchical construction in (4) is that each Gibbs update equation is analytic, with detailed update equations provided in the Appendix. Note that consistent use of atoms is encouraged because the active sets are defined by binary vectors {zi}i=1,N, and these are all drawn from a shared probability vector π; this is distinct from drawing the active sets independent and identically distributed (i.i.d.) from a Laplacian prior. Furthermore, the beta-Bernoulli prior imposes that many components of wi are exactly zero, whereas with a Laplacian prior, many components are small but not exactly zero (hence, the former is analogous to ℓ0 regularization, with the latter closer to ℓ1 regularization).
IV. Patch Clustering Via Dirichlet and PSBPs
A. Motivation
In the model previously discussed, each patch xi had a unique usage of dictionary atoms, defined by binary vector zi, which selects columns of D. One may wish to place further constraints on the model, thereby imposing a greater degree of statistical structure. For example, one may employ that the xi cluster and that within each cluster of each of the associated xi employ the same columns of D. This is motivated by the idea that a natural image may be clustered into different types of textures or general image structure. However, rather than imposing that all xi within a given cluster use exactly the same columns of D, one may want to impose that all xi within such a cluster share the same probability of dictionary usage, i.e., that all xi within cluster c share the same probability of using columns of D, defined by πc, rather than sharing a single π for all xi (as in the original model above). Again, such clustering is motivated by the idea that natural images tend to segment into different textural or color forms. Below, we perform clustering in terms of vectors πc, rather than explicit clustering of dictionary usage, which would entail cluster-dependent zc; the “softer” nature of the former clustering structure is employed to retain model flexibility while still encouraging sharing of parameters within clusters.
A question when performing such clustering concerns the number of clusters needed, this motivating the use of nonparametric methods, like those considered in the next subsections. Additionally, since the aforementioned clustering is motivated by the segmentations characteristic of natural images, it is desirable to explicitly utilize the spatial location of each image patch, encouraging that patches xi in a particular segment/cluster are spatially contiguous. This latter goal motivates use of a PSBP, as also detailed as follows.
B. DP
The DP [15] constitutes a popular means of performing non-parametric clustering. A random draw from a DP, i.e., G ~ DP(αG0), with precision α ∈ ℝ+ and “base” measure G0, may be constituted via the stick-breaking construction [34], i.e.,
(6) |
where and Vh ~ Beta(1, α). The βl may be viewed as a sequence of fractional breaks from a “stick” of original length one, where the fraction of the stick broken off on break l is Vl. The are model parameters, associated with the lth data cluster. For our problem, it has proven effective to set analogous to (3), and hence, , drawn from G0, corresponds to distinct probability vectors for using the K dictionary elements (columns of D). For sample i, we draw πi ~ G, and a separate sparse binary vector zi is drawn for each sample xi, as , with πik as the kth component of πi. In practice, we truncate the infinite sum for the to G to NL elements and impose VNL = 1, such that . A (conjugate) gamma prior is placed on the DP parameter α.
We may view this DP construction as an “Indian buffet franchise, ” generalizing the Indian buffet analogy [18]. Specifically, there are NL Indian buffet restaurants; each restaurant is composed of the same “menu” (columns of D) and is distinguished by different probabilities for selecting menu items. The “customers” {xi}i=1,N cluster based upon which restaurant they go to. The represent the probability of using each column of D in the respective NL different buffets. The {xi}i=1,N cluster themselves among the different restaurants in a manner that is consistent with the characteristics of the data, with the model also simultaneously learning the dictionary/menu D. Note that we typically make truncation NL large, and the posterior distribution infers the number of clusters actually needed to support the data, as represented by how many βl are of significant value. The model in (4), with the above DP construction for {zi}i=1,N, is henceforth referred to as DP-BPFA.
C. PSBP
The DP yields a clustering of {xi}i=1,N, but it does not account for our knowledge of the location of each patch within the image. It is natural to expect that if xi and xi′ are proximate, then, they are likely to be constituted in terms of similar columns of D.1 To impose this information, we employ the PSBP. A logistic stick-breaking process is discussed in detail in [32]. We employ the closely related probit version here because it may be easily implemented in a Gibbs sampler. We note that while the method in [32] is related to that discussed below, in [32], the concepts of learned dictionaries and beta-Bernoulli priors were not considered. Another related model, which employs a probit link function, is discussed in [7].
We augment the data as {xi, ri}i=1,N, where xi again represents pixel values from the ith image patch, and ri ∈ ℝ2 represents the 2-D location of each patch. We wish to impose that proximate patches are more likely to be composed of the same or similar columns of D. In the PSBP construction, all aspects of (4) are retained, except for the manner in which zi are constituted. Rather than drawing a single K-dimensional vector of probabilities π, as in (3), we draw a library of such vectors, i.e.,
(7) |
and each is associated with a particular segment in the image. One πi is associated with location ri and drawn
(8) |
with for all ri, and represents a point measure concentrated at . Once πi is associated with a particular xi, the corresponding binary vector zi is drawn as in the first line of (3). Note that the distinction between DP and PSBP is that, in the former, mixture weights {βl}l=1,NL are independent of spatial position r, whereas the latter explicitly utilizes r within {βl(r)}l=1,NL (and below, we impose that βl(r) smoothly changes with r).
The space-dependent weights are constructed as , where 0 < Vl(r) < 1 constitute space-dependent probabilities. We set VNL = 1, and for l ≤ NL − 1, the Vl are space-dependent probit functions, i.e.,
(9) |
where (r, ri; ψl) is a kernel characterized by parameter ψl and {ζli}i=0,N is a sparse set of real numbers. To implement the sparseness on {ζli}i=0,N, within prior , and (conjugate) αli ~ Gamma(a0, b0), with (a0, b0) set to favor most αli being large (if αli is large, draw is likely to be near zero, such that most {ζli}i=0,N are near zero). This sparseness-promoting construction is the same as that employed in the relevance vector machine (RVM) [39]. We here utilize radial basis function kernel (r, ri; ψl) = exp[−||ri−r||2/ψl].
Each gl(r) is encouraged to only be defined by a small set of localized kernel functions, and via the probit link function , the probability Vl(r) is characterized by localized segments over which the probability Vl(r) is contiguous and smoothly varying. The Vl(r) constitute a space-dependent stick-breaking process. Since VNL = 1, for all r.
The PSBP model is relatively simple to implement within a Gibbs sampler. For example, as indicated above, sparseness on ζli is imposed as in the RVM, and the probit link function is simply implemented within a Gibbs sampler (which is why it was selected, rather than a logistic link function). Finally, we define a finite set of possible kernel parameters {ψj}j=1,Np, and a multinomial prior is placed on these parameters, with the multinomial probability vector drawn from a Dirichlet distribution [32] (each of the gl(r) draws a kernel parameter from {ψj}j=1,Np). The model in (4), with the PSBP construction for {zi}i=1,N, is henceforth referred to as PSBP-BPFA.
D. Discussion of Proposed Sparseness-Imposing Priors
The basic BPFA model is summarized in (4), and three related priors have been developed for sparse binary vectors {zi}i=1,N: 1) the basic truncated beta-Bernoulli process in (3); 2) a DP-based clustering of the underlying {πi}i=1,N; and 3) a PSBP clustering of {πi}i=1,N that exploits knowledge of the location of the image patches. For 2 and 3, xi within a particular cluster have similar zi, rather than exactly the same binary vector; we also considered the latter, but this worked less well in practice. As discussed further when presenting results, for denoising and interpolation, all three methods yield comparable performance. However, for CS, 2 and 3 yield marked improvements in image-recovery accuracy relative to 1. In anticipation of these results, we provide a further discussion of the three priors on {zi}i=1,N and on the three image-processing problems under consideration.
For the denoising and interpolation problems, we are provided with data {xi}i=1,N, albeit in the presence of noise and potentially with substantial missing pixels. However, for this problem, N may be made quite large since we may consider all possible (overlapping) B×B patches. A given pixel (apart from near the edges of the image) is present in B2 different patches. Perhaps because we have such a large quantity of partially overlapping data, for denoising and interpolation, we have found that beta-Bernoulli process in (3) is sufficient for inferring the underlying relationships between the different data {xi}i=1,N and processing these data collaboratively. However, the beta-Bernoulli construction does not explicitly segment the image, and therefore, an advantage of the PSBP-BPFA construction is that it yields comparable denoising and interpolation performance as (3) while also simultaneously yielding an effective image segmentation.
For the CS problem, we measure yi = Σxi, and therefore, each of the n measurements associated with each image patch (Σ ∈ ℝn×P)loses the original pixels in xi (the projection matrix Σ may also change with each patch, denoted by Σi). Therefore, for CS, one cannot consider all possible shifts of the patches as the patches are predefined and fixed in the CS measurement (in the denoising and interpolation problems, the patches are defined in the subsequent analysis). Therefore, for CS imposition of the clustering behavior, via DP or PSBP provides important information, yielding state-of-the-art CS-recovery results.
E. Possible Extensions
The “Indian buffet franchise” and PSBP constructions considered above and in the results below draw the K-dimensional probability vectors independently. This implies that within the prior, we impose no statistical correlation between the components of vectors and , for l ≠ l′. It may be desirable to impose such structure, imposing that there is a “global” probability of using particular dictionary elements, and the different mixture components within the DP/PSBP constructions correspond to specific draws from global statistics of dictionary usage. This will encourage the idea that there may be some “popular” dictionary elements that are shared across different mixture components (i.e., popular across different buffets in the franchise). One can impose this additional structure via a hierarchical BP (HBP) construction [37], related to the hierarchical DP (HDP) [36]. Briefly, in an HBP construction, one may draw the via the hierarchical construction, i.e.,
(10) |
where ηk is the kth component of η. Vector η constitutes “global” probabilities of using each of the K dictionary elements (across the “franchise”), and defines the probability of dictionary usage for the lth buffet. This construction imposes statistical dependences among vectors { }.
To simplify the presentation, in the following example results, we do not consider the HBP construction as good results have already been achieved with the (truncated) DP and PSBP models previously discussed. We note that in these analyses, the truncated versions of these models may actually help inference of statistical correlations among { } within the posterior (since the set { } is finite, and each vector is of finite length). If we actually considered the infinite limit on K and on the number of mixture components, inference of such statistical relationships within the posterior may be undermined because specialized dictionary elements may be constituted across the different franchises, rather than encouraging sharing of highly similar dictionary elements.
While we do not focus on the HBP construction here, a recent paper has employed the HBP construction in related dictionary learning for image-processing applications, yielding very encouraging results [43]. We therefore emphasize that the basic hierarchical Bayesian construction employed here is very flexible and may be extended in many ways to impose additional structure.
V. Example Results
A. Reproducible Research
The test results and the MATLAB code to reproduce them can be downloaded from http://www.ee.duke.edu/~mz1/Results/BPFAImage/.
B. Parameter Settings
For all BPFA, DP-BPFA, and PSBP-BPFA computations, the dictionary truncation level was set at K = 256 or K = 512 based on the size of the image. Not all K dictionary elements are used in the model; the truncated beta-Bernoulli process infers the subset of dictionary elements employed to represent the data {xi}i=1,N. The larger the image, the more distinct types of structure are anticipated and, therefore, the more dictionary elements are likely to be employed; however, very similar results are obtained with K = 512 in all examples, just with more dictionary elements are not employed for smaller images (therefore, to save computational resources, we set K = 256 for the smaller images). The number of DP and PSBP sticks was set at NL = 20. The library of PSBP parameters is defined as in [32]; the PSBP kernel locations, i.e., {ri}i=1,N, were situated on a uniformly sampled grid in each image dimension, situated at every fourth pixel in each direction (the results are insensitive to many related definitions of {ri}i=1,N). The hyperparameters within the gamma distributions were set as c = d = e = f = 10−6, as is typically done in models of this type [39] (the same settings were used for the gamma prior for DP precision parameter α. The beta-distribution parameters are set as a = K and b = 1 if random initialization is used or a = K and b = N/8 if a singular value decomposition (SVD) based initialization is used. None of these parameters have been optimized or tuned. When performing inference, all parameters are randomly initialized (as a draw from the associated prior) or based on the SVD of the image under test. The Gibbs samplers for the BPFA, DP-BPFA, and PSBP-BPFA have been found to mix and quickly converge, producing satisfactory results with as few as 20 iterations. The inferred images represent the average from the collection samples. All software was written in nonoptimized MATLAB. On a Dell Precision T3500 computer with a 2.4-GHz central processing unit, for N = 148, 836 patches of size 8×8×3 with 20% of the RGB pixels observed at random, the BPFA required about 2 min per Gibbs iteration (the DP version was comparable), and PSBP-BPFA required about 3 min per iteration. For the 106-band hyperspectral imagery, which employed N = 428 578 patches of size 4×84×106 with 2% of the voxels observed uniformly at random, each Gibbs iteration required about 15 min.
C. Denoising
The BPFA denoising algorithm is compared with the original KSVD [13] and for both gray-scale and colored images. Newer denoising algorithms include block matching with 3-D filtering [8], the multiscale KSVD [28], and KSVD with the nonlocal mean constraints [26]. These algorithms assume that the noise variance is known, whereas the proposed model automatically infers, as part of the same model, the noise variance from the image under test. There are existing methods for estimation of the noise variance, as a preprocessing step, e.g., via wavelet shrinkage [10]. However, it was shown in [42] that the denoising accuracy of methods like that in [13] can be sensitive to small errors in the estimated variance, which are likely to occur in practice. Additionally, when doing joint denoising and image interpolation, as we consider below, methods like that in [10] may not be directly applied to estimate the noise variance, as there is a large fraction of missing data. Moreover, the BPFA, DP-BPFA, and PSBP-BPFA models infer a potentially nonstationary noise variance, with a broad prior on the variance imposed by the gamma distribution.
In the denoising examples, we consider the BPFA model in (3); similar results are obtained via the DP-BPFA and PSBP-BPFA models discussed in Section IV. To infer the final signal value at each pixel, we average the associated pixel contribution from each of the overlapping patches in which it is included (this is true for all results presented below in which overlapping patches were used).
In Table I, we consider images from [13]. The proposed BPFA performs very similarly to KSVD. As one representative example of the model’s ability to infer the noise variance, we consider the Lena image from Table I. The mean inferred noise standard deviations are 5.83, 10.59, 15.53, 20.48, 25.44, 50.46, and 100.54 for images contaminated by noise with the respective standard deviations of 5, 10, 15, 20, 25, 50, and 100. Each of these noise variances was automatically inferred using exactly the same model, with no changes to the gamma hyperparameters (whereas for the KSVD results, it was assumed that the noise variance was known exactly a priori).
TABLE I.
σ | C.man | House | Peppers | Lena | Barbara | Boats | F.print | Couple | Hill |
---|---|---|---|---|---|---|---|---|---|
5 | 37.87 | 39.37 | 37.78 | 38.60 | 38.08 | 37.22 | 36.65 | 37.31 | 37.02 |
37.32 | 39.18 | 37.24 | 38.20 | 37.94 | 36.43 | 36.29 | 36.77 | 36.24 | |
10 | 33.73 | 35.98 | 34.28 | 35.47 | 34.42 | 33.64 | 32.39 | 33.52 | 33.37 |
33.40 | 36.29 | 34.31 | 35.62 | 34.63 | 33.70 | 32.42 | 33.63 | 33.31 | |
15 | 31.42 | 34.32 | 32.22 | 33.70 | 32.37 | 31.73 | 30.06 | 31.45 | 31.47 |
31.34 | 34.52 | 32.46 | 33.93 | 32.61 | 31.97 | 30.23 | 31.73 | 31.64 | |
20 | 29.91 | 33.20 | 30.82 | 32.38 | 30.83 | 30.36 | 28.47 | 30.00 | 30.18 |
30.03 | 33.25 | 31.10 | 32.65 | 31.10 | 30.70 | 28.72 | 30.34 | 30.47 | |
25 | 28.85 | 32.15 | 29.73 | 31.32 | 29.60 | 29.28 | 27.26 | 28.90 | 29.18 |
28.99 | 32.24 | 30.00 | 31.63 | 29.88 | 29.70 | 27.58 | 29.28 | 29.57 | |
50 | 25.73 | 27.95 | 26.13 | 27.79 | 25.47 | 25.95 | 23.24 | 25.32 | 26.27 |
25.67 | 28.49 | 26.46 | 28.29 | 26.03 | 26.50 | 24.14 | 25.94 | 26.81 | |
100 | 21.69 | 23.71 | 21.75 | 24.46 | 21.89 | 22.81 | 18.30 | 22.60 | 23.98 |
21.93 | 24.37 | 22.73 | 24.95 | 22.13 | 23.32 | 20.44 | 23.01 | 24.22 |
In Table II, we present similar results, for denoising RGB images; the KSVD comparisons come from [27]. An example denoising result is shown in Fig. 1. As another example of the BPFA’s ability to infer the underlying noise variance, for the castle image, the mean (automatically) inferred variances are 5.15, 10.18, 15.22, and 25.23 for images with additive noise with true respective standard deviations of 5, 10, 15, and 25. The sensitivity of the KSVD algorithm to a mismatch between the assumed and true noise variances is shown in [42, Fig. 1], and the insensitivity of BPFA to changes in the noise variance and to requiring knowledge of the noise variance is deemed an important advantage.
TABLE II.
σ | Castle | Mushroom | Train | Horses | Kangroo |
---|---|---|---|---|---|
5 | 40.37 | 39.93 | 39.76 | 40.09 | 39.00 |
40.34 | 39.73 | 39.38 | 39.96 | 39.00 | |
10 | 36.24 | 35.60 | 34.72 | 35.43 | 34.06 |
36.28 | 35.70 | 34.48 | 35.48 | 34.21 | |
15 | 33.98 | 33.18 | 31.70 | 32.76 | 31.30 |
34.04 | 33.41 | 31.63 | 32.98 | 31.68 | |
25 | 31.19 | 30.26 | 28.16 | 29.81 | 28.39 |
31.24 | 30.62 | 28.28 | 30.11 | 28.86 |
It is also important to note that the gray-scale KSVD results in Table I were initialized using an overcomplete DCT dictionary, whereas the RGB KSVD results in Table II employed an extensive set of training imagery to learn dictionary D that was used to initialize the denoising computations. All BPFA, DP-BPFA, and PSBP-BPFA results employ no training data, with the dictionary initialized at random using draws from the prior or with the SVD of the data under test.
D. Image Interpolation
For the initial interpolation examples, we consider standard RGB images, with 80% of the RGB pixels uniformly missing at random (the data under test are shown in Fig. 2). Results are first presented for the Castle and Mushroom images, with comparisons between the BPFA model in (3) and the PSBP-BPFA model discussed in Section IV. The difference between the two is that the former is a “bag-of-patches” model, whereas the latter accounts for the spatial locations of the patches. Furthermore, the PSBP-BPFA simultaneously performs image recovery and segmentation. The results are shown in Fig. 3, presenting the mean reconstructed images and inferred segmentations. Each color in the inferred segmentation represents one PSBP mixture component, and the figure shows the last Gibbs iteration (to avoid issues with label switching between Gibbs iterations). While the BPFA does not directly yield a segmentation, its peak signal-to-noise ratio (PSNR) results are comparable to those inferred by PSBP-BPFA, as summarized in Table III.
TABLE III.
Castle 8×8×3 | Castle 5×5×3 | Mushroom 8×8×3 | Mushroom 5×5×3 | |
---|---|---|---|---|
BPFA | 29.32 | 28.48 | 31.63 | 31.17 |
PSBP-BPFA | 29.54 | 28.46 | 32.03 | 31.27 |
An important additional advantage of Bayesian models such as BPFA, DP-BPFA, and PSBP-BPFA is that they provide a measure of confidence in the accuracy of the inferred image. In Fig. 4, we plot the variance of the inferred error {εi}i=1,N, computed via the Gibbs collection samples.
To provide a more thorough examination of model performance, in Table IV, we present results for several well-studied gray-scale and RGB images, as a function of the fraction of pixels missing. All of these results are based upon BPFA, with DP-BPFA and PSBP-BPFA yielding similar results. Finally, in Table V, we perform interpolation and denoising simultaneously, again with no training data and without prior knowledge of the noise level (again, as previously discussed, it would be difficult to estimate the noise variance as a preprocessing step in this case using methods like that in [10] as, in this case, there is also a large number of missing pixels). An example result is shown in Fig. 5. To our knowledge, this is the first time that denoising and interpolation have been jointly performed while simultaneously inferring the noise statistics.
TABLE IV.
data ratio | C.man | House | Peppers | Lena | Barbara | Boats | F.print | Man | Couple | Hill |
---|---|---|---|---|---|---|---|---|---|---|
20% | 24.11 | 30.12 | 25.92 | 31.00 | 24.80 | 27.81 | 26.03 | 28.24 | 27.72 | 29.33 |
30% | 25.71 | 33.14 | 28.19 | 33.31 | 27.52 | 30.00 | 09.01 | 30.06 | 30.00 | 31.21 |
50% | 28.90 | 38.02 | 32.58 | 36.94 | 33.17 | 33.78 | 33.53 | 33.29 | 35.56 | 34.23 |
80% | 34.70 | 43.03 | 37.73 | 41.27 | 40.76 | 39.50 | 40.17 | 39.11 | 38.71 | 38.75 |
data ratio | Castle | Mushroom | Train | Horses | Kangroo |
---|---|---|---|---|---|
20% | 29.12 | 31.56 | 24.59 | 29.99 | 29.59 |
30% | 32.02 | 34.63 | 27.00 | 32.52 | 32.21 |
50% | 36.45 | 38.88 | 32.00 | 37.27 | 37.34 |
80% | 41.51 | 42.56 | 40.73 | 41.97 | 42.74 |
TABLE V.
σ | 10% | 20% | 30% | 50% | 100% |
---|---|---|---|---|---|
0 | 23.47 | 26.87 | 29.83 | 35.60 | 42.94 |
5 | 23.34 | 26.73 | 29.27 | 33.61 | 37.70 |
10 | 23.16 | 26.07 | 28.17 | 31.17 | 34.31 |
15 | 22.66 | 25.17 | 26.82 | 29.31 | 32.14 |
20 | 22.17 | 24.27 | 25.62 | 27.90 | 30.55 |
25 | 21.68 | 23.49 | 24.72 | 26.79 | 29.30 |
For all of the examples previously considered, for both gray-scale and RGB images, we also attempted a direct application of matrix completion based on the incomplete matrix X ∈ ℝP×N, with columns defined by the image patches (i.e., for N patches, with P pixels in each, the incomplete matrix is of size P×N, with the ith column defined by the pixels in xi). We considered the algorithm in [20], using software from Prof. Candès’ web-site. For most of the examples previously considered, even after very careful tuning of the parameters, the algorithm diverged, suggesting that the low-rank assumptions were violated. For examples for which the algorithm did work, the PSNR values were typically 4 to 5 dB worse than those reported here for our model.
E. Interpolation of Hyperspectral Imagery
The basic BPFA, DP-BPFA, and PSBP-BPFA technology may be also applied to hyperspectral imagery, and it is here where these methods may have significant practical utility. Specifically, the amount of data that need be measured and read off a hyperspectral camera is often enormous. By selecting a small fraction of voxels for measurement and readout, uniformly selected at random, the quantity of data that need be handled is substantially reduced. Furthermore, one may simply modify existing hyperspectral cameras. We consider hyperspectral data with 106 spectral bands, from HyMAP scene A.P. Hill, VA, with permission from U.S. Army Topographic Engineering Center. Because of the significant statistical correlation across the multiple spectral bands, the fraction of data that need be read is further reduced, relative to gray-scale or RGB imagery. In this example, we considered 2% of the voxels, uniformly selected at random, and used image patches of size 4×4×106. Other than the increased data dimensionality, nothing in the model was changed.
In Fig. 6, we show example (mean) inferred images, at two (arbitrarily selected) spectral bands, as computed via BPFA. All 106 spectral bands are simultaneously analyzed. The average PSNR for the data cube (size 845×512×106) is 30.96 dB. While the PSNR value is of interest, for data of this type, the more important question concerns the ability to classify different materials based upon the hyperspectral data. In a separate forthcoming paper, we consider classification based on the full datacube, and based upon the BPFA-inferred datacube using 2% of the voxels, with encouraging results reported. We also tried the low-rank matrix completion algorithm from [20] for the hyperspectral data, and even after extensive parameter tuning, the algorithm diverged for all hyperspectral data considered.
In Table VI, we summarize algorithm performance on another hyperspectral data set, composed of 210 spectral bands. We show the PSNR values as a function of percentage of observed data and as a function of the size of the image patch. Note that the 1×1 patches only exploit spectral information, whereas the other patch sizes exploit both spatial and spectral information.
TABLE VI.
Observed data (%) | 1×1 | 2×2 | 3×3 | 4×4 |
---|---|---|---|---|
2 | 15.34 | 21.09 | 22.72 | 23.46 |
5 | 17.98 | 23.58 | 25.30 | 25.88 |
10 | 20.41 | 25.27 | 26.36 | 26.68 |
20 | 22.22 | 26.50 | 27.02 | 27.16 |
F. CS
We consider a CS example in which the image is divided into 8×8 patches, with these constituting the underlying data {xi}i=1,N to be inferred. For each of the N blocks, a vector of CS measurements yi= Σxi is measured, where the number of projections per patch is n, and the total number of CS projections is n · N. In our first examples, the elements of Σ are randomly constructed, as draws from (0,1); many other random projection classes may be considered [3] (and below, we also consider optimized projections Σ, matched to dictionary D). Each xi is assumed represented in terms of dictionary xi = Dwi + εi, and three constructions for D were considered: 1) a DCT expansion; 2) learning of D using BPFA, using training images; and 3) using the BPFA to perform joint CS inversion and learning of D. For 2, the training data consisted of 4000 8×8 patches chosen at random from 100 images selected from the Microsoft database (http://research.microsoft.com/enus/projects/objectclassrecognition). The dictionary was set to K = 256, and the offline BP inferred a dictionary of size M = 237.
Representative CS reconstruction results are shown in Fig. 7 (left) based upon a DCT dictionary, for a gray-scale version of the “castle” image. The results in Fig. 7 (right) are based on a learned dictionary; except for the “online BP” results (where D and {wi}i=1,N are jointly learned), all of these results employ the same dictionary D learned offline, as previously mentioned, and the algorithms are distinguished by different ways of estimating {wi}i=1,N. A range of CS-inversion algorithms are considered from the literature, and several BPFA-based constructions are considered as well for CS inversion. The online BPFA results (with no training data) are quite competitive with those based on a dictionary learned offline.
Note that results based on a learned dictionary are markedly better than those based on the DCT; similar results were achieved when the DCT was replaced by a wavelet representation. For the DCT-based results, note that the DP-BPFA and PSBP-BPFA CS inversion results are significantly better than those of all other CS inversion algorithms. The results reported here are consistent with tests we performed using over 100 images from the aforementioned Microsoft database, not reported here in detail for brevity.
In all previous results, projection matrix Σ was randomly constituted. We now consider a simple means of matching Σ to a D learned offline, based upon representative training images. Assume a learned D ∈ ℝP×K, with K > P, which may be represented via SVD as D = UΛVT; U ∈ ℝP×P and V ∈ ℝK×P are each composed of orthonormal columns, and Λ is a P×P diagonal matrix. The columns of U span the linear subspace of ℝP in which the columns of D reside. Furthermore, since the columns of D are generally not orthonormal, each column of D is “spread out” when expanded in the columns of U. Therefore, one expects that U and D are incoherent. Hence, a simple means of matching CS projections to the data is to define the rows of Σ in terms of randomly selected columns of U. This was done in Fig. 8 for the gray-scale “castle” image, using the same learned dictionary as considered in Fig. 7. It is observed that this procedure yields a marked improvement in CS recovery accuracy, for all CS inversion algorithms considered.
Concerning computational costs, all CS inversions were efficiently run on personal computers, with the specifics computational times dictated by the detailed MATLAB implementation and the machine run on. A rough ranking of the computational speeds, from fastest to slowest, is as follows: StOMP-CFAR, Fast BCS, OMP, BPFA, LARS/Lasso, Online BPFA, DP-BPFA, PSBP-BPFA, VB BCS, Basis Pursuit; in this list, algorithms BPFA through Basis Pursuits have approximately the same computational costs.
The improved performance of the CS inversion based upon the learned dictionaries is manifested as a consequence of the structure that is imposed on the underlying image while performing inversion. The early CS inversion algorithms, of the type considered in the above comparisons, imposed that the underlying image is sparsely represented in an appropriate basis (we showed results here based upon a DCT expansion of the 8×8 blocks over which CS inversion was performed, and similar results were manifested using a wavelet expansion). The imposition of such sparseness does not take into account the additional structure between the wavelet coefficients associated with natural images. Recently, researchers have utilized such structure to move beyond sparseness and achieve even better CS-inversion quality [2], [19]; in this paper, structural relationships and correlations between basis-function coefficients are accounted for. Additionally, there has been recent statistical research that has moved beyond sparsity and that are of interest in the context of CS inversion [38]. In the tests, we omit, for brevity, the algorithms in [2] and [19] yield performance that is comparable to the best results to the right in Fig. 7. Consequently, the imposition of structure in the form of learned dictionaries can be achieved using conventional basis expansions but with more sophisticated inversion techniques, that account for structure in imagery. Therefore, the main advantage of the learned dictionary in the context of CS is that it provides a very convenient means of defining projection matrices that are matched to natural imagery, as done in Fig. 8. Once those projection matrices are specified, the CS inversion may be employed based upon dictionaries as discussed here or based upon more traditional expansions (DCT or wavelets) and newer CS inversion methods [2], [19].
VI. Conclusion
The truncated beta-Bernoulli process has been employed to learn dictionaries matched to image patches {xi}i=1,N. The basic nonparametric Bayesian model is termed a BPFA framework, and extensions have been also considered. Specifically, the DP has been employed to cluster the {xi}i=1,N, encouraging similar dictionary-element usage within respective clusters. Furthermore, the PSBP has been used to impose that proximate patches are more likely to be similarly clustered (imposing that they are more probable to employ similar dictionary elements). All inference has been performed by a Gibbs sampler, with analytic update equations. The PBFA, DP-BPFA, and PSBP-BPFA have been applied to three problems in image processing: 1) denoising; 2) image interpolation based upon a subset of pixels selected uniformly at random; and 3) learning dictionaries for CS and also CS inversion. We have also considered jointly performing 1 and 2. Important advantages of the proposed methods are as follows. 1) A full posterior on model parameters are inferred, and therefore, “error bars” may be placed on the inverted images. 2) The noise variance need not be known; it is inferred within the analysis and may be non-stationary, and it may be inferred in the presence of significant missing pixels. 3) While training data may be used to initialize the dictionary learning, this is not needed, and the BPFA results are highly competitive even based upon random initializations. In the context of CS, the DP-BPFA and PSBP-BPFA results are state of the art, which is significantly better than existing published methods. Finally, based upon the learned dictionary, a simple method has been constituted for optimizing the CS projections.
The interpolation problem is related to CS, in that we exploit the fact that {xi}i=1,N reside on a low-dimensional subspace of ℝP, such that the total number of measurements is small relative to N · P (recall xi ∈ ℝP). However, in CS, one employs projection measurements Σxi, where Σ ∈ ℝn×P, ideally with n ≪ P. The interpolation problem corresponds to the special case in which the rows of Σ are randomly selected rows of the P×P identity matrix. This problem is closely related to the problem of matrix completion [6], [23], [33], where the incomplete matrix X ∈ ℝP×N has columns defined by {xi}i=1,N.
While the PSBP-BPFA successfully segmented the image while performing denoising and interpolation of missing pixels, we found that the PSNR performance of direct BPFA analysis performed very close to that of PSBP-BPFA in those applications. The use of PSBP-BPFA utilizes the spatial location of the image patches employed in the analysis, and therefore, it removes the exchangeability assumption associated with the simple BPFA (the location of the patches may be interchanged within the BPFA, without affecting the inference). However, since in the denoising and interpolation problems we have many overlapping patches, the extra information provided by PSBP-BPFA does not appear to be significant. By contrast, in the CS inversion problem, we do not have overlapping patches, and PSBP-BPFA provided significant performance gains relative to BPFA alone.
Acknowledgments
This work was supported in part by DOE, by NSF, by ONR, by NGA, and by ARO. The associate editor coordinating the review of this manuscript and approving it for publication was Prof. Maya Gupta.
Biographies
Mingyuan Zhou received the B.Sc. degree in acoustics from Nanjing University, Nanjing, China, in 2005 and the M.Eng. degree in signal and information processing from the Chinese Academy of Sciences, Beijing, China, in 2008. He is currently working toward the Ph.D. degree with the Department of Electrical and Computer Engineering, Duke University, Durham, NC.
His current research interests include statistical machine learning and signal processing, with emphasis on dictionary learning, sparse coding, and image and video processing.
Haojun Chen received the M.S. and Ph.D. degrees in electrical and computer engineering from Duke University, Durham, NC, in 2009 and 2011, respectively.
His research interests include machine learning, data mining, signal processing, and image processing.
John Paisley received the B.S.E., M.S., and Ph.D. degrees from Duke University, Durham, NC, in 2004, 2007, and 2010.
He is currently a Postdoctoral Researcher with the Department of Computer Science, Princeton University, Princeton, NJ. His research interests include Bayesian nonparametrics and machine learning.
Lu Ren received the B.S. degree in electrical engineering from Xidian University, Xi’an, China, in 2002 and the Ph.D. degree in electrical and computer engineering from Duke University, Durham, NC, in 2010.
She is currently working with Yahoo! Lab as a Scientist. Her research interests include machine learning, data mining, and statistical modeling.
Lingbo Li received the B.Sc. degree in electronic information engineering with Xidian University, Xi’an, China, in 2008 and the M.Sc. degree in sensing and signals with Duke University, Durham, NC, in 2010. She is currently working toward the Ph.D. degree with the Department of Electrical and Computer Engineering, Duke University.
Her current research interests include statistical machine learning and signal processing, with emphasis on dictionary learning and topic modeling.
Zhengming Xing received the B.S. degree in telecommunications from the Civil Aviation University of China, Tianjin, China, in 2008 and the M.S. degree in electrical engineering from Duke University, Durham, NC, in 2010. He is currently working toward the Ph.D. degree in electrical and computer engineering with Duke University.
His research interests include statistical machine learning, hyperspectral image analysis, compressive sensing, and dictionary learning.
David Dunson is currently a Professor of statistical science with Duke University, Durham, NC. His recent projects have developed sparse latent factor models that scale to massive dimensions and improve performance in predicting disease and other phenotypes based on high-dimensional and longitudinal biomarkers. Related methods can be used for combining high-dimensional data from different sources and for massively dimensional variable selection. His research interests include the development and application of novel Bayesian statistical methods motivated by high-dimensional and complex data sets, with particular emphasis on nonparametric Bayesian methods that avoid assumptions, such as normality and linearity, and on latent factor models that allow dimensionality reduction in massively dimensional settings.
Dr. Dunson is a Fellow of the American Statistical Association and the Institute of Mathematical Statistics. He was a recipient of the 2007 Mortimer Spiegelman Award for the top public health statistician, the 2010 Myrto Lefkopoulou Distinguished Lectureship at Harvard University, and the 2010 COPSS Presidents’ Award for the top statistician under 41.
Guillermo Sapiro was born in Montevideo, Uruguay, on April 3, 1966. He received his B.Sc. (summa cum laude), M.Sc., and Ph.D. degrees from Technion Israel Institute of Technology, Hafia, Israel, in 1989, 1991, and 1993, respectively.
After his postdoctoral research with Massachusetts Institute of Technology, he became a member of the Technical Staff with the research facilities of HP Labs, Palo Alto, CA. He is currently with the Department of Electrical and Computer Engineering, University of Minnesota, where he holds the position of a Distinguished McKnight University Professor and the Vincentine Hermes-Luh Chair in electrical and computer engineering.
Lawrence Carin (F’01–SM’96) was born on March 25, 1963 in Washington, DC. He received the B.S., M.S., and Ph.D. degrees in electrical engineering from the University of Maryland, College Park, in 1985, 1986, and 1989, respectively.
In 1989, he joined the Department of Electrical Engineering, Polytechnic University, Brooklyn, Brooklyn, NY, as an Assistant Professor and, in 1994, became an Associate Professor. In September 1995, he joined the Department of Electrical Engineering, Duke University, Durham, NC. He is now the William H. Younger Professor of engineering with Duke University. He is a Cofounder with Signal Innovations Group, Inc., which is a small business where he serves as the Director of Technology. He is the author of over 200 peer-reviewed papers. His current research interests include signal processing, sensing, and machine learning.
Dr. Carin is a member of the Tau Beta Pi and Eta Kappa Nu Honor Societies.
Appendix. Gibbs Sampling Inference
The Gibbs sampling update equations are given below; we provide the update equations for the BPFA, and the DP and PSBP versions are relatively simple extensions. Below, Σi represents the projection matrix on the data, for image patch xi. For the CS problem, Σi is typically fully populated, whereas for the interpolation problem, each row of Σi is all zeros, except for a single one corresponding to the specific pixel that is measured. The update equations are the conditional probability of each parameter, conditioned on all other parameters in the model.
Sample dk:
It can be shown that dk can be drawn from a normal distribution as
with the covariance Σdk and mean μdk expressed as
where
Sample zk: = [z1k, z2k, ··· zNk]:
The posterior probability that zik = 1 is proportional to
and the posterior probability that zik = 0 is proportional to
Thus, zik can be drawn from a Bernoulli distribution as
(11) |
Sample sk: = [s1k, s2k, ··· sNk]:
It can be shown that sik can be drawn from a normal distribution, i.e.,
(12) |
variance Σsik and mean μsik expressed as
Note that zik is equal to either 1 or 0, and Σsik and μsik can be further expressed as
Sample πk:
It can be shown that πk can be drawn from a Beta distribution as
Sample γs:
It can be shown that γs can be drawn from a Gamma distribution as
Sample γε:
(13) |
It can be shown that γε can be drawn from a Gamma distribution as
(14) |
Note that is a sparse identity matrix, Σdk is a diagonal matrix, and Z is a sparse matrix; it is easy to find that only basic arithmetical operations are needed, and many unnecessary calculations can be avoided, leading to fast computation and low memory requirement.
Footnotes
Proximity can be modeled as in “spatial proximity,” as here developed in detail, or “feature proximity” as in nonlocal means and related approaches; see [26] and references therein.
Color versions of one or more of the figures in this paper are available online at http://ieeexplore.ieee.org.
Contributor Information
Mingyuan Zhou, Department of Electrical and Computer Engineering, Duke University, Durham, NC 27708-0291 USA.
Haojun Chen, Department of Electrical and Computer Engineering, Duke University, Durham, NC 27708-0291 USA.
John Paisley, Department of Electrical and Computer Engineering, Duke University, Durham, NC 27708-0291 USA.
Lu Ren, Department of Electrical and Computer Engineering, Duke University, Durham, NC 27708-0291 USA.
Lingbo Li, Department of Electrical and Computer Engineering, Duke University, Durham, NC 27708-0291 USA.
Zhengming Xing, Department of Electrical and Computer Engineering, Duke University, Durham, NC 27708-0291 USA.
David Dunson, Department of Statistics, Duke University, Durham, NC 27708-0291 USA.
Guillermo Sapiro, Department of Electrical and Computer Engineering, University of Minnesota, Minneapolis, MN 55455 USA.
Lawrence Carin, Department of Electrical and Computer Engineering, Duke University, Durham, NC 27708-0291 USA.
References
- 1.Aharon M, Elad M, Bruckstein AM. K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans Signal Process. 2006 Nov;54(11):4311–4322. [Google Scholar]
- 2.Baraniuk RG, Cevher V, Duarte MF, Hegde C. Model-based compressive sensing. IEEE Trans Inf Theory. 2010 Apr;56(4):1982–2001. [Google Scholar]
- 3.Baraniuk RG. Compressive sensing. IEEE Signal Process Mag. 2007 Jul;24(4):118–121. [Google Scholar]
- 4.Bruckstein AM, Donoho DL, Elad M. From sparse solutions of systems of equations to sparse modeling of signals and images. SIAM Rev. 2009 Jan;51(1):34–81. [Google Scholar]
- 5.Candès E, Tao T. Near-optimal signal recovery from random projections: Universal encoding strategies? IEEE Trans Inf Theory. 2006 Dec;52(12):5406–5425. [Google Scholar]
- 6.Candès EJ, Tao T. The power of convex relaxation: Near-optimal matrix completion. IEEE Trans Inf Theory. 2010 May;56(5):2053–2080. [Google Scholar]
- 7.Chung Y, Dunson DB. Nonparametric Bayes conditional distribution modeling with variable selection. J Amer Stat Assoc. 2009 Dec;104(488):1646–1660. doi: 10.1198/jasa.2009.tm08302. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Dabov K, Foi A, Katkovnik V, Egiazarian K. Image denoising by sparse 3D transform-domain collaborative filtering. IEEE Trans Image Process. 2007 Aug;16(8):2080–2095. doi: 10.1109/tip.2007.901238. [DOI] [PubMed] [Google Scholar]
- 9.Donoho DL. Compressed sensing. IEEE Trans Inf Theory. 2006 Apr;52(4):1289–1306. [Google Scholar]
- 10.Donoho DL, Johnstone IM, Kerkyacharian G, Picard D. Wavelet shrinkage: Asymptopia. J Roy Stat Soc B. 1995;57(2):301–369. [Google Scholar]
- 11.Duarte MF, Davenport MA, Takhar D, Laska JN, Sun T, Kelly KF, Baraniuk RG. Single-pixel imaging via compressive sampling. IEEE Signal Process Mag. 2008 Mar;25(2):83–91. [Google Scholar]
- 12.Duarte-Carvajalino JM, Sapiro G. Learning to sense sparse signals: Simultaneous sensing matrix and sparsifying dictionary optimization. IEEE Trans Image Process. 2009 Jul;18(7):1395–1408. doi: 10.1109/TIP.2009.2022459. [DOI] [PubMed] [Google Scholar]
- 13.Elad M, Aharon M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans Image Process. 2006 Dec;15(2):3736–3745. doi: 10.1109/tip.2006.881969. [DOI] [PubMed] [Google Scholar]
- 14.Elad M, Yavneh I. A Weighted Average of Sparse Representations is Better Than the Sparsest One Alone. 2010. Preprint. [Google Scholar]
- 15.Ferguson T. A Bayesian analysis of some nonparametric problems. Ann Stat. 1973 Mar;1(2):209–230. [Google Scholar]
- 16.Geman S, Geman D. Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images. IEEE Trans Pattern Anal Mach Intell. 1984 Nov;PAMI-6(6):721–741. doi: 10.1109/tpami.1984.4767596. [DOI] [PubMed] [Google Scholar]
- 17.Gleichman S, Eldar YC. Blind Compressed Sensing Technion—Israel Inst. Technol., Haifa, Israel. CCIT Rep. 2010;759 Preprint (on Arxiv.org) [Google Scholar]
- 18.Griffiths TL, Ghahramani Z. Infinite latent feature models and the Indian buffet process. Proc Adv Neural Inf Process Syst. 2005:475–482. [Google Scholar]
- 19.He L, Carin L. Exploiting structure in wavelet-based Bayesian compressive sensing. IEEE Trans Signal Process. 2009 Sep;57(9):3488–3497. [Google Scholar]
- 20.Candès EJ, Cai JF, Shen Z. A singular value thresholding algorithm for matrix completion. SIAM J Optim. 2008 Jan;20(4):1956–1982. [Google Scholar]
- 21.Ji S, Xue Y, Carin L. Bayesian compressive sensing. IEEE Trans Signal Process. 2008 Jun;56(6):2346–2356. [Google Scholar]
- 22.Knowles D, Ghahramani Z. Infinite sparse factor analysis and infinite independent components analysis. Proc Int Conf Ind Compon Anal Signal Separation. 2007:381–388. [Google Scholar]
- 23.Lawrence ND, Urtasun R. Non-linear matrix factorization with Gaussian processes. Proc Int Conf Mach Learn. 2009:601–608. [Google Scholar]
- 24.Mairal J, Bach F, Ponce J, Sapiro G. Online dictionary learning for sparse coding. Proc Int Conf Mach Learn. 2009:689–696. [Google Scholar]
- 25.Mairal J, Bach F, Ponce J, Sapiro G, Zisserman A. Supervised dictionary learning. Proc Neural Inf Process Syst. 2008:1033–1040. [Google Scholar]
- 26.Mairal J, Bach F, Ponce J, Sapiro G, Zisserman A. Non-local sparse models for image restoration. Proc Int Conf Comput Vis. 2009:2272–2279. [Google Scholar]
- 27.Mairal J, Elad M, Sapiro G. Sparse representation for color image restoration. IEEE Trans Image Process. 2008 Jan;17(1):53–69. doi: 10.1109/tip.2007.911828. [DOI] [PubMed] [Google Scholar]
- 28.Mairal J, Sapiro G, Elad M. Learning multiscale sparse representations for image and video restoration. SIAM Multisc Model Simul. 2008;7(1):214–241. [Google Scholar]
- 29.Paisley J, Carin L. Nonparametric factor analysis with beta process priors. Proc Int Conf Mach Learn. 2009:777–784. [Google Scholar]
- 30.Raina R, Battle A, Lee H, Packer B, Ng AY. Self-taught learning: Transfer learning from unlabeled data. Proc Int Conf Mach Learn. 2007:759–766. [Google Scholar]
- 31.Ranzato M, Poultney C, Chopra S, Lecun Y. Efficient learning of sparse representations with an energy-based model. Proc Neural Inf Process Syst. 2006:1137–1144. [Google Scholar]
- 32.Ren L, Du L, Dunson D, Carin L. The logistic stick breaking process. J Mach Learn Res. preprint. [PMC free article] [PubMed] [Google Scholar]
- 33.Salakhutdinov R, Mnih A. Bayesian probabilistic matrix factorization using Markov chain Monte Carlo. Proc Int Conf Mach Learn. 2008:880–887. [Google Scholar]
- 34.Sethuraman J. A constructive definition of Dirichlet priors. Stat Sin. 1994;4(2):639–650. [Google Scholar]
- 35.Shankar M, Pitsianis NP, Brady DJ. Compressive video sensors using multichannel imagers. Appl Opt. 2010 Feb;49(10):B9–B17. doi: 10.1364/AO.49.0000B9. [DOI] [PubMed] [Google Scholar]
- 36.Teh YW, Jordan MI, Beal MJ, Blei DM. Hierarchical Dirichlet processes. J Amer Stat Assoc. 2006 Dec;101(476):1566–1581. [Google Scholar]
- 37.Thibaux R, Jordan MI. Hierarchical beta processes and the Indian buffet process. Proc Int Conf Artif Intell Stat. 2007:564–571. [Google Scholar]
- 38.Tibshirani R, Saunders M, Rosset S, Zhu J, Knight K. Sparsity and smoothness via the fused Lasso. J Roy Stat Soc B. 2005 Feb;67(1):91–108. [Google Scholar]
- 39.Tipping M. Sparse Bayesian learning and the relevance vector machine. J Mach Learn Res. 2001 Jun;1:211–244. [Google Scholar]
- 40.Wright J, Yang AY, Ganesh A, Sastry SS, Ma Y. Robust face recognition via sparse representation. IEEE Trans Pattern Anal Mach Intell. 2009 Feb;31(2):210–227. doi: 10.1109/TPAMI.2008.79. [DOI] [PubMed] [Google Scholar]
- 41.Yang J, Wright J, Huang T, Ma Y. Image super-resolution via sparse representation. IEEE Trans Image Process. 2009 Nov;19(11):2861–2873. doi: 10.1109/TIP.2010.2050625. [DOI] [PubMed] [Google Scholar]
- 42.Zhou M, Chen H, Paisley J, Ren L, Sapiro G, Carin L. Non-parametric Bayesian dictionary learning for sparse image representations. Proc Neural Inf Process Syst. 2009:1–9. [Google Scholar]
- 43.Zhou M, Yang H, Sapiro G, Dunson D, Carin L. Dependent hierarchical beta process for image interpolation and denoising. Proc Int Conf Artificial Intelligence Statistics (AISTATS) JMLR W&CP. 2011;15:883–891. [Google Scholar]