Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2020 Aug 1.
Published in final edited form as: IEEE Signal Process Lett. 2019 Jun 7;26(8):1137–1141. doi: 10.1109/lsp.2019.2921446

Convolutional Analysis Operator Learning: Dependence on Training Data

Il Yong Chun 1,, David Hong 1,, Ben Adcock 2, Jeffrey A Fessler 3
PMCID: PMC7170269  NIHMSID: NIHMS1533018  PMID: 32313415

Abstract

Convolutional analysis operator learning (CAOL) enables the unsupervised training of (hierarchical) convolutional sparsifying operators or autoencoders from large datasets. One can use many training images for CAOL, but a precise understanding of the impact of doing so has remained an open question. This paper presents a series of results that lend insight into the impact of dataset size on the filter update in CAOL. The first result is a general deterministic bound on errors in the estimated filters, and is followed by a bound on the expected errors as the number of training samples increases. The second result provides a high probability analogue. The bounds depend on properties of the training data, and we investigate their empirical values with real data. Taken together, these results provide evidence for the potential benefit of using more training data in CAOL.

I. Introduction

LEARNING convolutional operators from large datasets is a growing trend in signal/image processing, computer vision, machine learning, and artificial intelligence. The convolutional approach resolves the large memory demands of patch-based operator learning and enables unsupervised operator learning from “big data,” i.e., many high-dimensional signals. See [1], [2] and references therein. Examples include convolutional dictionary learning [2], [3] and convolutional analysis operator learning (CAOL) [1], [4]. CAOL trains an autoencoding CNN in an unsupervised manner, and is useful for training multi-layer CNNs from many training images [1]. In particular, the block proximal gradient method using a majorizer [1], [2] leads to rapidly converging and memory-efficient CAOL [1]. However, a theoretical understanding of the impact of using many training images in CAOL has remained an open question.

This paper presents new insights on this topic. Our first main result provides a deterministic bound on filter estimation error, and is followed by a bound on the expected error when “model mismatch” has zero mean. (See Theorem 1 and Corollary 2, respectively.) The expected error bound depends on the training data, and we provide empirical evidence of its decrease with an increase in training samples. Our second main result provides a high probability bound that explicitly decreases with increasingly many i.i.d. training samples. The bound improves when model mismatch and samples are uncorrelated. (See Theorem 3.) Additional empirical findings provide evidence that the correlation can indeed be small in practice. Put together, our findings provide new insight into how using many samples can improve CAOL, underscoring the benefits of the low memory usage of CAOL.

II. Backgrounds and Preliminaries

A. CAOL with orthogonality constraints

CAOL seeks a set of filters that “best” sparsify a set of training images {xlN:l=1,,L} by solving the optimization problem [1, §II-A] (see Appendix for notation):

argminD=[d1,,dK]min{zl,k}F(D,{zl,k}), subj.toDDH=1RI, (P0)
F(D,{zl,k}):=l=1Lk=1Kdkxlzl,k22+αzl,k0,

where ⊛ denotes convolution, {dkR:k=1,,K} is a set of KR convolutional kernels, {zl,kN:l=1,,L,k=1,,K} is a set of sparse codes, α > 0 is a regularization parameter controlling the sparsity of features {zl,k}, and ||·||0 denotes the 0-quasi-norm. We group the K filters into a matrix:

D:=[d1dK]R×K. (1)

The orthogonality condition DDH=1RI in (P0) enforces 1) a tight-frame condition on the filters, i.e., k=1Kdkx22=x22, ∀x [1, Prop. 2.1]; and 2) filter diversity when R = K, since DDH=1RI implies DHD=1KI and each pair of filters is incoherent, i.e., |〈dk, dk〉|2 = 0, ∀kk′. One often solves (P0) iteratively, by alternating between optimizing D (filter update) and optimizing {zl,k : ∀l, k} (sparse code update) [1], i.e., at the i iteration, the current iterates are updated as {zl,k(i+1)}=argmin{zl,k}F(D(i),{zl,k}) and D(i+1)=argminDDH=1RIF(D,{zl,k(i+1)}).

B. Filter update in a matrix form

The key to our analysis lies in rewriting the filter update for (P0) in matrix form, to which we apply matrix perturbation and concentration inequalities. Observe first that

dkxl=[Π0xl,,ΠR1xl]=:ΨlN×Rdk=Ψldk,l=1,,L, (2)

where Π:=[0IN110]N×N is the circular shift operator and (·)n denotes the matrix product of its n copies. We consider a circular boundary condition to simplify the presentation of {Ψl} in (2), but our entire analysis holds for a general boundary condition with only minor modifications of {Ψl} as done in [1, §IV-A]. Using (2), the filter update of (P0) is rewritten as

D=argminDl=1LΨlDZlF2,subj.to DDH=1RI, (P1)

where Zl:=[zl,1,,zl,K]N×K contains all the current sparse code estimates for the lth sample, and we drop iteration superscript indices (·)(i) throughout. The next section uses this form to characterize the filter update solution D.

III. Main Results:

Dependence of CAOL on Training Data

The main results in this section illustrate how training with many samples can reduce errors in the filter D from (P1) and characterize the reduction in terms of properties of the training data. Throughout we model the current sparse codes estimates as

Zl=ΨlDtrue =:Ztrue ,l+El,l=1,,L (3)

where Dtrue is formed from optimal (orthogonal) filters analogously to (1), and ElN×K captures model mismatch in the current sparse codes, e.g., due to the current iterate being far from convergence or being trapped in local minima.

The following theorem provides a deterministic characterization.

Theorem 1.

Suppose that both matrices

l=1LΨlHZlR×Kandl=1LΨlHZtrue ,lR×K (4)

are full row rank, where {Ψl, Zl, Ztrue,l : l = 1, …, L} are defined in (2)–(3). Then, the solution D to (P1) has error with respect to Dtrue bounded as

DDtrue F25l=1LΨlHElF2λmin2(l=1LΨlHΨl), (5)

where λmin(·) denotes the smallest eigenvalue of its argument.

The full row rank condition on (4) ensures that the estimated filters D and the true filters Dtrue are unique, and it further guarantees that the denominator of (5) is strictly positive. When the model mismatches E1, …, EL are independent and mean zero, we obtain the following expected error bound:

Corollary 2.

Under the construction of Theorem 1, suppose that El is a zero-mean random matrix for l = 1, …, L, and is independent over l. Then,

EDDtrue F25σ¯2ρ2, (6)

where E() denotes the expectation,

σ¯2:=maxl=1,,Lλmax(E{ElElH}),ρ2:=tr(l=1LΨlHΨl)λmin2(l=1LΨlHΨl), (7)

λmax(·) denotes the largest eigenvalue of its argument, and the expectation is taken over the model mismatch.

Given fixed K and R, it is natural to expect that σ¯2 is bounded by some constant independent of L, and so the expected error bound in (6) largely depends on ρ2 in (7). When training samples are i.i.d., one may further expect (1/L)l=1LΨlHΨl to concentrate around its expectation, roughly resulting in ρ2 ∝ 1/L, with a proportionality constant that depends on R and the statistics of the training data. Fig. 1 illustrates ρ2 for various image datasets, providing empirical evidence of this decrease in real data.

Fig. 1.

Fig. 1.

Empirical values of ρ2 in (7) show a decrease with L for different datasets and filter dimensions. (The fruit and city datasets with L =10 and N = 104 were preprocessed with contrast enhancement and mean subtraction; see details of datasets and experiments in [1], [2] and references therein. For L < 10, the results are averaged over 50 datasets randomly selected from the full datasets.) Under the assumptions of Corollary 2, the decrease in this quantity leads to a better expected error bound in (5). Without preprocessing, the quantity ρ2 increases by a factor of around 103.

Our second theorem provides a probabilistic error bound via concentration inequalities, given i.i.d. training sample and model mismatch pairs (x1, E1), …, (xL, EL).1 It removes the zero-mean assumption for the model mismatches {El : ∀l} in Corollary 2 that might be strong, e.g., if training data are not preprocessed to have zero mean.

Theorem 3.

Suppose that training sample and model mismatch pairs (x1,E1),,(xL,EL)~iid(x,E), where x and E are almost surely bounded, i.e.,

x2γandEFσ, (8)

and the matrices in (4) are almost surely full row rank. Then, for any 0<δ<λmin(Λ¯)/(2Rγ2), the solution D to (P1) has error with respect to Dtrue bounded as

DDtrue F25{σtr(Λ¯)/L+E(ΨHE)F+2σγRδλmin(Λ¯)2γ2Rδ}2, (9)

with probability at least

13Rexp(Lδ2/23+δ/3), (10)

where Λ¯:=E(ΨHΨ) and Ψ is constructed from x as in (2).

Taking δ sufficiently small, the high probability error bound (9) is primarily driven by

ρ¯:=tr(Λ¯)/Lλmin(Λ¯)andχ¯:=E(ΨHE)Fλmin(Λ¯), (11)

where ρ¯ is analogous to ρ in (7), and χ¯ captures how correlated the model mismatch is to the training samples. As the number L of training samples increases, ρ¯ decreases as 1/L. On the other hand, χ¯ is constant with respect to L and provides a floor for the bound. Fig. 2 illustrates χ¯ for CAOL iterates from different image datasets, and provides empirical evidence that this term can indeed be small in real data. If the model mismatch is sufficiently uncorrelated with the training samples, i.e., χ¯ is practically zero, then only the ρ¯ term remains and this term decreases with L. Namely, if model mismatch is entirely uncorrelated with the training samples, then using many samples decreases the error bound to (effectively) zero.

Fig. 2.

Fig. 2.

Empirical estimate of χ¯ in (11) across iterations in the alternating optimization algorithm [1] that solves CAOL (P0) with α = 10−3. (The fruit and city datasets with L = 10 and N = 104 were preprocessed with contrast enhancement and mean subtraction; see details of datasets and experiments in [1], [2] and references therein. The model mismatches {El(i):l} at the ith iteration were calculated every 50 iterations based on (3), where we use the converged filters for Dtrue.) Observe that χ¯(i) generally decreases over iterations; when χ¯ is small, the high probability error bound (9) in Theorem 3 depends primarily on ρ¯ defined in (11).

IV. Related Works

Sample complexity [7] and synthesis (or reconstruction) error [8] have been studied in the context of synthesis operator learning (e.g., dictionary learning [9]); see the cited papers and references therein. A similar understanding for (C)AOL has however remained largely open; existing works focus primarily on establishing (C)AOL models and their algorithmic challenges [1], [10]–[13]. The authors in [14] studied sample complexity for a patch-based AOL method, but the form of their model differs from that of ours (P0). Specifically, they consider the following AOL problem: minDlf(DTx^l)+g(D), where f (·) is a sparsity promoting function (e.g., a smooth approximation of the 0-quasi-norm [14]), g(·) is a regularizer or constraint for the filter matrix D, and {x^l:l=1,,L} is a set of training patches (not images).

V. Proof of Theorem 1

Rewriting (P1) yields that D is a solution of the (scaled) orthogonal Procrustes problem [1, §S.VII]:

argminDΨ˜DZ˜F2, subj.toDDH=1RI, (12)

where Ψ˜LN×R arises by stacking Ψ1, …, ΨL vertically and Z˜LN×K arises likewise from Z1, …, ZL. Similarly, since Ψ1 Dtrue = Ztrue,l as in (3), Dtrue is a solution of the analogous (scaled) orthogonal Procrustes problem

argminDΨ˜DZ˜trueF2, subj.toDDH=1RI, (13)

where Z˜true LN×K arises by stacking Ztrue,1, …, Ztrue,L vertically.

By assumption, both Ψ˜HZ˜ and Ψ˜HZ˜true are full row rank and so (12) and (13) have unique solutions given by the unique (scaled) polar factors

D=1RQ(Z˜HΨ˜)HDtrue=1RQ(Z˜trueHΨ˜)H (14)

where Q(·) denotes the polar factor of its argument, and can be computed as Q(A) = WVH from the (thin) singular value decomposition A = WΣVH.

Thus we have

DDtrue F2=1RQ(Z˜trueHΨ˜)Q(Z˜HΨ˜)F21RE˜HΨ˜F2{[2σR(Z˜trueHΨ˜)+σR(Z˜HΨ˜)]2+[1max{σR(Z˜trueHΨ˜),σR(Z˜HΨ˜)}]2}1RE˜HΨ˜F2{[2σR(Z˜true HΨ˜)]2+[1σR(Z˜true HΨ˜)]2}=5RΨ˜HE˜F2σR2(Ψ˜HZ˜true)=5Rl=1LΨlHElF2σR2(l=1LΨlHZtrue,l) (15)

where E˜=Z˜Z˜true is exactly E1, …, EL stacked vertically, and σr (·) denotes the rth largest singular value of its argument. The first inequality holds by the perturbation bound in [15, Thm. 3], and the second holds since σR(Z˜HΨ˜)0. Recalling that Ztrue,l = ΨlDtrue, we rewrite the denominator of (15) as

σR2(l=1LΨlHZtrue ,l)=σR2(l=1LΨlHΨlDtrue )=1RσR2(l=1LΨlHΨl)=1Rλmin2(l=1LΨlHΨl), (16)

where the second equality holds because DtrneDtrueH=(1/R)I. Substituting (16) into (15) yields (5).

VI. Proof of Corollary 2

Taking the expectation of (5) over the model mismatch amounts to taking the expectation of the numerator of the upper bound in (5):

El=1LΨlHElF2=l=1LEΨlHElF2=l=1Ltr(ΨlHE{ElElH}Ψl)l=1Lλmax(E{ElElH})ΨlF2σ¯2l=1LΨlF2, (17)

where the first equality holds by using the assumption that El is zero-mean and independent over l, the second equality follows by expanding the Frobenius norm then applying linearity of the trace and expectation, the first inequality holds since vHMvλmax(M)v22 for any vector v and Hermitian matrix M, and the last inequality follows from the definition of σ¯2. Rewriting (17) using the identity l=1LΨlF2=l=1Ltr(ΨlHΨl)=tr(l=1LΨlHΨl) yields the result (6).

VII. Proof of Theorem 3

We derive two high probability bounds, one each for the numerator and denominator of (5). Then, the bound (9) with probability (10) follows by combining the two via a union bound. Before we begin, note that (8) implies that Ψ2ΨFγR almost surely; our proofs use this inequality multiple times.

A. Upper bound for numerator

Observe first that

l=1LΨlHElF=LE(ΨlHEl)+l=1L{ΨlHElE(ΨlHEl)}FLE(ΨlHEl)F+l=1Lξl2, (18)

where ξl:=vec{ΨlHElE(ΨlHEl)}RK for l = 1, …, L. We next bound l=1Lξl2 via the vector Bernstein inequality [16, Cor. 8.44]. Note that ξ1, …, ξL are i.i.d. with Eξl=0 (by construction). Furthermore, ξl is almost surely bounded as

ξl2=ΨlHElE(ΨlHEl)FΨlHElF+E(ΨlHEl)F(Triangleineq.)ΨlHElF+E(ΨlHEl)F(Jensens ineq.)ΨlFElF+EΨlFElF2σγR.

Thus the vector Bernstein inequality [16, Cor. 8.44] yields that for any t > 0,

l=1Lξl2σLtr(Λ¯)+t, (19)

with probability at least

1exp{t2/23L(2σγR)2+t(2σγR)/3}. (20)

We obtained (19) by the following simplification:

El=1Lξl2El=1Lξl22=LEξl22=LEΨlHElE(ΨlHEl)F2=L{EΨlHElF2E(ΨlHEl)F2}LEΨlHElF2LE(ΨlF2ElF2)Lσ2EΨlF2=σLtr(Λ¯),

where the third equality holds by EAEAF2=i,jE(Ai,jEAi,j)2=i,jEAi,j2(EAi,j)2=EAF2EAF2. We obtained (20) by the following simplifications:

supx21E|xHξl|2Eξl22(2σγR)2,
El=1Lξl2El=1Lξl2LEξl2L(2σγR).

Applying (19) and (20) with t=2σγRLδ to the square of (18) yields

l=1LΨlHElF2L2{σtr(Λ¯)/L+E(ΨlHEl)F+2σγRδ}2, (21)

with probability at least 1exp(Lδ2/23+δ/3).

B. Lower bound for denominator

Observe that l=1LΨlHΨl=LΛ¯+l=1LΛl, where Λl:=ΨlHΨlΛ¯, so Weyl’s inequality [17] yields

λmin(l=1LΨlHΨl)λmin(LΛ¯)l=1LΛl2, (22)

and it remains to bound l=1LΛl2. We do so by using the Matrix Bernstein inequality [16, Cor. 8.15].

Note that Λ1, …, ΛL are i.i.d. (since x1, …, xL are i.i.d.) and EΛl=0. Furthermore, Λl is almost surely bounded as

Λl2=ΨlHΨlE(ΨlHΨl)2ΨlHΨl2+E(ΨlHΨl)2(Triangleineq.)ΨlHΨl2+EΨlHΨl2(Jensens ineq.)=Ψl22+EΨl222γ2R.

Thus, the Matrix Bernstein inequality [16, Cor. 8.15] yields that for any t > 0,

{l=1LΛl2t}2Rexp{t2/2L(2γ2R)2+2γ2Rt/3}, (23)

where we use the following simplification:

l=1LEΛl22=LEΛl22LEΛl22L(2γ2R)2.

Applying (23) with t = 2γ2RLδ to the square of (22) yields

λmin2(l=1LΨlHΨl)L2{λmin(Λ¯)2γ2Rδ}2. (24)

with probability at least 12Rexp(Lδ2/21+δ/3).

C. Combined bound

Combining the bounds (21) and (24) via a union bound yields (9) with probability at least

1exp(Lδ2/23+δ/3)2Rexp(Lδ2/21+δ/3), (25)

which is greater than or equal to (10).

Acknowledgments

This work is supported in part by the Keck Foundation and NIH grant U01 EB018753. BA is supported by NSERC grant 611675.

Footnotes

1

We follow the natural convention in sample size analyses of assuming that {xl : ∀l} are i.i.d. samples from an underlying training distribution; see the references cited in Section IV and [5], [6] for other examples. Model mismatches {El : ∀l} also become i.i.d. across samples at all iterations of CAOL, if “fresh” training samples are used for each update, e.g., as can be done when solving (P1) via mini-batch stochastic optimization.

Contributor Information

Ben Adcock, Department of Mathematics, Simon Fraser University, Burnaby, BC V5A 1S6 Canada.

Jeffrey A. Fessler, Department of Electrical Engineering and Computer Science, The University of Michigan, Ann Arbor, MI 48019 USA.

REFERENCES

  • [1].Chun IY and Fessler JA, “Convolutional analysis operator learning: Acceleration and convergence,” submitted, January 2018. [Online]. Available: http://arxiv.org/abs/1802.05584 [DOI] [PMC free article] [PubMed]
  • [2].Chun IY and Fessler JA, “Convolutional dictionary learning: Acceleration and convergence,” IEEE Trans. Image Process, vol. 27, no. 4, pp. 1697–1712, April 2018. [DOI] [PubMed] [Google Scholar]
  • [3].Chun IY and Fessler JA, “Convergent convolutional dictionary learning using adaptive contrast enhancement (CDL-ACE): Application of CDL to image denoising,” in Proc. Sampling Theory and Appl. (SampTA), Tallinn, Estonia, July 2017, pp. 460–464. [Google Scholar]
  • [4].Chun IY and Fessler JA, “Convolutional analysis operator learning: Application to sparse-view CT,” in Proc. Asilomar Conf. on Signals, Syst., and Comput, Pacific Grove, CA, October 2018, pp. 1631–1635. [Google Scholar]
  • [5].Hastie T, Tibshirani R, and Friedman J, The elements of statistical learning: Data mining, inference, and prediction, ser. Springer series in statistics. New York, NY: Springer, 2009. [Google Scholar]
  • [6].Mohri M, Rostamizadeh A, and Talwalkar A, Foundations of machine learning. Cambridge, MA: MIT Press, 2018. [Google Scholar]
  • [7].Shakeri Z, Sarwate AD, and Bajwa WU, “Sample complexity bounds for dictionary learning from vector- and tensor-valued data,” in Information Theoretic Methods in Data Science, Rodrigues M and Eldar Y, Eds. Cambridge, UK: Cambridge University Press, 2019, ch. 5. [Google Scholar]
  • [8].Singh S, Poczos B, and Ma J, “Minimax reconstruction risk of convolutional sparse dictionary learning,” in Proc. Int. Conf. on Artif. Int. and Stat, ser. Proc. Mach. Learn. Res., vol. 84, Playa Blanca, Lanzarote, Canary Islands, April 2018, pp. 1327–1336. [Google Scholar]
  • [9].Aharon M, Elad M, and Bruckstein A, “K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation,” IEEE Trans. Signal Process, vol. 54, no. 11, pp. 4311–4322, November 2006. [Google Scholar]
  • [10].Yaghoobi M, Nam S, Gribonval R, and Davies ME, “Constrained overcomplete analysis operator learning for cosparse signal modelling,” IEEE Trans. Signal Process, vol. 61, no. 9, pp. 2341–2355, March 2013. [Google Scholar]
  • [11].Hawe S, Kleinsteuber M, and Diepold K, “Analysis operator learning and its application to image reconstruction,” IEEE Trans. Image Process, vol. 22, no. 6, pp. 2138–2150, June 2013. [DOI] [PubMed] [Google Scholar]
  • [12].Cai J-F, Ji H, Shen Z, and Ye G-B, “Data-driven tight frame construction and image denoising,” Appl. Comput. Harmon. Anal, vol. 37, no. 1, pp. 89–105, October 2014. [Google Scholar]
  • [13].Ravishankar S and Bresler Y, “ℓ0 sparsifying transform learning with efficient optimal updates and convergence guarantees,” IEEE Trans. Sig. Process, vol. 63, no. 9, pp. 2389–2404, May 2015. [Google Scholar]
  • [14].Seibert M, Wörmann J, Gribonval R, and Kleinsteuber M, “Learning co-sparse analysis operators with separable structures,” IEEE Trans. Signal Process, vol. 64, no. 1, pp. 120–130, January 2016. [Google Scholar]
  • [15].Li R-C, “New perturbation bounds for the unitary polar factor,” SIAM J. Matrix Anal. Appl, vol. 16, no. 1, pp. 327–332, January 1995. [Google Scholar]
  • [16].Foucart S and Rauhut H, A mathematical introduction to compressive sensing. New York, NY: Springer, 2013. [Google Scholar]
  • [17].Weyl H, “Das asymptotische verteilungsgesetz der eigenwerte linearer partieller differentialgleichungen (mit einer anwendung auf die theorie der hohlraumstrahlung),” Mathematische Annalen, vol. 71, no. 4, pp. 441–479, December 1912. [Google Scholar]

RESOURCES