In this discussion contribution, we connect the elegant proposal of Cai, Sun and Wang to multiview data, in which multiple sets of variables (or ‘views’) are measured on the same observations. Using ideas from Section 4, we show that we can exploit a secondary view to improve power for testing on the first view.
Consider independent and identically distributed observations of m random variables under two conditions. In condition observation of variable is given by (view 1 )
Where is zero mean, and we suppress the common intercept. The random-mean vectors μ(1) and μ(2) are sparse. Furthermore, for the same individuals, we also observe a second view of variables (view 2):
The mean vectors are sparse, is zero mean and again we suppress the intercept. Suppose that the two views satisfy a hierarchical sparsity constraint: for and
| (6) |
where maps the jth entry of μ(l) to its parent in Fig. 11.
Fig. 11.

Schematic diagram of constraint (6) with σ(3) = 1
Concretely, suppose that X(l) and Z(l) contain protein and gene expression measurements respectively. If transcripts that encode the jth protein are absent (i.e. ),then the jth protein cannot be present (i.e.
Suppose that is independent of . Further assume that the random errors are bivariate normal and independent across j, l and i, and independent of μ(l) and
Using the terminology of Cai, Sun and Wang the ‘primary statistic’ for testing is
for some constant Cj. We consider a pair of ‘auxiliary statistics’,
for some constants Dj and Ej. The statistic Rj is the same as T2j in the paper, whereas Sj is constructed by using the second data view. A small value of provides evidence for which by constraint (6) suggests that By analogy with proposition 1 in the paper, the oracle statistic is
Moreover, enjoys the properties in theorem 3 of the paper. Detailed proofs are available from https://hugogogo.github.io/paper/cars_discussion_supplement.pdf. If there is not a one-to-one mapping between σ(j) and j then must be estimated carefully.
Supplementary Material
Fig. 9.

(a) Power comparison and (b) empirical misclassification rates for two classes and based on 500 replications (FDR level α = 0.05; n1 = 50; n2 = 60; m = 1000; :
, method (5) based on Benjamini and Hochberg (1995);
, method (5) based on CARS;
, Bayes rule
Fig. 10.

Empirical misclassification rates when the same amount of locations are chosen for both methods:
, Benjamini and Hochberg (1995);
, CARS;
, Bayes rule
Contributor Information
Guo Yu, University of Washington, Seattle.
Jacob Bien, University of Southern California, Los Angeles.
Daniela Witten, University of Washington, Seattle.
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
