Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2024 Dec 26.
Published in final edited form as: Adv Neural Inf Process Syst. 2024 May 30;36:42762–42774.

Directed Cyclic Graph for Causal Discovery from Multivariate Functional Data

Saptarshi Roy 1, Raymond K W Wong 2, Yang Ni 3
PMCID: PMC11669967  NIHMSID: NIHMS2001062  PMID: 39726833

Abstract

Discovering causal relationship using multivariate functional data has received a significant amount of attention very recently. In this article, we introduce a functional linear structural equation model for causal structure learning when the underlying graph involving the multivariate functions may have cycles. To enhance interpretability, our model involves a low-dimensional causal embedded space such that all the relevant causal information in the multivariate functional data is preserved in this lower-dimensional subspace. We prove that the proposed model is causally identifiable under standard assumptions that are often made in the causal discovery literature. To carry out inference of our model, we develop a fully Bayesian framework with suitable prior specifications and uncertainty quantification through posterior summaries. We illustrate the superior performance of our method over existing methods in terms of causal graph estimation through extensive simulation studies. We also demonstrate the proposed method using a brain EEG dataset.

1. Introduction

Motivation.

Multivariate functional data arise in many fields such as biomedical research [Wei and Li, 2008, Chiou and Müller, 2016], environmental science [Korte-Stapff et al., 2022], finance [Kowal et al., 2017], plant science [Wong et al., 2019, Park et al., 2022], and sport science [Volkmann et al., 2021] where multiple variables are measured over time or other domains. The increasing availability of functional data in these fields provides us with great opportunities to discover causal relationships among random functions for the better understanding of complex systems, which is helpful for various machine learning and statistics tasks such as representation learning [Schölkopf et al., 2021], fairness [Tang et al., 2023], transfer learning [Rojas-Carulla et al., 2018], and reinforcement learning [Zeng et al., 2023]. One motivating example is electroencephalography (EEG) where electrical activity from the brain is recorded non-invasively from electrode channels by placing them on the scalp or directly on the surface of the brain. Given its continuous nature and the short time separation between the adjacent measuring points, it is natural to treat the data at each brain location/region as a function over time. A relevant scientific goal is to estimate brain effective connectivity among different regions, which will potentially allow us to make better decisions, design more effective interventions, and avoid unintended consequences. However, existing structural equation model (SEM) based causal discovery methods assume acyclic relationships among the random functions by imposing a directed acyclic graph (DAG) structure, which may be too restrictive for many real applications. For example, there are strong indications that in brain effective connectivity studies, due to reciprocal polysynaptic connections, the brain regions are far from exhibiting acyclicity [Friston, 2011, Markov et al., 2012], and that in genetic pathways, due to the presence of multiple upstream regulators and downstream targets for every signaling component, feedback loops/directed cycles are regular motifs [Brandman and Meyer, 2008]. Thus, in light of the prevalence of cycles in complex systems, it is desirable to have a flexible model for causal discovery among random functions that can account for such cyclic causal structures.

Challenges.

Causal discovery for multivariate functional data in the presence of cycles is an inherently difficult problem that is not yet well understood. We highlight three prominent challenges. (i) Functional data are infinite-dimensional in nature. It may so happen that the low-frequency spectrum of one curve might causally influence the high-frequency spectrum of another curve. This demands identification of pertinent features that can be used to create a finite-dimensional representation of the data, which is easier to work with and analyze. However, the challenge is that we may not know a priori what the relevant features are when dealing with infinite-dimensional objects. Blind adoption of standard (non-causal-adaptive) low-dimensional features can lead to errors or inaccuracies. (ii) Although the identifiability of causal models for multivariate functional data in the absence of cycles has been established in recent works [Zhou et al., 2022b, Lee and Li, 2022], showing identifiability of causal models from multivariate data, let alone multivariate functions, is still a challenging and complex task in cases where causal relationships are obscured by the presence of cycles. (iii) It is common that functional data are only observed over discrete time points with additional noises. Such incomplete and noisy observations of the functions add another layer of difficulty in probing the causal relationships of interest.

Related work.

Causal discovery from multivariate functional data has been studied by a few recent works [Zhou et al., 2022b, Lee and Li, 2022, Yang and Suzuki, 2022], which have already shown some promising results in discovering causality in, e.g., EEG data and longitudinal medical record data. However, all of them are limited to DAGs, which do not allow inference of cyclic causality. While there has been a surge of research on causal discovery methods for scalar random variables in the presence of feedback loops/cycles over the last few decades [Richardson, 1996, Lacerda et al., 2008, Mooij et al., 2011, Hyttinen et al., 2012, Huang et al., 2019, Mooij and Heskes, 2013, Mooij and Claassen, 2020, Zhou et al., 2022a], none of these approaches have been extended to discovering causal dependencies among random functions in multivariate settings. Therefore, how to handle cyclic causal relationships among multivariate functional data while addressing the aforementioned challenges remains a largely unsolved problem.

Contributions.

In this paper, we propose an operator-based non-recursive linear structural equation based novel causal discovery framework that identifies causal relationships among functional objects in the presence of cycles and additional measurement/sampling noises. Our major contribution is four-fold.

  1. We consider a causal embedding of the functional nodes into a lower-dimensional space for dimension reduction that adapts to causal relationships.

  2. We prove that the causal graph of the proposed model is uniquely identifiable under standard causal assumptions.

  3. We capture within-function dependencies using a data-driven selection of orthonormal basis that is both interpretable and computationally efficient.

  4. To perform inference and uncertainty quantification from finite-sample data, we adopt a fully Bayesian hierarchical formulation with carefully selected prior distributions. Posterior inference is performed using Markov chain Monte Carlo (MCMC). We demonstrate the effectiveness of the proposed method in identifying causal structure and key parameters through simulation studies and apply the framework to the analysis of brain EEG data, illustrating its real-world applicability. Code will be made available on the project’s website on Github.

2. Model definition and causal identifiability

2.1. Notations

Let [p]={1,,p} for any positive integer p. A causal directed cyclic graph (DCG) is a graph 𝒢=(𝒱,), which consists of a set of vertices or nodes 𝒱=[p] representing a set of random variables and a set of directed edges ={jj,𝒱} representing the direct causal relationships among the random variables. In a DCG, we do not assume the graph to be acyclic. A causal DCG model is an ordered pair (𝒢,P) where P is a joint probability distribution over 𝒱 (more rigorously, the random variables that 𝒱 represents) that satisfies conditional independence relationships encoded by the causal DCG 𝒢. A simple directed cycle is a sequence of distinct vertices v1,,vk such that the induced subgraph by these vertices is v1vkv1. For a vertex j𝒱, we use pa(j) to denote the set of parents (direct causes).

2.2. Model framework

Consider a multivariate stochastic process Y=Y1,,Yp where each Yj is defined on a compact domain 𝒯jR. Without loss of generality, we assume 𝒯1==𝒯p=[0,1]. Suppose Yjj where j is a Hilbert space of functions defined on 𝒯j. We let , denote the inner product of j. We propose a causal model that captures the relationships among Y1,,Yp.

Our proposed model considers an operator-based non-recursive linear structural equation model on the random functions Y as

Yj=pajjY+fj,jp, (1)

where j is a linear operator that maps to j, and fjj is an exogenous stochastic process. Clearly, for any j,𝒱 such that the edge j,j is not a null operator. Now by stacking the p equations in (1), we obtain

Y=BY+f, (2)

where B=jj,=1p is a matrix of operators and f=f1,,fp is a p-variate stochastic process. In DAGs, the causal effect matrix can be arranged into a lower block triangular structure given a topological/causal ordering. But since our model allows for cycles, we have no such restriction on the structure of the operator matrix B except that jj,j[p], is null, i.e., no self-loops.

Model (1) is infinite-dimensional and hence challenging to estimate and interpret. To alleviate such difficulties, we consider a low-dimensional causal embedding structure. Specifically, we assume that the causal relationships are preserved in an unknown low-dimensional subspace 𝒟j of j. Denote the dimension of 𝒟j by Kj. Let 𝒫j and 𝒬j be the projection onto 𝒟j and its orthogonal complement in j respectively. We assume j=𝒫jj𝒫, which implies that causal effects can be fully described within the low-dimensional subspaces 𝒟jj=1p. As such, (1) can be split into

𝒫jYj=pajj𝒫Y+𝒫jfj,𝒬jYj=𝒬jfj. (3)

We assume that 𝒫jfj and 𝒬jfj are independent of each other. Now, by defining αj=𝒫jYj and ϵj=𝒫jfj,j[p], (3) can be compactly written as

α=α+ϵ, (4)

where α=α1,,αp and ϵ=ϵ1,,ϵp with αj,ϵj𝒟j,j[p].

In practice, the random functions in Y can only be observed over a finite number of (input) locations, possibly with measurement errors. More specifically, for each random function Yj, we observe tju,Xjuu=1mj, where XjuR is the measurement of Yj at location tju𝒯j and mj is the number of measurements obtained from Yj. Defining βj=𝒬jYj, we consider the following measurement model:

Xju=Yjtju+eju=αjtju+βjtju+eju,umj,jp, (5)

with independent noises eju~N0,σj,umj.

More compactly, (5) can be written as

X=α(t)+β(t)+e, (6)

where X=X1,,Xp,α(t)=α1t1,,αptp,β(t)=β1t1,,βptp and e=e1,,ep with Xj=Xj1,,Xjmj,αjtj=αjtj1,,αjtjmj,βjtj=βjtj1,,βjtjmj and ej=ej1,,ejmj.

We call our proposed model, FENCE, which stands for’Functional Embedded Nodes for Cyclic causal Exploration’, reflecting its purpose.

2.3. Causal identifiability

In this section, we shall show that the graph structure of the proposed FENCE model is identifiable for functional data measured discretely with random noises under several causal assumptions. We start by defining causal identifiability and state our assumptions.

Definition 2.1. (Causal Identifiability) Suppose Y is a p-variate random function and X is the observed noisy version of Y given by (6). Assume X follows FENCE model 𝒮=(𝒢,P) where 𝒢 is the underlying graph and P is the joint distribution of X over 𝒢. We say that 𝒮 is causally identifiable from X if there does not exist any other 𝒮*=𝒢*,P* with 𝒢*𝒢 such that the joint distribution P* on X induced by 𝒢* is equivalent to P induced by 𝒢.

In other words, for a causal graph to be identifiable, there must not exist any other graph such that the joint distributions induced by the two different graphs are equivalent. Next, we list and discuss a few assumptions to establish the causal identifiability of the proposed model.

Assumption 1. (Causal Sufficiency) The model 𝒮=(𝒢,P) is causally sufficient, i.e., there are no unmeasured confounders.

Assuming no unmeasured confounders keeps the causal discovery task more manageable especially for cyclic graphs with purely observational data.

Assumption 2. (Disjoint Cycles) The cycles in 𝒢 are disjoint, i.e., no two cycles in the graph have two nodes that are common to both.

Assuming disjoint cycles induces a natural topological ordering and forms a directed acyclic hypergraph-like structure within the DCG. The same assumption was made in Lacerda et al. [2008].

Assumption 3. (Stability) For the model 𝒮, the moduli of the eigenvalues of the finite rank operator are less than or equal to 1, and none of the real eigenvalues are equal to 1.

According to Fisher, 1970, the SEM in (4) can be viewed as being in a state of equilibrium, where the finite rank operator represents coefficients in a set of dynamical equations that describe a deterministic dynamical system observed over small time intervals as the time lag approaches zero. The eigenvalue conditions are deemed necessary and sufficient for the limiting behavior to hold, as argued by Fisher, 1970. Such an assumption is widely adopted in e.g., econometrics, and Lacerda et al., 2008 made this assumption as well.

Assumption 4. (Non-Gaussianity) The exogenous variables have independent mixture of Gaussian distributions. i.e., ϵjk~indm=1MjkπjkmNμjkm,τjkm with Mjk2.

The assumption of non-Gaussianity on the exogenous variables has been proven useful in causal discovery as it induces model identifiability in the linear SEM framework [Aapo and Petteri, 1999, Shimizu et al., 2006, Lacerda et al., 2008, Spirtes and Zhang, 2016]. Mixture of Gaussian can approximate any continuous distribution arbitrarily well given a sufficiently large number of mixture components [Titterington et al., 1985, McLachlan and Peel, 2000, Rossi, 2014]. It is also easy to sample, which facilitates our posterior inference.

Assumption 5. (Non-causal dependency) We assume β(t)=C(t)γ, where C(t)=diagC11t1,,Cpptp and γ represent another exogenous component of the model. Here Cjjtj is a mixing matrix that mixes the independent entires in γ to generate temporal dependence within the j-th block. We assume γjk~indm=1MjkπjkmNμjkm,τjkm with Mjk1.

Since the model assumes that all causal information in Y is preserved in the lower-dimensional space 𝒟j and not in its orthogonal complement, it is apparent that while each βjtj within a block can have temporal dependence, it is independent of βt when j and j,[p].

For some basis ϕjkk=1Kj that spans the low-dimensional causal embedded space 𝒟j,αj in (5) can be further expanded by, αjtju=k=1Kjα˜jkϕjktju. Therefore (6) can be written more compactly as

X=Φtα˜+βt+e, (7)

where Φ(t)=diagΦ1t1,,Φptp with Φjtj=ϕjvtjuu=1,v=1mj,Kj.

Assumption 6. (Sufficient sampling locations) The basis matrix Φ(t) of size j=1pmj×j=1pKj has a full column rank.

This assumption implies enough sampling locations, over which each random function Yj is observed, to capture all the causal information that Yj contains.

Given these six assumptions, our main theorem establishes the causal identifiability of the proposed model.

Theorem 2.1. Under Assumptions 1-6,𝒮=(𝒢,P) is causally identifiable.

The proof essentially involves two steps as shown in Figure 1. On the left-hand side (LHS) of the diagram, we depict the hypergraph-like structure that emerges when assuming the existence of disjoint cycles (Assumption 2), whereas, on the right-hand side (RHS), we offer a magnified view of the hypernodes (nodes containing simple directed cycle). Our approach to proving causal identifiability progresses from the LHS to the RHS. That is, we first prove the identifiability of the hypergraph-like structure depicted on the LHS of Figure 1, and then we proceed to establish the identifiability of each simple directed cycle within every hypernode in the hypergraph. The detailed exposition of the proof can be found in Section A of the Supplementary Materials.

Figure 1:

Figure 1:

Two important components of causal identifiability proof: (I) identifiability of directed acyclic hypergraph induced by disjoint cycles, and (II) identifiability of each disjoint cycle.

3. Bayesian model formulation

In this section, we will describe the inference procedure of the proposed model. A straightforward approach would be a two-step procedure where the first step performs functional principal component analysis on each function marginally to reduce the dimension, and then the second step learns causal structure based on the principal components. However, this simple approach has several disadvantages. First, the estimated functional principal components that explain the most variation of each individual function marginally may not optimally capture the cause-effect dependence relationships among different functions. Second, this procedure is unreliable since estimation uncertainty fails to propagate correctly from the first step to the second step. As such, we propose a fully Bayesian approach, which reduces the dimension of functional data adaptively for causal structure learning.

3.1. Model parameters

Let E=Ejj,=1p denote the adjacency matrix where Ej=1 indicates the existence of a directed edge from node to node j, and 0 otherwise. Let ϕkk=1S be a set of S common unknown basis functions that approximate each random function Yj i.e., Yj=k=1Sα˜jkϕk where α˜jkk=1S denote the set of basis coefficients. Note that ϕk is not the basis for the lower-dimensional causal embedded subspace 𝒟j. However, we assume that the first Kj of them actually spans 𝒟j and our goal is to hunt for them through a properly designed inference procedure. Moreover, according to our assumption, we build our SEM on the first Kj of the basis coefficients α˜j=α˜j1,,α˜jKj. Defining α-j=α˜j,Kj+1,,α˜jS with α-j=γj, jointly they can be written as

α˜=B˜α˜+ϵ˜, (8)

where α˜=α˜1,,α˜p,α-1,,α-p,ϵ˜=ϵ˜1,,ϵ˜p,γ1,,γp with ϵ˜j=ϵ˜j1,,ϵ˜jKj and γj=γj,Kj+1,,γjS. Here B˜=B000 where B=Bj(a,b)a=1,b=1Kj,Klj,=1p with Bjj=0 since we assume the absence of self loops. To carry out inference, we assume ϵ˜jk,γjk~indm=1MjkπjkmNμjkm,τjkm.

3.2. Adaptive basis expansion

As the ϕk’s are specifically useful for restricting the original function space for each Yj to a lower-dimensional causally embedded smooth space of dimension Kj, we make the basis ϕk adaptive for causal structure learning by further expanding them with known spline basis functions [Kowal et al., 2017], ϕk()=r=1RAkrbr(), where b=b1,,bR is the set of fixed cubic B-spline basis functions with equally spaced knots and Ak=Ak1,,AkR are the corresponding spline coefficients. Since we do not fix Ak’s a priori, the basis functions ϕk’s can be learned from data a posteriori and hence are adaptive to both data and causal structure (i.e., the basis functions, the functional data, and the causal graph are dependent in their joint distribution).

3.3. Prior specifications

Prior on spline coefficients.

The prior on Ak is chosen to serve multiple purposes. (i) It sorts the basis functions by decreasing smoothness and therefore helps to identify the spanning set of size Kj for the underlying smooth causally embedded space 𝒟j. (ii) Although not a strict requirement for modelling purpose, it forces ϕk’s to be orthonormal, i.e. ϕk(ω)ϕk(ω)dω=Ik=k. As such, the orthogonality constraints help eliminate any information overlap between the basis functions, which keeps the total number of necessary basis functions that actually contribute to the causal structure learning to a minimum. (iii) It regularizes the roughness of ϕk’s to prevent overfitting.

For (iii), more specifically, we restrict the roughness of the basis functions ϕk() by assigning a prior that penalizes its second derivatives [Gu, 1992, Wahba, 1978, Berry et al., 2002]:

Ak~N0,λk-1Ω-,

where Ω- is the pseudoinverse of Ω=b(t)b(t)dt. Let Ω=UDU be the singular value decomposition of Ω. Following Wand and Ormerod, 2010, to facilitate computation, we reparameterize ϕk=r=1RA˜krb˜k with b˜()=1,t,bT()U˜D˜-12 where D˜ is the (R-2)×(R-2) submatrix of D corresponding to non-zero singular values (note that the rank of Ω is R-2 by definition) and U˜ is the corresponding R×(R-2) submatrix of U. This induces a prior on A˜k given by

A˜k~N0,SkwithSk=diag,,λk1,,λk1.

In other words, the intercept and the linear term are unpenalized but the non-linear terms are penalized, the degree of which is controlled by λk. In practice, we set the first two diagonal elements of Sk as 108. We constrain the regularization parameters λ1>>λS>0 by putting a uniform prior:

λk~UniformLk,Uk,kS,U1=108,Lk=λk+1kS-1,Uk=λk-1k2,,S,LS=10-8,

which implies that the smoothness of ϕk() decreases as k gets larger.

Priors on the adjacency matrix.

We propose to use an independent uniform-Bernoulli prior on each entry Ej of E, i.e., Ejρ~indBernoulli(ρ) and ρ~Uniform(0,1). The marginal distribution of E with ρ integrated out is given by

pE=pEρpρdρ=BetajEj+1,j1Ej+1.

Now, for example, if E0 denotes the null adjacency matrix and E1 denotes the adjacency matrix with only one edge, then we can see that pE0/pE1=p2-p. Therefore, an empty graph is favored over a graph with one edge by a factor of p2-p, and, importantly, this penalty increases with p. Thus, the uniform-Bernoulli prior prevents false discoveries and leads to a sparse network by increasing the penalty against additional edges as the dimension p grows.

Prior on the causal effect matrix.

Now given E, we assume an independent spike and slab prior on the entries of B=Bjj,=1p:

BjEj~1-EjMVNBj;0,sγIKj,IK+EjMVNBj;0,γIKj,IK,

where MVNBj;0,γIKj,IK is a matrix-variate normal distribution with row and column co-variance matrices as γIKj and IK, respectively. We assume a conjugate inverse-gamma prior on the causal effect size, γ~InverseGammaaγ,bγ. We choose aγ=bγ=1. We fix s=0.02 so that when Ej=0,Bj is negligibly small.

Priors on the parameters of the Gaussian mixture distribution.

We choose conjugate priors for the parameters of the Gaussian mixture distribution:

πjk1,,πjkMjk~Dirichlet(β,,β),j[p],k[S]μjkm~Naμ,bμ,τjkm~InverseGammaaτ,bτ,j[p],k[S],mMjk

We have fixed values for the hyperparameters, β=1,aμ=0,bμ=100,aτ=bτ=1.

Prior on the noise variances.

We assume a conjugate prior for σj~InverseGammaaσ,bσ,j[p]. We choose aσ=bσ=0.01.

We simulate posterior samples through Markov chain Monte Carlo (MCMC). Details are given in Section B of the Supplementary Materials. Sensitivity analyses will be conducted to test the hyperparameters including aγ,bγ,aτ,bτ,aσ,bσ,s,R,S,M and β.

4. Simulation study

Data generation

The data were simulated according to various combinations of sample size (n), number of nodes (p), and grid size (mj=dj[p]) where n{75,150,300},p{20,40,60}, and d{125,250}. The grid evenly spans the unit interval [0,1]; the results with unevenly spaced grids are presented in Section C of the Supplementary Materials. The true causal graph 𝒢 was generated randomly with edge formation probability 2/p. Given 𝒢, each non-zero block Bj of the causal effect matrix was generated from the standard matrix-variate normal distribution. We set the true number of basis functions to be K=4. In order to generate K=4 orthonormal basis functions, we first simulated unnormalized basis functions by expanding them further with 6 cubic B-spline basis functions where the coefficients were drawn from the standard normal distribution and then empirically orthonormalized them. The basis coefficients α˜ were generated following (8) with the exogenous variables ϵ˜j drawn independently from Laplace distribution with location parameter μ=0 and scale parameter b=0.2. We have also considered other non-Gaussian distributions for the exogenous variables; the corresponding results are provided in Section C of the Supplementary Materials. Finally, noisy observations were simulated following (6) with the signal-to-noise ratio, i.e., the mean value of Yj(i)(t)/σj across all i and t, set to 5. Here, superscript (i) denotes the ith sample, where i[n].

For the implementation of the proposed FENCE, we fixed the number of mixture components to be 10 and ran MCMC for 5,000 iterations (discarding the first 2,000 iterations as burn-in and retaining every 5th iteration after burn-in). The causal graph G was then estimated by using the median probability model [Barbieri and Berger, 2004], i.e., by thresholding the posterior probability of inclusion at 0.5.

Methods for comparison.

We compared our method with fLiNG [Zhou et al., 2022b], a recently proposed directed acyclic graph (DAG) for multivariate functional data. Codes for Lee and Li [2022], Yang and Suzuki [2022] are not publicly available. Hence for more comparison, we considered two ad hoc two-step approaches. In the first step of both approaches, we obtained the basis coefficients by carrying out functional principal component analysis (fPCA) using the fdapace [Zhou et al., 2022c] package in R. Then in the second step, given the basis coefficients, we estimated causal graphs using existing causal discovery methods, (i) LiNGAM [Shimizu et al., 2006] (ii) PC [Spirtes and Glymour, 1991] and (iii) CCD [Richardson, 1996]; we call these three approaches fPCA-LiNGAM, fPCA-PC and fPCA-CCD respectively. Note that we did not use SEM-based cyclic discovery algorithm, LiNG-D [Lacerda et al., 2008] in the second step due to the unavailability of the code.

LiNGAM estimates a causal DAG based on the linear non-Gaussian assumption whereas PC generally returns only an equivalence class of DAGs based on conditional independence tests. CCD algorithm is a constraint-based causal discovery method, which yields an equivalence class of cyclic causal graphs. LiNGAM and PC are implemented in the pcalg package [Kalisch et al., 2018] in R. CCD algorithm is implemented in the py-tetrad [Ramsey et al., 2018] package in python.

Performance metrics.

To assess the graph recovery performance, we calculated the true positive rate (TPR), false discovery rate (FDR), and Matthew’s correlation coefficient (MCC). For TPR and MCC, higher is better, whereas lower FDR is better.

Results.

Table 1 summarizes the results of 50 repeat simulations, demonstrating that the proposed FENCE model outperforms all competitors (fLiNG, fPCA-LiNGAM, and fPCA-CCD) across all combinations of n,p, and d. We provide the results of fPCA-PC in Section D of the Supplementary Materials, which are similar to those of fPCA-LiNGAM. We favored fPCA-PC and fPCA-CCD by counting a non-invariant edge between two nodes as a true positive as long as the two nodes are adjacent in the true graph. The superiority of FENCE is not unexpected for three reasons. First, fLiNG, fPCA-LiNGAM, and fPCA-PC are not specifically designed for learning cyclic graphs. Second, two-step approaches like fPCA-LiNGAM, fPCA-PC, and fPCA-CCD do not necessarily capture the causally embedded space through the functional principal components. Third, although fPCA-CCD can handle cyclic graphs, it, being a two-step approach, fails to capture true functional dependencies. Overall, these findings provide strong evidence of the effectiveness of FENCE compared to existing methods.

Table 1:

Comparison of performance of various methods under 50 replicates. Since LiNGAM is not applicable to cases where q>n with q=Kp being the total number of extracted basis coefficients across all functions, the results from those cases are not available and indicated by “-”. The metrics reported are based on 50 repetitions are reported; standard deviations are given within the parentheses.

n p d FENCE fLiNG fPCA-LINGAM fPCA-CCD
TPR FDR MCC TPR FDR MCC TPR FDR MCC TPR FDR MCC
75 20 125 0.85(0.09) 0.19(0.07) 0.88(0.05) 0.41(0.09) 0.79(0.05) 0.36(0.04) 0.35(0.19) 0.84(0.04) 0.11(0.08) 0.69(0.03) 0.41(0.04) 0.23(0.03)
75 40 125 0.79(0.08) 0.23(0.06) 0.86(0.04) 0.37(0.08) 0.82(0.06) 0.33(0.05) - - - 0.73(0.02) 0.47(0.04) 0.21(0.05)
75 60 125 0.75(0.07) 0.27(0.05) 0.83(0.04) 0.34(0.07) 0.83(0.06) 0.32(0.04) - - - 0.68(0.03) 0.61(0.05) 0.19(0.03)
150 20 125 0.88(0.07) 0.14(0.06) 0.89(0.05) 0.45(0.07) 0.75(0.06) 0.39(0.05) 0.28(0.22) 0.86(0.05) 0.08(0.09) 0.71(0.03) 0.42(0.03) 0.25(0.04)
150 40 125 0.81(0.07) 0.21(0.06) 0.87(0.05) 0.39(0.06) 0.79(0.05) 0.37(0.04) 0.35(0.22) 0.91(0.02) 0.08(0.06) 0.73(0.04) 0.47(0.05) 0.23(0.03)
150 60 125 0.79(0.06) 0.24(0.05) 0.86(0.04) 0.36(0.04) 0.80(0.06) 0.36(0.05) - - - 0.72(0.05) 0.54(0.04) 0.22(0.02)
300 20 125 0.91(0.03) 0.09(0.04) 0.90(0.04) 0.51(0.04) 0.73(0.06) 0.41(0.04) 0.30(0.19) 0.84(0.05) 0.11(0.09) 0.81(0.03) 0.39(0.04) 0.26(0.03)
300 40 125 0.87(0.04) 0.15(0.05) 0.87(0.05) 0.47(0.05) 0.75(0.06) 0.38(0.05) 0.27(0.20) 0.91(0.02) 0.08(0.06) 0.77(0.03) 0.45(0.02) 0.24(0.03)
300 60 125 0.85(0.05) 0.17(0.03) 0.86(0.03) 0.45(0.05) 0.76(0.04) 0.38(0.03) 0.28(0.17) 0.91(0.05) 0.05(0.03) 0.72(0.03) 0.49(0.02) 0.22(0.03)
75 20 250 0.81(0.04) 0.23(0.02) 0.85(0.05) 0.39(0.07) 0.80(0.05) 0.39(0.04) 0.32(0.14) 0.82(0.03) 0.09(0.04) 0.67(0.03) 0.46(0.03) 0.22(0.04)
75 40 250 0.73(0.04) 0.28(0.05) 0.82(0.04) 0.35(0.04) 0.85(0.06) 0.33(0.05) - - - 0.68(0.02) 0.51(0.04) 0.21(0.03)
75 60 250 0.67(0.03) 0.34(0.05) 0.79(0.04) 0.34(0.04) 0.85(0.03) 0.31(0.04) - - - 0.63(0.04) 0.56(0.04) 0.19(0.04)
150 20 250 0.83(0.06) 0.17(0.05) 0.86(0.05) 0.46(0.07) 0.73(0.07) 0.42(0.05) 0.32(0.19) 0.79(0.05) 0.13(0.05) 0.73(0.04) 0.43(0.02) 0.24(0.03)
150 40 250 0.79(0.02) 0.26(0.06) 0.82(0.03) 0.41(0.05) 0.71(0.05) 0.40(0.03) 0.31(0.14) 0.81(0.02) 0.13(0.06) 0.71(0.03) 0.47(0.04) 0.23(0.03)
150 60 250 0.69(0.05) 0.31(0.05) 0.79(0.04) 0.43(0.03) 0.79(0.06) 0.43(0.05) - - - 0.69(0.02) 0.52(0.03) 0.21(0.02)
300 20 250 0.86(0.02) 0.16(0.04) 0.85(0.04) 0.68(0.02) 0.77(0.07) 0.47(0.04) 0.45(0.13) 0.86(0.05) 0.17(0.09) 0.78(0.02) 0.44(0.06) 0.27(0.03)
300 40 250 0.79(0.08) 0.16(0.05) 0.84(0.06) 0.73(0.05) 0.71(0.06) 0.43(0.05) 0.39(0.16) 0.87(0.05) 0.16(0.07) 0.76(0.05) 0.49(0.06) 0.23(0.05)
300 60 250 0.76(0.05) 0.21(0.03) 0.80(0.03) 0.77(0.05) 0.74(0.03) 0.42(0.03) 0.28(0.17) 0.90(0.04) 0.13(0.04) 0.72(0.06) 0.53(0.03) 0.22(0.04)

Additional simulations.

We considered additional simulation scenarios with unevenly spaced grids, general exogenous variable distributions, true acyclic graphs and data generated using non-linear SEM, and also conducted sensitivity analyses of FENCE with respect to several hyperparameters; the results are presented in Section C of the Supplementary Materials. The performance of the proposed method is consistently better than competing methods and is relatively robust with respect to hyperparameter choices.

5. Real data application

Brain EEG data.

We demonstrate the proposed FENCE model on a brain EEG dataset from an alcoholism study [Zhang et al., 1995]. This dataset was earlier used to demonstrate functional undirected graphical models [Zhu et al., 2016, Qiao et al., 2019] and functional Bayesian network [Zhou et al., 2022b]. Data were initially obtained from 64 electrodes placed on subjects’ scalps, which captured EEG signals at 256 Hz (3.9 ms epoch) during a one-second period. The study consists of 122 subjects, out of which 77 are in the alcoholic group and 45 are in the control group. Each subject completed 120 trials. During each trial, the subject was exposed to either a single stimulus (a single picture) or two stimuli (a pair of pictures) shown on a computer monitor. We particularly focus on the EEG signals filtered at α frequency bands between 8 and 12.5Hz using the eegfilt function of the eeglab toolbox of Matlab as α band signals are associated with inhibitory control [Knyazev, 2007]. Given that the EEG measurements were recorded from each subject over multiple trials, these measurements are not independent of each other due to the time dependency of the trials. Moreover, since the measurements were obtained under various stimuli, the signals may have been affected by different stimulus effects. To mitigate these issues, we calculated the average of the band-filtered EEG signals for each subject across all trials under a single stimulus, resulting in a single event-related potential curve per electrode per subject. By doing so, we eliminated the potential dependence between the measurements and the influence of different stimulus types. We performed separate analyses of the two groups to identify both the similarities and dissimilarities in their brain effective connectivity.

We conducted a Shapiro-Wilk normality test on the observed functions for each of the p=64 scalp positions at each of the mj=256j[p] time points to evaluate their Gaussianity. The results showed that for numerous combinations of scalp position and time point, the null hypothesis (which assumes that the observations are marginally Gaussian) was rejected. Thus, we conclude that the non-Gaussianity of the proposed model is appropriate. Next, for posterior inference, we ran MCMC for 20,000 iterations, discarded the first half as burn-in, and retained every 10th iteration after burn-in. The estimated causal graph by thresholding the posterior inclusion probability to 0.9 is given below in Figure 2.

Figure 2:

Figure 2:

Estimated causal brain connectivity from EEG records by FENCE with posterior probability of inclusion 0.9, separately for the alcoholic (left) and control (right) group. The bi-directed edges are just directed cycles, i.e., ij means ij and ij.

Results.

There are some interesting findings. First, for both groups (alcoholic and control), brain regions that are located in adjacent positions tend to be more connected than the brain regions that are far apart. Second, dense connectivity is observed in the frontal region of the brain in both groups, with multiple cycles being formed. Third, compared to the control group, the alcoholic group has more connectivity across the left parietal and occipital lobes. Fourth, the same cycle of Iz, Cz, and RPA is observed in both groups.

Validity.

We now discuss the validity of our real data results. In Hayden et al., 2007, it was observed that alcohol-dependent subjects exhibited frontal asymmetry, distinguishing them from the control group. Our own investigation aligns well with these results, as we have identified denser connectivity across various brain regions in the middle and left areas of the frontal lobe among alcoholic subjects, when compared to controls. Furthermore, Winterer et al., 2003 documented coherent differences between alcoholics and controls in the posterior hemispheres, specifically in the temporal, parietal, and occipital lobes. In accordance with their findings, our study provides additional support for this claim, as we have observed heightened activity with several cycles formed in those same regions within the alcoholic group when compared to the control group.

6. Discussion

We briefly highlight here several potential avenues for future development of our current work. First, an intriguing and important direction would be to explore the relaxation of the causal sufficiency assumption in the model identifiability. Second, our current model is based on a linear non-Gaussian assumption over the exogenous variables, but a nonlinear model could be considered as an alternative. Lastly, an alternative approach to determining the effective number of basis functions that span the lower-dimensional causal embedded space would be to utilize the ordered shrinkage priors [Bhattacharya and Dunson, 2011, Legramanti et al., 2020] in order to adaptively eliminate redundant components, resulting in a more flexible methodology.

Supplementary Material

Supplementary

Acknowledgements

Ni’s research was partially supported by NSF DMS-2112943 and NIH 1R01GM148974-01.

Contributor Information

Saptarshi Roy, Department of Statistics Texas A&M University College Station, TX 77843.

Raymond K. W. Wong, Department of Statistics Texas A&M University College Station, TX 77843

Yang Ni, Department of Statistics Texas A&M University College Station, TX 77843.

References

  1. Aapo Hyvärinen and Petteri Pajunen. Nonlinear independent component analysis: Existence and uniqueness results. Neural Networks, 12:429–439, 1999. URL 10.1016/S0893-6080(98)00140-3. [DOI] [PubMed] [Google Scholar]
  2. Barbieri Maria Maddalena and Berger James O.. Optimal predictive model selection. The Annals of Statistics, 32(3):870 – 897, 2004. doi: 10.1214/009053604000000238. [DOI] [Google Scholar]
  3. Berry Scott M, Carroll Raymond J, and Ruppert David. Bayesian smoothing and regression splines for measurement error problems. Journal of the American Statistical Association, 97 (457):160–169, 2002. doi: 10.1198/016214502753479301. [DOI] [Google Scholar]
  4. Bhattacharya A and Dunson DB. Sparse Bayesian infinite factor models. Biometrika, 98(2): 291–306, 06 2011. ISSN 0006–3444. doi: 10.1093/biomet/asr013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Brandman Onn and Meyer Thomas. Feedback loops shape cellular signals in space and time. Science, 322(5900):390–395, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Chiou Jeng-Min and Müller Hans-Georg. A pairwise interaction model for multivariate functional and longitudinal data. Biometrika, 103(2):377–396, 2016. [DOI] [PubMed] [Google Scholar]
  7. Fisher Franklin M.. A correspondence principle for simultaneous equation models. Econometrica, 38 (1):73–92, 1970. ISSN 00129682, 14680262. URL http://www.jstor.org/stable/1909242. [Google Scholar]
  8. Friston Karl J.. Functional and effective connectivity: A review. Brain Connectivity, 1(1):13–36, 2011. doi: 10.1089/brain.2011.0008. [DOI] [PubMed] [Google Scholar]
  9. Gu Chong. Penalized likelihood regression: A bayesian analysis. Statistica Sinica, 2:255–264, 1992. [Google Scholar]
  10. Hayden Elizabeth, Wiegand Ryan, Meyer Eric, Bauer Lance, Sean O’Connor John Nurnberger, Chorlian David, Porjesz Bernice, and Begleiter Henri. Patterns of regional brain activity in alcohol-dependent subjects. Alcoholism, clinical and experimental research, 30:1986–91, 01 2007. doi: 10.1111/j.1530-0277.2006.00244.x. [DOI] [PubMed] [Google Scholar]
  11. Huang Biwei, Zhang Kun, Zhang Jiji, Ramsey Joseph D., Sanchez-Romero Ruben, Glymour Clark, and Schölkopf Bernhard. Causal discovery from heterogeneous/nonstationary data. CoRR, abs/1903.01672, 2019. URL http://arxiv.org/abs/1903.01672. [Google Scholar]
  12. Hyttinen Antti, Eberhardt Frederick, and Hoyer Patrik O.. Causal discovery of linear cyclic models from multiple experimental data sets with overlapping variables. In de Freitas Nando and Murphy Kevin P., editors, Proceedings of the Twenty-Eighth Conference on Uncertainty in Artificial Intelligence, Catalina Island, CA, USA, August 14–18, 2012, pages 387–396. AUAI Press, 2012. URL https://dslpitt.org/uai/displayArticleDetails.jsp?mmnu=1&smnu=2&article_id=2301&proceeding_id=28. [Google Scholar]
  13. Kalisch Markus, Hauser A, Maathuis Marloes H., and Mächler Martin. An overview of the pcalg package for r, 2018.
  14. Knyazev Gennady G. Motivation, emotion, and their inhibitory control mirrored in brain oscillations. Neuroscience & Biobehavioral Reviews, 31(3):377–395, 2007. [DOI] [PubMed] [Google Scholar]
  15. Korte-Stapff Moritz, Yarger Drew, Stoev Stilian, and Hsing Tailen. A multivariate functional-data mixture model for spatio-temporal data: inference and cokriging, 2022.
  16. Kowal Daniel R, Matteson David S, and Ruppert David. A bayesian multivariate functional dynamic linear model. Journal of the American Statistical Association, 112(518):733–744, 2017. [Google Scholar]
  17. Lacerda Gustavo, Spirtes Peter L., Ramsey Joseph, and Hoyer Patrik O.. Discovering cyclic causal models by independent components analysis. In Conference on Uncertainty in Artificial Intelligence, 2008. [Google Scholar]
  18. Lee Kuang-Yao and Li Lexin. Functional structural equation model. Journal of the Royal Statistical Society Series B, 84(2):600–629, April 2022. doi: 10.1111/rssb.12471. [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Legramanti Sirio, Durante Daniele, and Dunson David B. Bayesian cumulative shrinkage for infinite factorizations. Biometrika, 107(3):745–752, 05 2020. ISSN 0006–3444. doi: 10.1093/biomet/asaa008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Markov NT, Ercsey-Ravasz MM, Ribeiro Gomes AR, Lamy C, Magrou L, Vezoli J, Misery P, Falchier A, Quilodran R, Gariel MA, Sallet J, Gamanut R, Huissoud C, Clavagnier S, Giroud P, Sappey-Marinier D, Barone P, Dehay C, Toroczkai Z, Knoblauch K, Van Essen DC, and Kennedy H. A Weighted and Directed Interareal Connectivity Matrix for Macaque Cerebral Cortex. Cerebral Cortex, 24(1):17–36, 09 2012. ISSN 1047–3211. doi: 10.1093/cercor/bhs270. [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. McLachlan GJ and Peel D. Finite mixture models. Wiley Series in Probability and Statistics, New York, 2000. [Google Scholar]
  22. Mooij Joris M. and Claassen Tom. Constraint-based causal discovery using partial ancestral graphs in the presence of cycles. In Adams Ryan P. and Gogate Vibhav, editors, Proceedings of the Thirty-Sixth Conference on Uncertainty in Artificial Intelligence, UAI 2020, virtual online, August 3–6, 2020, volume 124 of Proceedings of Machine Learning Research, pages 1159–1168. AUAI Press, 2020. URL http://proceedings.mlr.press/v124/m-mooij20a.html. [Google Scholar]
  23. Mooij Joris M. and Heskes Tom. Cyclic causal discovery from continuous equilibrium data. In Nicholson Ann E. and Smyth Padhraic, editors, Proceedings of the Twenty-Ninth Conference on Uncertainty in Artificial Intelligence, UAI 2013, Bellevue, WA, USA, August 11–15, 2013 AUAI Press, 2013. URL https://dslpitt.org/uai/displayArticleDetails.jsp?mmnu=1&smnu=2&article_id=2404&proceeding_id=29. [Google Scholar]
  24. Mooij Joris M., Janzing Dominik, Heskes Tom, and Schölkopf Bernhard. On causal discovery with cyclic additive noise models. In Shawe-Taylor John, Zemel Richard S., Bartlett Peter L., Pereira Fernando C. N., and Weinberger Kilian Q., editors, Advances in Neural Information Processing Systems 24: 25th Annual Conference on Neural Information Processing Systems 2011. Proceedings of a meeting held 12–14 December 2011, Granada, Spain, pages 639–647, 2011. URL https://proceedings.neurips.cc/paper/2011/hash/d61e4bbd6393c9111e6526ea173a7c8b-Abstract.html. [Google Scholar]
  25. Park Yeonjoo, Li Bo, and Li Yehua. Crop yield prediction using bayesian spatially varying coefficient models with functional predictors. Journal of the American Statistical Association, pages 1–14, 2022.35757777 [Google Scholar]
  26. Qiao Xinghao, Guo Shaojun, and James Gareth M. Functional graphical models. Journal of the American Statistical Association, 114(525):211–222, 2019. [Google Scholar]
  27. Ramsey Joseph D, Zhang Kun, Glymour Madelyn, Sanchez Romero Ruben, Huang Biwei, Ebert-Uphoff Imme, Samarasinghe Savini, Barnes Elizabeth A, and Glymour Clark. Tetrad—a toolbox for causal discovery. In 8th international workshop on climate informatics, 2018. [Google Scholar]
  28. Richardson Thomas. A discovery algorithm for directed cyclic graphs. In Proceedings of the Twelfth International Conference on Uncertainty in Artificial Intelligence, UAI’96, page 454–461, San Francisco, CA, USA, 1996. Morgan Kaufmann Publishers Inc. ISBN 155860412X. [Google Scholar]
  29. Rojas-Carulla Mateo, Schölkopf Bernhard, Turner Richard, and Peters Jonas. Invariant models for causal transfer learning. J. Mach. Learn. Res, 19(1):1309–1342, jan 2018. ISSN 1532–4435. [Google Scholar]
  30. Rossi Peter E.. Bayesian Non- and Semi-parametric Methods and Applications. Princeton University Press, 2014. ISBN 9780691145327. URL http://www.jstor.org/stable/j.ctt5hhrfp. [Google Scholar]
  31. Schölkopf Bernhard, Locatello Francesco, Bauer Stefan, Ke Nan Rosemary, Kalchbrenner Nal, Goyal Anirudh, and Bengio Yoshua. Toward causal representation learning. Proceedings of the IEEE, 109(5):612–634, 2021. doi: 10.1109/JPROC.2021.3058954. [DOI] [Google Scholar]
  32. Shimizu Shohei, Hoyer Patrik O., Hyvrinen Aapo, and Kerminen Antti. A linear non-gaussian acyclic model for causal discovery. Journal of Machine Learning Research, 7(72):2003–2030, 2006. URL http://jmlr.org/papers/v7/shimizu06a.html. [Google Scholar]
  33. Spirtes Peter and Glymour Clark. An algorithm for fast recovery of sparse causal graphs. Social Science Computer Review, 9(1):62–72, 1991. doi: 10.1177/089443939100900106. [DOI] [Google Scholar]
  34. Spirtes Peter and Zhang Kun. Causal discovery and inference: concepts and recent methodological advances. Applied Informatics, 3, 2016. doi: 10.1186/s40535-016-0018-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Tang Zeyu, Zhang Jiji, and Zhang Kun. What-is and how-to for fairness in machine learning: A survey, reflection, and perspective. ACM Computing Surveys, 55(13s):1–37, jul 2023. doi: 10.1145/3597199. [DOI] [Google Scholar]
  36. Titterington DM, Smith AFM, and Makov UE. Statistical Analysis of Finite Mixture Distributions. Wiley, New York, 1985. [Google Scholar]
  37. Volkmann Alexander, Stöcker Almond, Scheipl Fabian, and Greven Sonja. Multivariate functional additive mixed models, 2021.
  38. Wahba Grace. Improper priors, spline smoothing and the problem of guarding against model errors in regression. Journal of the Royal Statistical Society: Series B (Methodological), 40(3): 364–372, 1978. doi: 10.1111/j.2517-6161.1978.tb01050.x. [DOI] [Google Scholar]
  39. Wand MP and Ormerod JT. Corrigendum: On semiparametric regression with o’sullivan penalised splines. Australian & New Zealand Journal of Statistics, 52(2):239–239, 2010. doi: 10.1111/j.1467-842X.2010.00578.x. [DOI] [Google Scholar]
  40. Wei Zhi and Li Hongzhe. A hidden spatial-temporal Markov random field model for network-based analysis of time course gene expression data. The Annals of Applied Statistics, 2(1):408 – 429, 2008. doi: 10.1214/07--AOAS145. [DOI] [Google Scholar]
  41. Winterer G, Enoch M-A, White K, Saylan Mete, Coppola Richard, and Goldman D. Eeg phenotype in alcoholism: Increased coherence in the depressive subtype. Acta psychiatrica Scandinavica, 108:51–60,08 2003. doi: 10.1034/j.1600-0447.2003.00060.x. [DOI] [PubMed] [Google Scholar]
  42. Wong Raymond KW, Li Yehua, and Zhu Zhengyuan. Partially linear functional additive models for multivariate functional data. Journal of the American Statistical Association, 114(525):406–418, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Yang Tianle and Suzuki Joe. The functional LiNGAM. In Proceedings of The 11th International Conference on Probabilistic Graphical Models, volume 186 of Proceedings of Machine Learning Research, pages 25–36. PMLR, 05–07 Oct 2022. URL https://proceedings.mlr.press/v186/yang22a.html. [Google Scholar]
  44. Zeng Yan, Cai Ruichu, Sun Fuchun, Huang Libo, and Hao Zhifeng. A survey on causal reinforcement learning, 2023.
  45. Zhang Xiao Lei, Begleiter Henri, Porjesz Bernice, Wang Wenyu, and Litke Ann. Event related potentials during object recognition tasks. Brain Research Bulletin, 38:531–538, 1995. [DOI] [PubMed] [Google Scholar]
  46. Zhou Fangting, He Kejun, and Ni Yang. Causal discovery with heterogeneous observational data. In Cussens James and Zhang Kun, editors, Uncertainty in Artificial Intelligence, Proceedings of the Thirty-Eighth Conference on Uncertainty in Artificial Intelligence, UAI 2022, 1–5 August 2022, Eindhoven, The Netherlands, volume 180 of Proceedings of Machine Learning Research, pages 2383–2393. PMLR, 2022a. URL https://proceedings.mlr.press/v180/zhou22a.html. [Google Scholar]
  47. Zhou Fangting, He Kejun, Wang Kunbo, Xu Yanxun, and Ni Yang. Functional bayesian networks for discovering causality from multivariate functional data, 2022b. [DOI] [PMC free article] [PubMed]
  48. Zhou Yidong, Bhattacharjee Satarupa, Carroll Cody, Chen Yaqing, Dai Xiongtao, Fan Jianing, Gajardo Alvaro, Hadjipantelis Pantelis Z., Han Kyunghee, Ji Hao, Zhu Changbo, Müller Hans-Georg, and Wang Jane-Ling. fdapace: Functional Data Analysis and Empirical Dynamics, 2022c. URL https://CRAN.R-project.org/package=fdapace. R package version 0.5.9.
  49. Zhu Hongxiao, Strawn Nate, and Dunson David B. Bayesian graphical models for multivariate functional data. Journal of Machine Learning Research, 17(204):1–27, 2016. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary

RESOURCES