Skip to main content
PLOS One logoLink to PLOS One
. 2022 Jun 27;17(6):e0270556. doi: 10.1371/journal.pone.0270556

Comparing the reliability of different ICA algorithms for fMRI analysis

Pengxu Wei 1,2, Ruixue Bao 3, Yubo Fan 2,*
Editor: Pew-Thian Yap4
PMCID: PMC9236259  PMID: 35759502

Abstract

Independent component analysis (ICA) has been shown to be a powerful blind source separation technique for analyzing functional magnetic resonance imaging (fMRI) data sets. ICA can extract independent spatial maps and their corresponding time courses from fMRI data without a priori specification of time courses. Some popular ICA algorithms such as Infomax or FastICA generate different results after repeated analysis from the same data volume, which is generally acknowledged as a drawback for ICA approaches. The reliability of some ICA algorithms has been explored by methods such as ICASSO and RAICAR (ranking and averaging independent component analysis by reproducibility). However, the exact algorithmic reliability of different ICA algorithms has not been examined and compared with each other. Here, the quality index generated with ICASSO and spatial correlation coefficients were used to examine the reliability of different ICA algorithms. The results demonstrated that Infomax running 10 times with ICASSO could generate consistent independent components from fMRI data sets.

Introduction

Independent component analysis (ICA) has been shown to be a powerful blind source separation technique for analyzing functional magnetic resonance imaging (fMRI). ICA can decompose a set of signal mixtures into a corresponding set of statistically independent component signals (source signals) [14]. The basic concept of ICA can be expressed as X = M ∙ C, where X is the observed data (i.e., the data matrix of fMRI signal), C is the component map (a matrix of voxel values), and M is the mixing matrix determining the time-varying contribution of each component map to the observed fMRI data. To estimate M and C simultaneously, ICA determines the unmixing matrix W (a permuted version of the inverse of the mixing matrix M) through iterative calculation. The component maps and corresponding time courses can be acquired using the following equation: C = W ∙ X [5, 6].

A number of ICA algorithms have been used for fMRI data analysis, including Infomax (Information Maximization), FastICA, ERICA (Equivalent Robust Independent Component Analysis), and SIMBEC (SIMultaneous Blind signal Extraction using Cumulants). ICA can extract independent spatial maps and their corresponding time courses from fMRI data without a priori specification of time courses. However, most ICA algorithms are based on minimization or maximization of an objective function and will find different local minima depending on where the “initial point” is. The randomness of the initialization that inherently exists in such ICA algorithms introduces randomness into the ICA decomposition and leads to different results when running ICA repeatedly. As a result, some popular ICA algorithms such as Infomax or FastICA will generate different results after repeated analysis from the same data volume, which is generally acknowledged as a drawback for ICA approaches [7, 8].

The reliability of some ICA algorithms has been explored by methods such as RAICAR (Ranking and Averaging Independent Component Analysis by Reproducibility) [7] and ICASSO (http://research.ics.aalto.fi/ica/icasso/) [8]. Such methods repeat the analysis multiple times with different random initializations and subsequently quantify the consistency of the outcomes [713]. ICASSO provides several indices including the quality index (Iq) to assess the reliability of independent components (ICs) after multiple runs. The Iq is introduced to be the main index reflecting the compactness and isolation of a cluster that contains ICs with high similarities from multiple runs. The Iq is computed as the difference between the average intra-cluster similarities and average extra-cluster similarities. RAICAR uses the spatial correlation coefficients (SCCs) to measure the component reproducibility. The reliability of the FastICA algorithm was assessed by using RAICAR or its extensions/variaions [710], and Correa et al. compared the results from Infomax, FastICA, JADE (Joint Approximate Diagonalization of Eigenmatrices), and EVD (Eigenvalue Decomposition)and found inconsistency between the results of EVD [14].

Some issues regarding the reliability of different ICA algorithms still remain. (1) The exact algorithmic reliability of different ICA algorithms is still unknown. Yang et al. assessed the algorithmic reliability of FastICA by using SCC [7]. Correa et al. applied ICASSO to evaluate the consistency of results from multiple runs of FastICA by using Infomax as a benchmark; however, they did not measure the reliability of Infomax [14]. No published reports examined the difference of different ICA algorithms with quantified index such as SCC (used by RAICAR) or Iq (provided by ICASSO). (2) ICA algorithms including AMUSE (Algorithm for Multiple Unkown Signal Extraction), JADE, ERICA (Extended Robust Independent Components Analysis), RADICAL (Robust, Accurate, Direct ICA aLgorithm), and SIMBEC produce identical results after each run (deterministic algorithms), whereas Infomax, FastICA, EVD, and COMBI (Combination of WASOBI and EFICA; WASOBI: Weights-Adjusted Second Order Blind Identification, EFICA: Efficient variant of algorithm FastICA) generate different results after repeated analysis (non-deterministic algorithms). How many times a non-deterministic algorithm should be run to acquire a reliable result remains unknown. Yang et al. [7] found that more than 20 times was not necessary, but only FastICA was tested.

ICASSO presents clusters after integrating the results of multiple runs. The estimates of an IC will be incorporated into a cluster if these estimates are consistent, and the number of estimates in a cluster is equivalent to the number of runs if the corresponding ICs are estimated consistently. Therefore, the number of estimates in each cluster can also be used to assess the reliability of an ICA algorithm.

This study aimed to solve three problems. (1) The Iq and the number of estimates in each cluster provided by ICASSO were used together to quantify the reliability of four non-deterministic algorithms (i.e., Infomax, FastICA, EVD, and COMBI). (2) Each of the four non-deterministic algorithms was repeated at different times with ICASSO (from 10 times to 100 times for each algorithm), and the SCC was used to determine how many times were needed to acquire ICA results with good reliability. (3) The SCC was used to quantify the consistency among the results of nine ICA algorithms (the four non-deterministic algorithms and the five deterministic algorithms, including AMUSE, JADE, ERICA, RADICAL, and SIMBEC) and determine the algorithm with the best reliability.

We explored real fMRI data involving sensory stimulation and motor task execution in this study.

Materials and methods

Subjects, stimulation/task paradigms, and image acquisition

The sensory stimulation data were acquired by using a 1.5T MRI scanner from a female patient (aged 18 years) with traumatic brain injury. The fMRI data were acquired with a gradient echo type echo planar imaging (GRE–EPI) sequence (Repetition Time = 3000 ms, Echo Time = 40 ms, flip angle = 90°, field of view = 240 mm × 240 mm, matrix size = 64 × 64, slice thickness = 5 mm, no gap, 23 contiguous axial slices). In one fMRI session, a block design protocol was applied with successive blocks alternating between rest and transcutaneous electrical nerve stimulation, starting with rest. There was a 24 s dummy period before MR data collection. Each block duration of rest or electrical stimulation was 30 s. The rest–stimulation cycle was repeated six times. Thus, the total duration of the fMRI session was 6 min 24 s. The electrical stimulation was performed at a rate of 2 Hz with surface electrodes. The cathode was placed on the space between the spinous processes of L2 and L3, whereas the anode was placed on the space between the spinous processes of L4 and L5. The stimulation was performed with a WQ-10D stimulator (Beijing Electronics Instrument Co., Ltd.). The intensity was set to a score of 2 rated on a 0–10 verbal numerical rating scale. T1-weighted images were also acquired to provide an anatomical reference.

The motor task data were acquired by using the same scanner from a healthy male subject (aged 30 years). The fMRI data were acquired with a GRE–EPI sequence (Repetition Time = 3000 ms, Echo Time = 40 ms, flip angle = 90°, field of view = 240 mm × 240 mm, matrix size = 64 × 64, slice thickness = 5 mm, no gap, 28 contiguous axial slices). T1-weighted images were also acquired. Two types of motor tasks were applied (i.e., imagined movements and motor execution task) in one fMRI session. The motor execution task consisted of 90° isometric dorsiflexion of the right foot. During the imagined movement task, the subject was required to imagine performing the same type of foot movement instead of actually executing it. The total duration of the fMRI session was 6 min. The experiment consisted of eight 21 s periods of baseline alternating with eight 21 s periods of motor tasks. The imagined movements were followed by the motor execution task in sequence. The beginning and ending of each task block were signaled with verbal commands “Ready, imaging, right, start” and “stop” or “Ready, move, right, start” and “stop.” The duration of each command was 3 s. An audio cue repeated every 3 s was broadcasted during each task block to prompt the subject to persistently focus on the task. The two fMRI data sets contained three types of status, i.e., sensory stimulation, imagined movements and motor execution task. The fMRI data sets can be found in the S1 Data.

The experiment was conducted with the approval from the institutional review board of the National Research Center for Rehabilitation Technical Aids, and written informed consents were obtained from the subjects. The procedure of the experiment was in accordance with the principles of the Declaration of Helsinki.

Performing ICA

The Statistical Parametric Mapping (SPM) software (https://www.fil.ion.ucl.ac.uk/spm) was used to perform motion correction, coregistration, and spatial smoothing with a Gaussian kernel of 10 mm full-width at half-maximum.

ICA was performed using the Group ICA of fMRI Toolbox (GIFT) software (http://mialab.mrn.org/software/gift/index.html). The number of ICs was estimated by using the minimum description length (MDL) criteria [15]. Except for the algorithm and the number of ICs, default settings/parameters defined by the GIFT software were used during analysis.

In group analysis containing a number of subjects, subjects in a group can be concatenated as a single ensemble by the GIFT software. In this study, there was a single ensemble for each subject since each data set contained only one subject. During data reduction steps, for one subject and one session, the data reduction actually would be disabled since the number of principal components extracted from the data is the same as the number of independent components.

Assessing the reliability of four non-deterministic ICA algorithms with ICASSO

The ICASSO toolbox embedded in the GIFT software was employed to run 10 times with different repetition number k (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, and 100) for each of the four non-deterministic algorithms.

The RandInit mode (algorithm starts with Randomizing different Initial values) was used. This mode in ICASSO uses the original data whereas the data will be resampled if the bootstrapping method is used. Additionally, the RandInit mode generates correlation coefficients with straightforward calculations whereas some extra normalization is necessary for bootstrapping [8]. When we run ICASSO 10 times by using the RandInit mode, the algorithm (e.g., Infomax) would run 10 times; each time the algorithm started with randomizing different initial conditions.

For a given algorithm, each of the 10 results of ICASSO contained N clusters; the number of N was determined by the MDL criteria. For an ICASSO result with a repetition number k, clusters presented by the ICASSO toolbox included estimated ICs from each run if these ICs presented high mutual similarities. The centrotype of a cluster holds the maximum sum of similarities to other points in the cluster and was used as the representation of this cluster by ICASSO. In ICASSO, the similarity between one pair of ICs (i and j) is quantified by the absolute value of their mutual correlation coefficients σi j. The clustering process is performed by using the distance between the two ICs. The distance is determined by transforming the similarity matrix into a dissimilarity (distance) matrix: di j = 1-σi j [8].

The reliability of the four non-deterministic algorithms was assessed with two approaches. The Iq was used to be the main index of reliability. First, for each algorithm, we compared the Iq values among results with different repetition number k to determine whether there is any difference when the repetition number k increases and to find the best k value with which a given algorithm could provide the most reliable result. Second, the number of ICs in each cluster will be equal to the repetition number k if an ICA algorithm is reliable, whereas an unreliable algorithm will generate clusters containing less number of ICs than the repetition number k. Thus, we compared how many clusters contained different numbers of ICs from the repetition number k for each of the four non-deterministic algorithms.

Afterward, we compared the Iq values between the most reliable results of the four non-deterministic algorithms to determine which ICA algorithm exhibited the highest reliability (i.e., the highest Iq value).

Comparing the reliability among four non-deterministic ICA algorithms with SCC

The SCCs between each pair of ICs from each pair of ICASSO results using the same algorithm were calculated by using MATLAB function corrcoef. The SCC was the Pearson correlation coefficient value between spatial maps of a pair of ICs. Afterward, the SCCs were used in two steps.

Step 1 was a process to select the maximum value. For a non-deterministic algorithm, suppose that the total number of ICs in each ICASSO result (when k = 10, 20, 30, 40, 50, 60, 70, 80, 90, and 100, the corresponding result was A, B, C, …, J, respectively) is i. Then, the j-th IC map (j = 1, 2, …, i) in ICASSO result A would have i SCC values with each IC map in ICASSO result B for this algorithm. The highest one among these i SCC values indicates the best spatial consistency/similarity. Thus, an IC with the highest SCC value in ICASSO result B can be considered to match the j-th IC in ICASSO result A. In this way, each pair of matched ICs in each pair of ICASSO results was determined. As a result, we acquired a list of SCCs for each pair of groups.

In Step 2, for each non-deterministic algorithm, the SCC values between the most reliable result (determined in the former section) and the other nine results were compared by using the Kruskal–Wallis test to measure the impact of repetition time k on the consistency of spatial maps.

Comparing the reliability among nine ICA algorithms with SCC

Finally, we compared the results from the nine algorithms with the SCC value as an index to investigate the spatial difference of results acquired with different algorithms. For the non-deterministic ICA algorithms Infomax, FastICA, EVD, and COMBI, the most reliable results were used to represent the most approximation of the source.

Results

Using the MDL criteria, the estimated numbers of ICs were 12 for the sensory stimulation data and 8 for the motor task data.

The most reliable result for the four non-deterministic algorithms when using the Iq value as an index

Sensory stimulation data

The median Iq values of different results for each of the four non-deterministic algorithms are shown in Fig 1. The fluctuation of Infomax results appeared to be much less than those of the other algorithms. For each of the four algorithms, the ICASSO repetition number k did not influence the Iq values (Table 1). For all of the four algorithms, the repetition number k with the highest median is displayed in Table 2. The ICASSO results with such number k were the most reliable ones for each non-deterministic algorithm.

Fig 1. Median Iq values of different results for non-deterministic ICA (sensory data).

Fig 1

The unit of the x-axis is the ICASSO repetition number k, and the unit of the y-axis is the Iq value. The results of COMBI presented clusters containing only one estimate when k = 10, 20, 40, and 70; under such conditions, the corresponding Iq value could not be calculated, which suggested very poor reliability.

Table 1. Repetition number k did not influence Iq values for non-deterministic ICA (sensory data).
Infomax FastICA EVD COMBI
Chi-sq 1.04 1.11 11.67 0.8406
P 0.9993 0.9992 0.2323 0.9744

For each non-deterministic algorithm, the Iq values of different repetition number k (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, and 100) did not present statistically significant difference (Kruskal–Wallis test). The result of the Kruskal–Wallis test of COMBI was based on six groups of Iq values because four results (when k = 10, 20, 40, and 70) of COMBI had clusters without the Iq value.

Table 2. ICASSO repetition number k with the highest median Iq of each algorithm (sensory data).
Infomax FastICA EVD COMBI
k 60 60 60 60
median 0.9843 0.9522 0.9667 0.9190

The result of COMBI was based on six groups of Iq values because four results (when k = 10, 20, 40, and 70) had clusters without the Iq value.

Motor task data

The median Iq values of different results for each of the four non-deterministic algorithms are shown in Fig 2. The fluctuation of results of Infomax also appeared to be much less than those of the other three algorithms. For each of the four algorithms, the ICASSO repetition number k did not influence the Iq values (Table 3). For all of the four algorithms, the repetition number k with the highest median is displayed in Table 4.

Fig 2. Median Iq values of different results for non-deterministic ICA (motor data).

Fig 2

The unit of the x-axis is the ICASSO repetition number k, and the unit of the y-axis is the Iq value.

Table 3. Repetition number k did not influence Iq values for non-deterministic ICA (motor data).
Infomax FastICA EVD COMBI
Chi-sq 15.37 8.83 1.99 9.14
P 0.0814 0.4535 0.9916 0.4241

For each non-deterministic algorithm, the Iq values of different repetition number k (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, and 100) did not present statistically significant difference (Kruskal–Wallis test).

Table 4. ICASSO repetition number k with the highest median Iq of each algorithm (motor data).
Infomax FastICA EVD COMBI
k 10 20 80 20
median 0.9954 0.9902 0.9242 0.9756

Taken together, Infomax presented better reliability than the other three non-deterministic algorithms.

Most reliable non-deterministic algorithm when using Iq values as the index

For each non-deterministic algorithm, the ICASSO result with the highest median Iq values was used as the most reliable one (i.e., those shown in Table 2 for the sensory data and Table 4 for the motor data). Here, these results were used to represent each algorithm and were then compared. The results of the Kruskal–Wallis test are shown in Table 5.

Table 5. Differences among the most reliable results of each non-deterministic algorithm.

Sensory data Motor data
Chi-sq 5.0128 22.0256
P 0.1709 0.00006

For both sensory and motor data, Infomax generated higher median Iq values than other algorithms although the difference was not statistically significant (p = 0.1709, Table 5). For the motor data, the median Iq value of the most reliable results from EVD (k = 80) was significantly lower than those of the other three non-deterministic algorithms. The result of COMBI was significantly lower than those of Infomax and FastICA when comparing the results from Infomax, FastICA, and COMBI (Chi-sq 15.7650, p = 0.0003773, the medians shown in Table 4). The results of Infomax and FastICA were not statistically different (Chi-sq 0.8934, p = 0.3446).

Thus, Infomax presented better reliability than the other three non-deterministic algorithms.

Assessing the reliability using the number of clusters containing different numbers of ICs from ICASSO repetition number k as an index

Sensory stimulation data

For Infomax, all clusters of each ICASSO result contained the same number as the repetition number k; for FastICA, each ICASSO result contained at least two clusters presenting different numbers of ICs from ICASSO repetition number k when k > 30, and the total number of such clusters was 21; for EVD, the ICASSO results contained two clusters presenting different numbers of ICs from the repetition number k when k = 80; for COMBI, each ICASSO result contained at least two clusters presenting different numbers of ICs from the repetition number k, and the total number of such clusters was 48.

Motor task data

For Infomax, all clusters of each ICASSO result contained the same number as the repetition number k; for FastICA, when k = 50, 70, and 80, each ICASSO result contained two clusters presenting different numbers of ICs from the repetition number k, and the total number of such clusters was 6; for EVD, when k = 90, the ICASSO results contained two clusters presenting different IC numbers from the repetition number k; for COMBI, when k > 50, each ICASSO result contained at least two clusters presenting different numbers of ICs from the repetition number k, and the total number of such clusters was 13.

Taken together, Infomax exhibited the best reliability when using the number of clusters containing IC numbers different from the ICASSO repetition number k as an index.

Difference in SCC values between the most reliable results and the other results for each non-deterministic algorithm

For each algorithm, no statistical difference in SCC values was found between the most reliable ICASSO results and each of the other nine results for either sensory data (Table 6) or motor data (Table 7).

Table 6. Differences in SCC values between the most reliable ICASSO results and the other nine results (sensory data).

Infomax FastICA EVD COMBI
Chi-sq 2.3734 0.7240 6.1818 5.0999
P 0.9674 0.9995 0.6269 0.7468

For each non-deterministic algorithm, ICASSO was run k times (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, 100) to acquire 10 results. The most reliable ICASSO results (shown in Table 2 for the sensory data) and the other results were compared by using the Kruskal–Wallis test.

Table 7. Differences in SCC values between the most reliable ICASSO results and the other nine results (motor data).

Infomax FastICA EVD COMBI
Chi-sq 3.2546 4.0372 7.0266 2.9837
P 0.9174 0.8537 0.5338 0.9354

For each non-deterministic algorithm, ICASSO was run k times (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, 100) to acquire 10 results. The most reliable ICASSO results (shown in Table 4 for the motor data) and the other results were compared by using the Kruskal–Wallis test.

The range of SCC values for each non-deterministic algorithm is presented in Tables 8 and 9. Note for the sensory data, FastICA and COMBI had very low SCC values, indicating poor spatial reliability.

Table 8. Range of SCC values between the most reliable ICASSO results and the other nine results (sensory data).

Infomax FastICA EVD COMBI
Max 1 0.999999 0.999357 1.000000
Min 0.969351 0.005471 0.994896 0.441637
Number of clusters with SCC<0.9 0 5 0 7

Table 9. Range of SCC values between the most reliable ICASSO results and the other nine results (motor data).

Infomax FastICA EVD COMBI
Max 1 1 1 1
Min 0.998238 0.997670 0.998731 0.998982
Number of clusters with SCC<0.9 0 0 0 0

For the sensory data, if the SCC values ≤ 0.88 (found in the results of COMBI), there would be an IC (in the corresponding ICASSO result) that matched two IC maps of the most reliable result (i.e., presenting similar SCC values), and there was another IC (in the corresponding ICASSO result) that did not match any IC in the most reliable result. In other words, when compared with the most reliable result, the other nine results with SCC ≤ 0.88 indicated unreliable performance. Thus, we set 0.88 as a threshold to numbers of such SCC values in Tables 8 and 9 (using 0.9 referring to [7]).

Taken together, Infomax exhibited the best reliability.

Comparing SCC values among the nine ICA algorithms

In previous sections, Infomax always presented better reliability than other non-deterministic algorithms. Infomax is a widely used ICA algorithm. It has been proven to be quite reliable for fMRI data analysis [6, 14, 16] and is a well-performing algorithm for other data types [17, 18]. Thus, we used ICASSO results of Infomax as a reference to compare the performance of other ICA algorithms against Infomax. For either sensory data or motor data, SCC values were computed between Infomax (the most reliable ICASSO result) and the results of the other eight ICA algorithms. For FastICA, EVD, and COMBI, the most reliable ICASSO results were used.

For both sensory data and motor data, only the results of FastICA exhibited good spatial consistency with the results of Infomax when considering median and minimum SCC values (Tables 10 and 11). Other algorithms had very low minimum SCC values in the results of either sensory or motor data, or both of them.

Table 10. SCC values between the most reliable Infomax results and the other nine results (sensory data).

AMUSE ERICA JADE RADICAL SIMBEC FastICA EVD COMBI
Median 0.708041 0.656005 0.95573 0.910922 0.685752 0.9972918 0.51767 0.991084
Max 0.916648 0.992157 0.998284 0.996636 0.992512 0.9999235 0.819438 0.999003
Min 0.519998 0.279998 0.096291 0.033701 0.186971 0.9786125 0.244201 0.018344

The results of FastICA (the most reliable results) presented higher SCC values than other algorithms.

Table 11. SCC values between the most reliable Infomax results and the other nine results (motor data).

AMUSE ERICA JADE RADICAL SIMBEC FastICA EVD COMBI
Median 0.801476 0.75061 0.94187 0.98748 0.739344 0.998879 0.623538 0.99123
Max 0.977052 0.939287 0.999795 0.998379 0.992799 0.9998937 0.97587 0.999588
Min 0.652854 0.476254 0.669816 0.723642 0.491713 0.963321 0.243115 0.88953

If there was an SCC value less than 0.669816, there would be at least one “unmatched” IC in the result of such algorithm, which means that this IC could not exclusively match any IC map in the results of Infomax and thus suggests a poor spatial consistency between the results of the algorithm and Infomax.

Discussion

Whichever index was used (Iq values, cluster numbers, and SCC values), Infomax always presented better reliability than other non-deterministic algorithms (FastICA, EVD, and COMBI). Infomax also exhibited better consistency than the five deterministic algorithms (AMUSE, JADE, ERICA, RADICAL, and SIMBEC) when using SCC as the index. Compared with Infomax, FastICA presented weaker consistency with some unreliable separated ICs. Other algorithms exhibited poor consistency under some conditions.

When running Infomax with ICASSO, repetition times > 10 did not lead to significant differences between results, indicating that 10 times can be used to acquire reliable ICs when using this algorithm. In additional to the high reliability of Infomax, this finding can also be attributed to the algorithm of ICASSO that generates centrotype of each cluster as the representative. The centrotype is calculated as the maximum sum of similarities to other points in a cluster [8], and running Infomax with ICASSO can therefore generate consistent results after a few repetition times.

When using ICA to explore cerebral networks, especially for conditions with low signal-to-noise ratio such as resting state [19], consistent ICA results generated by using a reliable algorithm such as Infomax can lead to higher study efficiency. Otherwise, variations of ICA results from different runs enhance the difficulty in explaining acquired ICs.

Our results demonstrate that Infomax running 10 times with ICASSO can generate consistent ICs from fMRI data sets. This finding provides an easily accessible approach to generate reliable ICs. These ICs can be used to develop advanced analyses such as personalized brain networks [20].

The results also demonstrate that the algorithms other than Infomax produce more or less unreliable separated ICs. Therefore, caution needs to be taken in explaining ICA results when the consistency of ICs cannot be assured.

The SCC was the correlation coefficient value between two IC spatial maps, that is, two spatial matrices. When comparing two groups of ICs A and B, the correlation coefficient value between each IC in group A and each IC in group B was calculated. The best-matched ICs must be a pair presenting the highest correlation value. Such a pair was unique since the corresponding correlation value was the highest one, and we did not find an IC in one group presenting two equally highest correlation values with two different ICs in the other group. We used the correlation coefficient value to compare different conditions, and the correlation values can be transformed into normal variables by using the Fisher-Z transform. Suppose that the population of correlation coefficient values is ρ, and the sample of correlation coefficients is r (the sample is a part of the population). When the Fisher-Z transformation is applied to the sample r, the sampling distribution of the transformed variable is approximately normal. Without the Fisher transformation, the variance of r (between two variables X and Y) grows smaller as |ρ| gets closer to 1 and thus is not normally distributed. In this study, however, for each comparison between two groups of SCCs, the normal distribution of the transformed SCC values in each group cannot be assured. The reason is that the SCC values (r values) come from different populations. For example, the SCC value between j-th IC in group A and k-th IC in group B can be Fisher-Z transformed, which leads to a normal distribution of the transformed SCC value. Such a normal distribution means that the transformed r values meet normal distribution for the two populations where two samples (j-th IC in group A and k-th IC in group B) are drawn. But now we have other SCC values between other pairs of ICs (e.g., m-th IC in group A and n-th IC in group B). These are samples from other populations. Therefore, the transformed r values may not be normally distributed when put into a group. As the solution, we used the Kruskal–Wallis test in Tables 6 and 7 since the Kruskal–Wallis test is a nonparametric test and does not assume the normality of the measurement variable.

This study used fMRI data from two subjects. These fMRI data sets contained three types of status, i.e., sensory stimulation, imagined movements, and motor execution task. Single-subject ICA analysis is the foundation of group-level analysis. However, results from two single-subject analyses may not well represent the performance of an ICA algorithm for all types of data. Whether the reliability of Infomax is superior to other algorithms needs to be further proved with other types of data.

Supporting information

S1 Data

(RAR)

Acknowledgments

We appreciate the suggestions of reviewers and the clarification of the origin of the word ICASSO by Prof. Aapo Hyvärinen (that is, ICASSO is not an acronym but evolved from the acronym ICA and the name of the artist Picasso).

Data Availability

All relevant data are within the manuscript and its Supporting Information files.

Funding Statement

This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700 by WP and ZL), the National Natural Science Foundation of China (Grant No. 81972160 by WP), and the Beijing Natural Science Foundation (Grant No. 17L20019 by WP). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Xu L. Independent Component Analysis and Extensions with Noise and Time: A Bayesian Ying-Yang Learning Perspective. 2003; 1(1): 1–52. [Google Scholar]
  • 2.Pontifex MB, Gwizdala KL, Parks AC, Billinger M, Brunner C. Variability of ICA decomposition may impact EEG signals when used to remove eyeblink artifacts. Psychophysiology. 2017; 54:386–98. doi: 10.1111/psyp.12804 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Grandchamp R, Braboszcz C, Makeig S, Delorme A. Stability of ICA decomposition across within-subject EEG datasets. Annu Int Conf IEEE Eng Med Biol Soc. 2012; 2012:6735–9. doi: 10.1109/EMBC.2012.6347540 [DOI] [PubMed] [Google Scholar]
  • 4.Risk BB, Matteson DS, Ruppert D, Eloyan A, Caffo BS. An evaluation of independent component analyses with an application to resting-state fMRI. Biometrics. 2014; 70:224–36. doi: 10.1111/biom.12111 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Chen H, Yao D, Zhuo Y, Chen L. Analysis of fMRI data by blind separation of data in a tiny spatial domain into independent temporal component. Brain Topogr. 2003; 15:223–32. doi: 10.1023/a:1023958024689 [DOI] [PubMed] [Google Scholar]
  • 6.Esposito F, Formisano E, Seifritz E, Goebel R, Morrone R, Tedeschi G, et al. Spatial independent component analysis of functional MRI time-series: to what extent do results depend on the algorithm used. Hum Brain Mapp. 2002; 16:146–57. doi: 10.1002/hbm.10034 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Yang Z, LaConte S, Weng X, Hu X. Ranking and averaging independent component analysis by reproducibility (RAICAR). Hum Brain Mapp. 2008; 29:711–25. doi: 10.1002/hbm.20432 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Himberg J, Hyvarinen A, Esposito F. Validating the independent components of neuroimaging time series via clustering and visualization. Neuroimage. 2004; 22:1214–22. doi: 10.1016/j.neuroimage.2004.03.027 [DOI] [PubMed] [Google Scholar]
  • 9.Pendse GV, Borsook D, Becerra L. A simple and objective method for reproducible resting state network (RSN) detection in fMRI. PLoS One. 2011; 6:e27594. doi: 10.1371/journal.pone.0027594 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Yang Z, Zuo XN, Wang P, Li Z, LaConte SM, Bandettini PA, et al. Generalized RAICAR: discover homogeneous subject (sub)groups by reproducibility of their intrinsic connectivity networks. Neuroimage. 2012; 63:403–14. doi: 10.1016/j.neuroimage.2012.06.060 [DOI] [PubMed] [Google Scholar]
  • 11.Chai R, Tran Y, Ling SH, Craig A, Nguyen HT. Combining ICA Clustering and Power Spectral Density for Feature Extraction of Mental Fatigue of Spinal Cord Injury Patients. Annu Int Conf IEEE Eng Med Biol Soc. 2019; 2019:530–3. doi: 10.1109/EMBC.2019.8857137 [DOI] [PubMed] [Google Scholar]
  • 12.Zhou C, Gao T, Guo T, Wu J, Guan X, Zhou W, et al. Structural Covariance Network Disruption and Functional Compensation in Parkinson’s Disease. Front Aging Neurosci. 2020; 12:199. doi: 10.3389/fnagi.2020.00199 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Nenert R, Allendorfer JB, Szaflarski JP. A model for visual memory encoding. PLoS One. 2014; 9:e107761. doi: 10.1371/journal.pone.0107761 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Correa N, Adali T, Calhoun VD. Performance of blind source separation algorithms for fMRI analysis using a group ICA method. Magn Reson Imaging. 2007; 25:684–94. doi: 10.1016/j.mri.2006.10.017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Rissanen J. A Universal Prior for Integers and Estimation by Minimum Description Length. The Annals of Statistics. 1982; 11:416–31. [Google Scholar]
  • 16.Calhoun V, Pearlson G, Adali T. Independent Component Analysis Applied to fMRI Data: A Generative Model for Validating Results. The Journal of VLSI Signal Processing. 2004; 37:281–91. [Google Scholar]
  • 17.Tresch MC, Cheung VC, d’Avella A. Matrix factorization algorithms for the identification of muscle synergies: evaluation on simulated and experimental data sets. J Neurophysiol. 2006; 95:2199–212. doi: 10.1152/jn.00222.2005 [DOI] [PubMed] [Google Scholar]
  • 18.Yang Q, Logan D, Giszter SF. Motor primitives are determined in early development and are then robustly conserved into adulthood. Proc Natl Acad Sci U S A. 2019; 116:12025–34. doi: 10.1073/pnas.1821455116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Roquet DR, Pham BT, Foucher JR. Manual selection of spontaneous activity maps derived from independent component analysis: criteria and inter-rater reliability study. J Neurosci Methods. 2014; 223:30–4. doi: 10.1016/j.jneumeth.2013.11.014 [DOI] [PubMed] [Google Scholar]
  • 20.Mejia AF, Nebel MB, Wang Y, Caffo BS, Guo Y. Template Independent Component Analysis: Targeted and Reliable Estimation of Subject-level Brain Networks using Big Data Population Priors. J Am Stat Assoc. 2020; 115:1151–77. doi: 10.1080/01621459.2019.1679638 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Kewei Chen

25 Jan 2022

PONE-D-21-19322Comparing the reliability of different ICA algorithms for fMRI analysisPLOS ONE

Dear Dr. Wei, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Sorry for the long review process so far, and we had only one reviewer's feedback, the experts comments are comprehensive and in-depth. I feel it is fair to move forward with you starting the revision process and I personally went over the MS couple time and concur with the reviewer.

Please submit your revised manuscript by Mar 11 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Kewei Chen, Ph.D

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. 

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

3. Thank you for stating the following in the Acknowledgments Section of your manuscript: 

"This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700), the National Natural Science Foundation of China (Grant No. 81972160), and the Beijing Natural Science Foundation (Grant No. 17L20019)."

We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. 

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows: 

"This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700 by WP and ZL), the National Natural Science Foundation of China (Grant No. 81972160 by WP), and the Beijing Natural Science Foundation (Grant No. 17L20019 by WP).

The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript."

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

4. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access.

We will update your Data Availability statement to reflect the information you provide in your cover letter.

5. Please include your full ethics statement in the ‘Methods’ section of your manuscript file. In your statement, please include the full name of the IRB or ethics committee who approved or waived your study, as well as whether or not you obtained informed written or verbal consent. If consent was waived for your study, please include this information in your statement as well. 

6. Please include a copy of Table S8, S9, S2 and S4 which you refer to in your text on pages 6, 11 and 12.

7. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information. 

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This is an interesting article. There is real value. However, I have several concerns regarding details missing, and technical questions that are unanswered in the article in its present form. The article is generally well written, but all acronyms are not fully spelled out on first use, and software packages and links are not always provided in the methods. The authors would have made both review and their revision tasks easier if they had added line numbers in the draft.

I will detail the various data and methodology concerns in detail below.

1. ACRONYMS must be spelled out on first use and links to software and citations provided. The omissions are numerous. A few (not complete list) MDL, AMUSE, JADE, ERICA and surprisingly, even ICASSO which is a pivotal method here.

2. Only two human subject data sets are used in this paper. This might be acceptable if we were convinced that the data was sufficiently rich, but this information is missing. The reader needs to know fully data set sizes: numbers of imaging sessions, full number of tasks, full number of stimuli in tasks etc. The richness of data is a factor in ICA outcomes so these aspects MUST be reported. Similarly, the method of data setup for ICA needs better description. Was data analyzed in each subject ultimately as a single concatenated ensemble (and if so, of what matrix dimensions), or as a number of subset ensembles (and if so, of what dimensions, and how many such ensembles). Because only two individuals data were used, there remains the unfortunate possibility that there are structural individual-based idiosyncracies in the data here that favor Infomax. I do not believe this likely, but it cannot be fully discounted given the data. Other research applying ICA or testing dimensionality separation methods including ICA have used larger numbers of individuals, or/and artificial data generation under constraints to explore comparisons, or reliability. Examples in motor field are Tresch, D'Avella, Cheung, papers in J Neurophysiology for synergy separation analyses, and Yang, Logan and Giszter in PNAS exploring SCC-like measures on motor synergy outcomes across individual animals. An aspect not explored by the authors of this paper (and given lack of detail provided, this may or may not be possible in their data), that might be used to enrich the analysis they perform is bootstrapping or jack-knifing the data sets used, in addition to subsequent multiple iterations of ICA methods on each subjects subset data. At the very least, there needs to be discussion of the data limits in this paper resulting from only 2 subjects and only 1 subject per type of experiment, and the caveats resulting, i.e. possibility of individual idiosyncracy of data favoring specific ICA methods.

3. The SCC method as described in section 2.4 is unclear. It is possible to do this in two ways - correlate the individual IC spatial components picking best correlations (where pairings of and IC with other sets may not be unique- same IC best in two or more correlations), or correlate the spatial matrix (e.g., in MATLAB using matperm, matcorr) and then IC correlations will be unique, based on the matrix permutation use in the unique IC matchings. The authors used the former, non-unique method I believe, but this might have biased results. If they did, with the possibility of non-unique matching of ICs, then I think the reader needs to know if there were non-unique best IC spatial correlations in each method, and if so how many. This might be an additional metric on ICA algorithm quality.

4. Finally, in relation to the SCC statistics used: Correlation, especially here, is non-gaussian if used directly. This may present difficulties in interpreting the statistics in tables 6 and 7 that could be avoided. Using the Fisher-Z transform (see Yang, Logan, Giszter, PNAS 2019 for example with Infomax ICA data stats) the correlations are transformed to normal variables, improving the interpretation of parametric methods.

5. Equations would be helpful throughout the methods if very clearly written. The goal is reproducibility of methods.

6. The data sharing statement is insufficient, for this study an anonymized repository should be chosen - there are many.

This paper is a solid contribution, but marred by the omissions in detail, and possibly improved by attention to the technical points noted.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Jun 27;17(6):e0270556. doi: 10.1371/journal.pone.0270556.r002

Author response to Decision Letter 0


3 Feb 2022

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match.

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

3. Thank you for stating the following in the Acknowledgments Section of your manuscript:

"This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700), the National Natural Science Foundation of China (Grant No. 81972160), and the Beijing Natural Science Foundation (Grant No. 17L20019)."

We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

"This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700 by WP and ZL), the National Natural Science Foundation of China (Grant No. 81972160 by WP), and the Beijing Natural Science Foundation (Grant No. 17L20019 by WP).

The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript."

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

4. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability.

Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized.

Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access.

We will update your Data Availability statement to reflect the information you provide in your cover letter.

5. Please include your full ethics statement in the ‘Methods’ section of your manuscript file. In your statement, please include the full name of the IRB or ethics committee who approved or waived your study, as well as whether or not you obtained informed written or verbal consent. If consent was waived for your study, please include this information in your statement as well.

6. Please include a copy of Table S8, S9, S2 and S4 which you refer to in your text on pages 6, 11 and 12.

7. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

We prepared the new version on the basis of “Journal Requirements” mentioned in the Decision letter:

1. We referred to the The PLOS ONE style templates.

2. The correct grant numbers for the Funding are those in the Acknowledgments Section of the first version. That is, "This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700), the National Natural Science Foundation of China (Grant No. 81972160), and the Beijing Natural Science Foundation (Grant No. 17L20019)."

3. We have removed any funding-related text from the manuscript. Please move the following contents to the Funding Statement: "This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700), the National Natural Science Foundation of China (Grant No. 81972160), and the Beijing Natural Science Foundation (Grant No. 17L20019)."

4. The minimal data set has been included in the Supporting information.

5. The information on IRB has been added.

6. Table S8, S9, S2 and S4 should be Table 8, 9, 2 and 4. We have corrected the error in the revised version.

7. We have included captions for the Supporting Information files at the end of the manuscript, and updated in-text citations.

=========================

Reviewer #1: This is an interesting article. There is real value. However, I have several concerns regarding details missing, and technical questions that are unanswered in the article in its present form. The article is generally well written, but all acronyms are not fully spelled out on first use, and software packages and links are not always provided in the methods. The authors would have made both review and their revision tasks easier if they had added line numbers in the draft.

Response:We really appreciate the comments and suggestions. The revised manuscript, we hope, can solve the raised issues and is clearly expressed.

I will detail the various data and methodology concerns in detail below.

1. ACRONYMS must be spelled out on first use and links to software and citations provided. The omissions are numerous. A few (not complete list) MDL, AMUSE, JADE, ERICA and surprisingly, even ICASSO which is a pivotal method here.

Response: We did not find the origin of the acronym “ICASSO” when we prepare the manuscript. The developers provided a link http://research.ics.aalto.fi/ica/icasso/ where two publications (http://research.ics.aalto.fi/ica/icasso/publications.shtml) can be found. However, neither the content on the website nor the publications introduced the origin of ICASSO. We also failed to find the origin of the acronym in other publications. To find the answer, we asked the developer Prof. Aapo Hyvärinen and learn: it is not a real acronym. It is a kind of a joke on the acronym "ICA" and the name of the artist Picasso. We think this information is useful and have supplemented it in the Acknowledgments since no answer can be found publicly elsewhere.

The origins of other acronyms, links to software, and citations have been added.

2. Only two human subject data sets are used in this paper. This might be acceptable if we were convinced that the data was sufficiently rich, but this information is missing. The reader needs to know fully data set sizes: numbers of imaging sessions, full number of tasks, full number of stimuli in tasks etc. The richness of data is a factor in ICA outcomes so these aspects MUST be reported. Similarly, the method of data setup for ICA needs better description. Was data analyzed in each subject ultimately as a single concatenated ensemble (and if so, of what matrix dimensions), or as a number of subset ensembles (and if so, of what dimensions, and how many such ensembles). Because only two individuals data were used, there remains the unfortunate possibility that there are structural individual-based idiosyncracies in the data here that favor Infomax. I do not believe this likely, but it cannot be fully discounted given the data. Other research applying ICA or testing dimensionality separation methods including ICA have used larger numbers of individuals, or/and artificial data generation under constraints to explore comparisons, or reliability. Examples in motor field are Tresch, D'Avella, Cheung, papers in J Neurophysiology for synergy separation analyses, and Yang, Logan and Giszter in PNAS exploring SCC-like measures on motor synergy outcomes across individual animals. An aspect not explored by the authors of this paper (and given lack of detail provided, this may or may not be possible in their data), that might be used to enrich the analysis they perform is bootstrapping or jack-knifing the data sets used, in addition to subsequent multiple iterations of ICA methods on each subjects subset data. At the very least, there needs to be discussion of the data limits in this paper resulting from only 2 subjects and only 1 subject per type of experiment, and the caveats resulting, i.e., possibility of individual idiosyncracy of data favoring specific ICA methods.

Response: The fMRI data sets contained three types of status, sensory stimulation, imagined movement and motor execution task. The imagined movement is actually one type of cognitive task. We used these data to represent several different types. Based on the suggestion, missing information such as numbers of imaging sessions has been added.

More details of the method of data setup for ICA have been added. Except the algorithm and the number of ICs, default settings/parameters defined by the GIFT software were used during analysis.

There was a single ensemble for each subject since each data set contained only one subject. During data reduction steps, for one subject one session, the data reduction actually would be disabled since the number of principal components extracted from the data is the same as the number of independent components, as introduced by the manual of the GIFT software, i.e., the matrix dimensions would not be changed. (In group analysis containing a number of subjects, subjects in a group can also be concatenated as a single ensemble by the GIFT software).

As you pointed out, some published studies such as Tresch, D'Avella, Cheung’s and Yang, Logan and Giszter’s have successfully performed ICA with Infomax. We suppose that many study groups have found some merits of Infomax empirically. The merit of this study is, on the aspect of reliability, to show the priority of Infomax and to uncover in which index Infomax is superior to other tested ICA algorithms in the single-subject level.

We used the RandInit mode (algorithm starts with Randomizing different Initial values) to run ICASSO. This information has been added in the manuscript. The RandInit mode was chosen because 1) the RandInit mode in ICASSO uses the original data whereas the data will be resampled in the bootstrapping method; 2) the RandInit mode generates correlation coefficients with straightforward calculations whereas some extra normalization is necessary for bootstrapping [J. Himberg, A. Hyvärinen and F. Esposito. NeuroImage 2004(3):1214-1222.]. If we run ICASSO 10 times, the algorithm (e.g., Infomax) will run 10 times; in each time, the algorithm starts with randomizing different initial conditions.

This study used fMRI data from two subjects. Discussions of this limitation have been added at the end of the manuscript. These fMRI data sets contained three types of status, i.e., sensory stimulation, imagined movements, and motor execution task. Single-subject ICA analysis is the foundation of group-level analysis. However, results from two single-subject analyses may not well represent the performance of an ICA algorithm for all types of data. Whether the reliability of Infomax is superior to other algorithms needs to be further proved with other types of data.

3. The SCC method as described in section 2.4 is unclear. It is possible to do this in two ways - correlate the individual IC spatial components picking best correlations (where pairings of and IC with other sets may not be unique- same IC best in two or more correlations), or correlate the spatial matrix (e.g., in MATLAB using matperm, matcorr) and then IC correlations will be unique, based on the matrix permutation use in the unique IC matchings. The authors used the former, non-unique method I believe, but this might have biased results. If they did, with the possibility of non-unique matching of ICs, then I think the reader needs to know if there were non-unique best IC spatial correlations in each method, and if so how many. This might be an additional metric on ICA algorithm quality.

Response: The SCC was the correlation coefficient value between two IC spatial maps, that is, two spatial matrices. When comparing two groups of ICs A and B, correlation coefficient value between each IC in group A and each IC in group B was calculated.

The best matched ICs must be a pair showing the highest correlation value. This best- matched pair is unique since the corresponding r value is the highest one; we did not find one IC in one group with two equal highest r values with two ICs in the other group.

In our data, we only found lower thresholds: 1) (section 3.4 Difference in SCC values between the most reliable results and the other results for each non-deterministic algorithm) For the sensory data, if the SCC values ≤ 0.88 (found in the results of COMBI), there would be an IC (in the corresponding ICASSO result) that matched two IC maps of the most reliable result (i.e., presenting similar SCC values), and there was another IC (in the corresponding ICASSO result) that did not match any IC in the most reliable result. In other words, when compared with the most reliable result, the other nine results with a SCC ≤ 0.88 indicated unreliable performance; 2) (section 3.5 Comparing SCC values among the nine ICA algorithms). If there was an SCC value less than 0.669816, there would be at least one “unmatched” IC in the result of such algorithm, which means that this IC could not exclusively match any IC map in the results of Infomax and thus suggests a poor spatial consistency between the results of the algorithm and Infomax. This information has been presented in the manuscript.

4. Finally, in relation to the SCC statistics used: Correlation, especially here, is non-gaussian if used directly. This may present difficulties in interpreting the statistics in tables 6 and 7 that could be avoided. Using the Fisher-Z transform (see Yang, Logan, Giszter, PNAS 2019 for example with Infomax ICA data stats) the correlations are transformed to normal variables, improving the interpretation of parametric methods.

Response: We used correlation coefficient value to compare different conditions. The correlation values can be transformed to normal variables with the Fisher-Z transform.

The population correlation coefficient is ρ, and the sample correlation coefficient is r. The sample is a part of the population.

When Fisher-Z the transformation is applied to the sample correlation coefficient r, the sampling distribution of the transformed variable is approximately normal. Without the Fisher transformation, the variance of r (between two variables X and Y) grows smaller as |ρ| gets closer to 1 and thus is not normally distributed. (referring to publications listed on https://en.wikipedia.org/wiki/Fisher_transformation)

In our case, however, for each comparison between two groups of SCC, normal distribution of transformed SCC values in each group cannot be assured. The reason is that the SCC values (r values) come from different populations: For example, a SCC value between j-th IC in group A and k-th IC in group B can be Fisher-Z transformed, which leads to a normal-distributed transformed SCC value. Such normal distribution means the transformed r values meet normal distribution for the two populations where two samples (j-th IC in group A and k-th IC in group B) are drawn. But now we have other SCC values between other pairs of ICs (e.g., m-th IC in group A and n-th IC in group B). These are samples from other populations. Therefore, the transformed r values may not be normally distributed when putted into a group. For a large data set it may be possible (normally distributed), but we have only a small sample size here (12 ICs for the sensory data and 8 ICs for motor data). As the solution, we used the Kruskal–Wallis test in Table 6 and 7. The Kruskal–Wallis test is a nonparametric test that does not rely on normal distribution.

5. Equations would be helpful throughout the methods if very clearly written. The goal is reproducibility of methods.

Response: We supplemented an equation in section 2.3:

In ICASSO, the similarity between one pair of ICs (i and j) is quantified by the absolute value of their mutual correlation coefficients �i j. The clustering process is performed by using distance between the two ICs. The distance is determined by transforming the similarity matrix into a dissimilarity (distance) matrix: di j=1-�i j.

There are two commonly used methods to transform the similarity matrix into a distance matrix: di j=1-�i j or di j=1/�i j. Therefore, the supplemented information is helpful if someone want to verify the results since the two methods generate different values. There are some long and complex equations in the cited literature. We do not present these equations in that they need detailed introductions and may interfere with the topic of the manuscript.

6. The data sharing statement is insufficient, for this study an anonymized repository should be chosen - there are many.

Response: We have added the fMRI data to the Supporting information so that other groups can verify the results.

This paper is a solid contribution, but marred by the omissions in detail, and possibly improved by attention to the technical points noted.

Response: Thanks very much again for the comments and suggestions. We hope the new version responds well to the concerns.

Attachment

Submitted filename: Response to Reviewers.docx

Decision Letter 1

Kewei Chen

16 Mar 2022

PONE-D-21-19322R1Comparing the reliability of different ICA algorithms for fMRI analysisPLOS ONE

Dear Dr. Wei,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Apr 30 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Kewei Chen, Ph.D

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This is a nice revision. Thank you for the work on the term ICASSO origin. This revision is coupled with a very responsive letter, but not all comments in the letter appear in the revised manuscript as I read it. Since readers may have the same concerns as reviewers it is crucial to include these in the manuscript, unless the full review history is also going to be published. In particular, the issue I could not find in the materials and methods:

That the component correlation method used (in section "Comparing the reliability among four ICA algorithms with SCC") did not in the present data produce any double use correlations must be stated, and that this was observed to be true, although this is not guaranteed by the method. I still wasn't clear if the authors were relying here on other code or directly checked this in their work, and this should be clearly stated.

Line 331 - probably, term 'well-performing' is better than 'well-performed' here

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Jun 27;17(6):e0270556. doi: 10.1371/journal.pone.0270556.r004

Author response to Decision Letter 1


28 Mar 2022

Thank you very much for the comments and suggestions. The explanations in the former letter have been added in the manuscript if the contents were not included in the former version.

In section "Comparing the reliability among four non-deterministic ICA algorithms with SCC” (and other sections where SCCs were calculated), the SCC was calculated by using MATLAB function corrcoef. We have added this information in the manuscript.

In this section, the SCC values are used in two steps.

Step 1 is a simple picking-the maximum-value process, introduced in the second paragraph of this section. This is not a statistical comparison but just picks the highest SCC value. As a result, we get a list of SCC values for each pair of groups.

Step 2 is a statistical comparison between different lists of SCCs by using the Kruskal–Wallis test, introduced in the last paragraph of this section.

Thus, this is not a double use of correlations. Step 1 provides values for statistical analysis in Step 2.

In the new version, we have added this information and highlighted supplemented words. We hope the contents will be more clearly expressed.

The word 'well-performed' has been changed to 'well-performing' based on your suggestion.

Attachment

Submitted filename: Response to Reviewers.docx

Decision Letter 2

Pew-Thian Yap

14 Jun 2022

Comparing the reliability of different ICA algorithms for fMRI analysis

PONE-D-21-19322R2

Dear Dr. Wei,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Pew-Thian Yap

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: No

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I think the use of Fisher Z transforms on your data would be necessary for parametric tests, and your argument about the distributions of i th and jth correlation pairings potentially having different distributions is not particularly good, and only meaningful insofar as your number of subjects is too small. The correlations have different normal distributions, as you correctly state, but the central limit theorem indicates a sum of normal distributions will tend to normal. The distribution problems you may have with normality in your study are primarily due to the very small subject sample size you have (2 subjects) meaning the distribution might be multimodal if variances are low for different correlations across your analysis runs, because not enough subject variations for distributions are contributing. It would be better to directly acknowledge this as the basis of the non parametric testing. Since you are choosing using non-parametric tests here, rather than parametric tests, the ordinal testing you used will be unaffected by the z transform and it is not needed. The best study would have 6-10 subjects per group and use z-transformed correlation data.

I suggest a note to this effect, but it is your choice.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

Acceptance letter

Pew-Thian Yap

17 Jun 2022

PONE-D-21-19322R2

Comparing the reliability of different ICA algorithms for fMRI analysis

Dear Dr. Wei:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Pew-Thian Yap

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Data

    (RAR)

    Attachment

    Submitted filename: Response to Reviewers.docx

    Attachment

    Submitted filename: Response to Reviewers.docx

    Data Availability Statement

    All relevant data are within the manuscript and its Supporting Information files.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES