Abstract
0.1. Background
In this age of big data, certain models require very large data stores in order to be informative and accurate. In many cases however, the data are stored in separate locations requiring data transfer between local sites which can cause various practical hurdles, such as privacy concerns or heavy network load. This is especially true for medical imaging data, which can be constrained due to the health insurance portability and accountability act (HIPAA) which provides security protocols for medical data. Medical imaging datasets can also contain many thousands or millions of features, requiring heavy network load.
0.2. New Method
Our research expands upon current decentralized classification research by implementing a new singleshot method for both neural networks and support vector machines. Our approach is to estimate the statistical distribution of the data at each local site and pass this information to the other local sites where each site resamples from the individual distributions and trains a model on both locally available data and the resampled data. The model for each local site produces its own accuracy value which are then averaged together to produce the global average accuracy.
0.3. Results
We show applications of our approach to handwritten digit classification as well as to multi-subject classification of brain imaging data collected from patients with schizophrenia and healthy controls. Overall, the results showed comparable classification accuracy to the centralized model with lower network load than multishot methods.
0.4. Comparison with Existing Methods
Many decentralized classifiers are multishot, requiring heavy network traffic. Our model attempts to alleviate this load while preserving prediction accuracy.
0.5. Conclusions
We show that our proposed approach performs comparably to a centralized approach while minimizing network traffic compared to multishot methods.
1. Introduction
Due to advances in analytic tools, researchers have the opportunity to assemble informative models given large datasets. Because of this, research institutes and hospitals are increasing collaborative efforts [Ming et al., 2017, Plis et al., 2016, Carter et al., 2015, Thompson et al., 2014]. The simplest and easiest method is for these institutes to share data, which typically involves an individual group downloading data from various sources and performing large, centralized analyses. Such an approach is limited in what data can be shared openly due to the health insurance portability and accountability act (HIPAA) or other regulatory concerns. More specifically, these institutes need to protect the personal information of patients which can be compromised when transferring or centralizing the data. While this can be addressed by anonymizing data, in some cases, notably when there are many local sites, this is not sufficient. Another issue is that large computational resources and extensive download time is typically required in a centralized approach, which is very problematic for neuroimaging datasets, which can contain hundreds or thousands of subjects, with many thousands or millions of features. Decentralized models may solve both of these issues by eliminating the requirement of data transfer.
There is a current body of various decentralized models [Gazula et al., 2018, Saha et al., 2017, Wojtalewicz et al., 2017, Baker et al., 2015], and more specifically, decentralized neural networks [Lewis et al., 2017] and support-vector machines (SVM)[Forero et al., 2010]. However, these models are multishot, meaning they pass statistical information many times during the training process, which can require a great deal of network traffic. The multishot neural network, or decentralized-data neural network (dDNN) [Lewis et al., 2017] requires heavy network traffic at least once every epoch, or one full iteration through the entire dataset during the training process. This is because the dDNN model passes all gradient information from local sites to a centralized location after every epoch, then calculates the average of these gradients, and passes the averaged gradients to the local sites. As neural networks can require many thousands of epochs, the overall network traffic would be unmanageable for neuroimaging data. This same problem occurs for multishot SVMs, which also require a high number of steps in which gradients are passed between local sites.
Algorithm 1.
Decentralized Data-sampled Neural Network
1: | procedure Decentralized Data-sampled Neural Network(NN, Sites) |
2: | NN ← New Neural Network Model |
3: | Sites ← array of Network pipes to local sites |
4: | for each i ∈Sites do |
5: | Gi = Gaussian Distribution of local data |
6: | //Calculate distribution per local data and pass to every other site |
7: | for each j ∈Sites\i do |
8: | Gi →j |
9: | end for |
10: | end for |
11: | //Initialize Models at Local Sites and Train |
12: | for each site ∈ Sites do |
13: | site ← NN |
14: | site → Train Model |
15: | end for |
16: | end procedure |
In this research, we attempt to mitigate these issues for certain classifiers by introducing a singleshot method. Singleshot methods require statistical information to be passed only once, either before or after the local models have been trained. In our case, statistical information is passed to the local sites and then each site trains separately. The statistical information is an estimated distribution of the local data, which is comprised of the per-feature mean and a covariance matrix of the features. We refer to this model as a decentralized distribution-sampled classifier (dDSC). We demonstrate the efficacy of the dDSC approach applied to both neural networks and SVMs on two datasets. We quantify the data at each local site by building local distributions using a Gaussian mixture model (GMM) and pass these distributions to the remaining local sites which will then be used in training models at the local sites. Each local site combines artificial data sampled from the given distributions with locally available data to train the models.
2. Methods
2.1. Multishot decentralized modalities
In the previous multishot models, dDNN and the consensus-based SVM, high-level statistical information (i.e. gradients) are passed between local sites many times during the training of the models. However, this requires a high traffic load and as the number of training iterations increases, the chance of network failure also increases. The dDNN aggregates the local gradients every iteration (or epoch through the data), averages the gradients, and passes these updated gradients to the local sites. The multishot SVM uses the alternating direction method of multipliers (ADMoM) to accumulate the updating parameters, or the model weights, [Forero et al., 2010]. These model weights are used, as in a non-decentralized model, to update the Lagrangian multipliers. This process is repeated as many times as necessary to complete the training.
2.2. Statistical inference models
Our approach for singleshot classifiers-dDSC gathers statistical information about the datasets rather than the models, as is the case with multishot algorithms, at the local sites and passes this information between the sites before the models are trained. We use a GMM to estimate the distribution of the local site data for each class. Although more sophisticated models like Dirichlet mixture models may perform better in density estimation, GMMs were chosen for their simplicity and could also act as a baseline model. Once the distribution is gathered from the model for each site, this distribution is passed to the other sites. The other sites then draw artificial samples from the remaining sites’ distributions and trains their own model on both locally available data as well as the artificial samples. This approach also shows a much smaller amount of network traffic, as the mixture model is transferred once. This is the case for both the neural network and SVM methods.
(1) |
(2) |
(3) |
(4) |
(5) |
2.3. Gaussian Mixture Model
A GMM uses the expectation-maximization (EM) algorithm to fit the data to a Gaussian distribution. The EM algorithm begins by generating a random Gaussian distribution (Equation 1) with randomized mean, μk and variance, σk for each class k. γik (Equation 2) is used to estimate the probability that a given data point, xi is within the distribution of a given class k. πk is the mixture weight for class k. The algorithm creates a random Gaussian distribution for each class, and then calculates γik for each data point and class pair. It then updates the class-wise μk (equation 4), σk (equation 5), and πk (equation 3) using γik. The data points are then reassigned to the k distributions over many iterations until the model converges.
GMMs use covariance matrices to store the variances of the features, which determine the contours of the distribution in an N-dimensional space. The covariance matrices can be represented in different ways, the most common being a full-covariance matrix; in which the variance of all pairs of features is stored in the matrix. Another representation of covariance matrices is a diagonal matrix, in which the variances of only individual features are stored in memory. This distinction is important for our work as it determines the network traffic of the model. When analyzing the MNIST dataset, we use a full covariance matrix as the feature set is relatively small compared to the sMRI dataset. However, in the case of the sMRI data, we use a diagonal matrix due to the sheer size of the feature space and to show that the dDSC also works with a very small subset of the distribution information, which reduces total network traffic.
2.4. Experiments
The experiments use two different datasets: the Mixed National Institute of Standards and Technology (MNIST) dataset of handwritten numbers [LeCun et al., 2010] (examples found in figure 2), and a set of real-world structural magnetic resonance images (sMRI) of schizophrenia patients and healthy subjects from an aggregated multisite dataset [Potkin et al., 2008, Gollub et al., 2013, Hanlon et al., 2011, Aine et al., 2017]. We tested the dDSC models’ performances on MNIST with three cases: the data is uniformly and randomly distributed across three sites, three sites have access to only certain classes, and the datasets are uniformly and randomly distributed across 20 sites. We also tested the models via the sMRI data in two cases: the data is uniformly distributed across four sites at random, and four sites have access to only certain classes. A full break-down of the experiments can be seen in table 1.
Figure 2:
Examples of the hand-drawn digits in the MNIST dataset.
Table 1:
The break-down of the experiments. The rows are the two different datasets, and the columns are the three experiments. Each cell explains the important details of an experiment for a given dataset.
MNIST Data | sMRI Data | Models | |
---|---|---|---|
Fully Randomized | -60,000 images -3 local sites -Data distributed uniformly at random to all sites |
-10-fold cross-validation -584 subjects -4 local sites -Data distributed uniformly at random to all sites |
NN, SVM |
Biased by Class Label | -53,869 images -3 sites, each with 3 digits -site 1: 0–2 -site 2: 4–6 -site 3: 7–8 |
-584 subjects -4 sites, each with either healthy controls or patients -sites 1 & 2: healthy -sites 3 & 4: patients |
NN, SVM |
Fully Random, with Many Local Sites | -60,000 images -20 sites -Data distributed uniformly at random to all sites |
N/A | NN |
2.5. MNIST Experiments
For all MNIST experiments, the images are vectorized from 28×28 pixel images into vectors of size 784. The labels are one-hot-encoded into a vector of size 10 for each image. For all experiments, we tested the models using the test set established by the dataset creators.
In the first experiment, we randomly select (without replacement) 20,000 images from the entire dataset of 60,000 images for each of the three sites. This process is used for both the dDNN and dDSC (NN as well as SVM) models. However, the centralized model will have only one site with access to all 60,000 images. All of the neural networks (centralized, decentralized and decentralized with density sampling) use the same architecture: two hidden layers of size 256, a learning rate of 0.001, and batch normalization is also used. The models are initialized with the Xavier initialization algorithm [Glorot and Bengio, 2010]). The models are run for 1000 epochs, and the accuracy of the MNIST test data is captured every 100 epochs. In the dDSC model, the data at each local site are sampled with a separate GMM process for each class, and the distributions are passed to every other site. The local sites then sample 20,000 images (synthetic data) from each of the incoming distributions. We validate the procedure by testing each site’s model on the MNIST testing set and averaging the accuracies across every site.
In the second MNIST experiment, we bias the per-site data by class label. One site had access to digits 0 through 2, the second site had digits 4–6, and the last site contained digits 7–9. As the number 3 is not included, there are only 53,869 total images. As with the first experiment, in the dDSC model each site sampled synthetic data from the other sites’ per-class distributions of the respective local datasets and the accuracies of the fitted model from the local sites are averaged.
The third experiment, which was only used to test the neural network version of the dDSC, uses 60,000 images as in the first experiment. The SVM was not tested with this experiment due to memory and time constraints. The data is processed the same way as the first experiment, and the models are also of the same architecture. However, the primary difference is that the data is separated into 20 local sites as opposed to 3. This means that there are a total of 3,000 images at each local site distributed uniformly at random. Then, as in the previous experiments, the accuracies between the three NN models are compared.
2.6. sMRI data demographics
The sMRI dataset is aggregated from three separate datasets: the Mind Clinical Imaging Consortium (MCIC) [Gollub et al., 2013], the Function Biomedical Informatics Research Network (fBIRN) [Potkin et al., 2008], and the Center for Biomedical Research Excellence (COBRE). In total, the three datasets consists of 584 subjects, of which, 269 are schizophrenia patients; the remaining subjects are healthy controls. For a more detailed breakdown of the demographic information, please see table 2.
Table 2:
sMRI Subject Demographics
MCIC | COBRE | FBIRN | |
---|---|---|---|
Schizophrenia Subjects | 101 | 93 | 75 |
Healthy Subjects | 119 | 97 | 99 |
Schizophrenia Male/Female Split | 74/27 | 76/17 | 63/12 |
Healthy Male/Female Split | 73/46 | 69/28 | 69/30 |
Total | 220 | 190 | 174 |
2.7. sMRI dataset experiments
The sMRI dataset is very large with over 58 thousand features. As such, with or without statistical inference, the network traffic would be problematic. Due to this, we use a diagonal matrix to store the estimated variances of the GMM. Diagonal matrices store fewer data points and greatly reduces the required network traffic as it requires N values, whereas full covariance matrices require N2 values. However, this loss of information may also have an impact on accuracy.
The sMRI dataset is analyzed using two separate experiments: one in which the entire dataset is randomized across all local sites, which is tested on both the neural network and SVM, and another in which the local site data are biased by class, which is tested on the neural networks only. For this dataset, both experiments used four sites, but due to the size of the dataset, 10-fold cross-validation is used to measure the efficacy of the model. For all experiments involving neural network, we use two hidden layers of 1000 nodes and 200 nodes respectively. The batches are normalized and the weight sets are initialized with the Xavier algorithm, and we used a learning rate of 0.01. The models are run for a total of 2000 epochs throughout the entire dataset and the accuracy is measured on the per-fold test data every 100 epochs.
In the first experiment, for the neural network, the three approaches (centralized NN, dDNN and dDSC) were tested the same way as in the MNIST experiment. All three approaches had access to both classes and the data is uniformly disbursed to the four sites at random. The dDSC SVM and centralized SVM were also tested and compared to each other.
In the second experiment, for the neural network only, we bias the per-site data by class label. The schizophrenia patients are evenly and randomly distributed between two of the local sites. The healthy controls are evenly and randomly split between the remaining two sites. No subject overlaps between the local sites, and each site had access to data with only one label; either schizophrenia or healthy. The distributions for each local dataset are then calculated and passed to the remaining local sites and each site trained a model on the sampled data and available local data. The local models are built with the same architecture, but since they are trained on different data they are not an exact equivalent to the dDNN model. This is measured by averaging the accuracies across all four local sites. The goal is to show the model’s robustness to extremely biased data.
3. Results
3.1. MNIST experiments
The first experiment of the neural network, in which the entire dataset is randomly distributed across all three sites, shows near identical accuracies between the dDNN and centralized approaches. Both converge at approximately 97.1% accuracy (Figure 3). The dDSC neural network slightly worse, but still quite well, approximately converging towards 96.4% accuracy.
Figure 3:
The MNIST data are randomly and evenly distributed between three sites for all three neural network approaches, a centralized neural network, the dDNN model, and the dDSC model. The accuracy and standard deviation per approach is outlined above. The dDNN accuracy is the average over all three sites.
In the second neural network experiment, the data are biased in such a way that each site had access to only three of the possible classes. Digit ‘3’ is removed so as to give each site an equal number of classes. From Figure 4, we see that, as in the previous experiment, the dDNN and centralized approaches are almost identical, converging toward 97.8%. The dDSC neural network approach is slightly less accurate, converging towards 95.5% accuracy. The discrepancy between the accuracies of the same procedures is most likely due to the impact from the exclusion of digit three. It appears that including digit three makes the MNIST problem slightly more difficult.
Figure 4:
Results comparing the Centralized, dDNN, and dDSC approaches when the local MNIST data are biased by class label such that each of the three sites has access to only three of the digits. The accuracy and standard deviation of the accuracy is reported for the dDSC model over all three sites. The standard deviation is shown as stems within the plot.
The third neural network experiment shows the limitations of the dDSC model. This model converges towards 92%. Whereas the centralized and DDNN are almost identical and converge toward approximately 96.4% (Figure 5). The goal of this experiment is to show a major limitation in the procedure; which is its poor performance with many local sites.
Figure 5:
The MNIST data are randomly and evenly distributed randomly and uniformly over 20 sites for all three approaches. The accuracy and standard deviation per approach is outlined above, including the average accuracy of the dDSC models over all sites.
The SVM and neural network dDSC models perform similarly, although slightly worse, than their centralized counter-parts in the case in which all of the data are distributed uniformly at random to all three sites. The average accuracy for the three site case for the dDSC SVM model is 91% (Figure 8). This is in contrast to the centralized method which had an accuracy of 91.5%. These accuracies reflect the differences between SVMs and neural networks when applied to the MNIST dataset [Deng, 2012].
Figure 8:
Results from the SVM experiments. The MNIST data (left) showed similar results between the centralized and dDSC models. The uniformly and randomly distributed experiment of the MNIST dataset is on the far left and the case in which the data is biased by class label is in the center. The dDSC also showed similar results compared to a centralized model when applied to the sMRI data (right).
For the case in which there are three sites and the data are biased by class label, the dDSC had an accuracy of 90.8%, while the centralized method had an accuracy of 91.5%.
3.2. sMRI experiments
For all sMRI experiments, a diagonal covariance matrix is assumed while estimating the parameters of the distribution. This would have an impact on the performance of a model, which can be seen in the results. However, a diagonal covariance matrix greatly reduces the required network traffic from O(N2 * L) to O(N * L). Although the accuracy does decline by a small factor, the multishot dDNN and SVM models require a much higher order of network traffic.
The first neural network experiment, in which the entire dataset is randomly distributed across all three sites, show near identical accuracies between the dDNN and centralized approaches. Both converge similarly: the dDNN method converges to about 72.1% accuracy (Figure 6), averaged across all 10 folds, and the centralized approach converges to about 72.9% accuracy. The dDSC converges toward 68.5% accuracy.
Figure 6:
Results comparing the 10-fold cross-validation accuracies between a Centralized neural network, dDNN, and dDSC approaches with randomly and uniformly distributed data from the sMRI dataset. Stems show standard deviation and the lines show the mean accuracies across the 10 folds for each approach.
When the sMRI data is modeled with an SVM, we uniformly and randomly distribute the data across three sites (i.e. the same method as the first neural network experiment). 10-fold cross validation is used, covering all data samples, and the data are uniformly distributed to four different sites at random. The mean of the accuracies across all 10 folds from the dDSC model is 67.5%, whereas the mean from the centralized method across all folds is 72% (Figure 8).
In the second neural network sMRI experiment of the dDSC model, the data is evenly distributed across four sites, but is biased in such a way that each site had access to only one of the possible classes. This means that two sites had access to only patients and the remaining two sites had access to controls. From Figure 7, we see that, as in the previous experiment, the dDNN and centralized approaches are almost identical, converging towards 72.8% accuracy. The dDSC approach converges towards 65.1% accuracy.
Figure 7:
Results comparing the 10-fold cross-validation accuracies between a centralized neural network, dDNN, and dDSC approaches with data from the sMRI dataset in which the data are biased by class label, such that each each site only has access to one of the two classes, either patients with schizophrenia and healthy controls. Stems show standard deviation and the lines show the mean accuracies across the 10 folds for each of the three approaches.
4. Discussion
Decentralized models have an important place in machine learning and data analytics and are based on the concept of distributed computation [Dijkstra, 1965, Dean et al., 2012]. Currently, decentralized modeling has focused on basic machine learning techniques [Gazula et al., 2018, Ming et al., 2017, Plis et al., 2016, Baker et al., 2015, Forero et al., 2010], with scant research on neural networks [Lewis et al., 2017]. Previous research on decentralized deep learning focused on eliminating significant differences in accuracy between decentralized neural networks and their centralized counterpart. However, this proved to be problematic as the size of the network traffic is unmanageable for even the fastest network, given a large enough dataset, such as fMRI scans for many subjects. As the research is geared towards biological data, consisting of very large datasets, this problem is especially important.
Mixture models have previously been used to enhance neural networks [Viroli and McLachlan, 2017], and the concept of generating new samples to improve neural network accuracy has also been explored [He et al., 2008, Goodfellow et al., 2014a]. We adapt these concepts to greatly improve decentralized neural networks and solve the problem of high network traffic with minimal impact on accuracy. GMMs, although simple, have a history of improving the accuracy of classification models [Chen et al., 2014, Wang and Jo, 2006].
4.1. Network Traffic Analysis
The importance of this work rests in the dDSC model’s ability to reduce total network traffic. Our asymptotic analysis of the models provides a method to quantify the total network load as a function of the magnitude of the input data. Asymptotic analysis is a widely used method to understand the speed or space constraints of algorithms [de Bruijn, 1959]. The asymptotics are computed as the big-O, or the upper bound on the total amount of space required with respect to a given input of size N. This is done by simply computing the upper bound on how many gradients the multishot methods must pass between sites and how often this transfer must occur. The asymptotics of our singleshot method are estimated as a function of the size of the distributions passed between sites, including the means of each component and the covariance matrices of the local data.
The dDNN model proves to be very network intensive, as all gradients are passed between sites for every iteration. In order to reduce the total number of iterations required, the dDNN would train on one epoch ever iteration, which is analogous to stochastic gradient descent. Given this, the asymptotic analysis of the network traffic is straight forward: O(W * I), where W is the total number of weights in the model, and I is the number of iterations the model requires to converge. Given that the first weight matrix is size N*w1, W ≥ N for any possible N. With complex datasets, the number of weights can be much higher than N, as even first weight set, wi, is a multiplicative combination of the number of input features and the number of hidden nodes, and the number of hidden nodes can be on the order of size N, and the subsequent weight set is a multiplicative combination of w1. This suggests an upper bound of O(N2). It also must be noted that this network traffic occurs for every local site twice; the first time when the local sites pass gradient information to the central server, and the second time when the central server passes the averaged gradients to the local sites. This network load occurs in every iteration through the local dataset, where each iteration can be small batches of data, or an entire epoch of the local datasets. This means that the total space complexity of the multishot neural network is O(N2 * I). As neural networks can require many hundreds or even thousands of iterations, it grows as complexity of the data and model increases.
The current decentralized SVM model calculates the support vector weights, vi for every site i and broadcasts vi to all other local sites in a given iteration. Every feature n has a corresponding vi weight value, meaning vi is of size N for every v ∈ V. With I iterations and N input features, the space complexity of an SVM is O(N * I * L). The dDSC model however, requires statistical information to be passed only once. When a full covariance matrix is used to estimate the distribution of the GMM, the model requires space for the variance of each feature pair, for each possible class. Meaning, at most, O(N2 * L) bytes of information are passed to each other site for each site with a number of labels L. However, in the case of a diagonal matrix, which is the method used to analyze the sMRI data, the variances and means of only the individual features are stored in memory for each label, meaning this method requires O(N * L) traffic across a network.
Given this analysis, it is clear why the dDSC model is effective at reducing the upper bound of the total network traffic, as well as greatly reducing the number of times information is passed between sites. Our analysis showed that our singleshot method does decrease network traffic by at least one order of magnitude. Beyond the reduction in total network traffic, it is important to note that the dDSC model reduces the total number of network broadcasts required during training. With multiple network broadcasts, there are more fault points, as any interruption in the network connection halts training, and the probability of interruption increases as the number of required broadcasts increases. This problem is entirely mitigated with a singleshot method.
4.2. Limitations and Future Work
The dDSC model appears to be valid for many criteria, however it does have limitations. As seen in Figure 5, the accuracy of the dDSC model decreases with many more local sites. The local distributions appear to be less accurate due to a lack of sample data. This result is what would be expected for small enough sample sizes. However, there is a body of literature showing how to improve mixture model accuracy given small datasets [Liu et al., 2008]. Although, it should be noted that small datasets will always be a limiting factor for machine learning, and a certain amount of error is expected with small datasets.
Although the accuracy was similar between the dDSC, dDNN, and centralized modalities, there was a large discrepancy between the performances of the two datasets. MNIST was selected primarily because it is a relatively easy dataset to model. Although neural networks and SVMs perform similarly when compared [Liu et al., 2003], our SVM was not finely tuned which may explain differences in accuracy between our SVM and neural network models. However, schizophrenia classification from sMRI data is vastly more difficult, both in feature size and fewer available samples. Previous research has also shown prediction accuracy similar to what we achieved [Gould et al., 2014].
We suggest that future researchers could develop models that are more accurate in general, or at least in the case of many local sites. This could include testing different mixture model or methods to generate artificial data to improve the accuracy of local distributions for small datasets [Fei-Fei et al., 2007]. Another avenue for future work would be to use the Dirichlet process to generate new samples using an infinite mixture model. Future research could also focus on using generative adversarial networks (GAN) [Goodfellow et al., 2014b] to train on local data and produce artificial samples for other local sites. Finally, a more basic avenue for continued research would be to use our method for different neural network architectures or different machine learning models. This model shows that statistical inference works as a way to model a classifier in a decentralized setting. However, statistical inference could also be used for other decentralized methods.
5. Conclusion
This work expands on the current body of decentralized classification models with a technique to improve network efficiency by estimating distributions of the local data and sampling from these distributions at every other local site. We accomplished this with a Gaussian mixture model to produce artificial samples from the local data. The dDSC neural network produces accuracy comparable to the dDNN model, but with greatly reduced network traffic. This research also concludes that the dDSC SVM performs on par with a centralized SVM model while also reducing network traffic compared to the original decentralized SVM when using a diagonalized matrix. The results from the experiments show a promising expansion to decentralized neural networks.
Figure 1:
The three paradigms from left to right: dDSC, DDNN, and a centralized model. In the dDSC model, for every site i, every other site calculates the distribution of the local data and passes the distribution (in the form of a matrix) to site i. Site i then samples data from these distributions and uses this artificial data as well as the local data to train its own model. In the dDNN model, each local site trains its own model and the available local data and passes the gradient data to a centralized server. The centralized server then averages the local gradients and passes this average to the local sites to train the local models. The centralized paradigm uses all possible data in a single model at a central site.
0.6. Highlights.
A novel yet simple approach to decentralized classification
Reduces total network load compared to current multishot algorithms
Maintains a prediction accuracy comparable to the centralized approach
References
- Aine CJ, Bockholt HJ, Bustillo JR, Cañive JM, Caprihan A, Gasparovic C, Hanlon FM, Houck JM, Jung RE, Lauriello J, Liu J, Mayer AR, Perrone-Bizzozero NI, Posse S, Stephen JM, Turner JA, Clark VP, and Calhoun VD (2017). Multimodal neuroimaging in schizophrenia: Description and dissemination. Neuroinformatics, 15(4):343–364. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Baker BT, Silva RF, Calhoun VD, Sarwate AD, and Plis SM (2015). Large scale collaboration with autonomy: Decentralized data ica. In 2015 IEEE 25th International Workshop on Machine Learning for Signal Processing (MLSP). [Google Scholar]
- Carter KW, Francis RW, Carter K, Francis R, Bresnahan M, Gissler M, Grønborg T, Gross R, Gunnes N, Hammond G, et al. (2015). Vipar: a software platform for the virtual pooling and analysis of research data. International journal of epidemiology, 45(2):408–416. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Chen Z, Pears N, Freeman M, and Austin J. (2014). A gaussian mixturemodel and support vector machine approach to vehicle type and colour classification. IET Intelligent Transport Systems, 8(2):135–144. [Google Scholar]
- de Bruijn NG (1959). Asymptotic methods in analysis. Bull. Amer. Math. Soc, 65(3):160–163. [Google Scholar]
- Dean J, Corrado G, Monga R, Chen K, Devin M, Mao M, aurelio Ranzato M, Senior A, Tucker P, Yang K, Le QV, and Ng AY (2012). Large scale distributed deep networks In Pereira F, Burges CJC, Bottou L, and Weinberger KQ, editors, Advances in Neural Information Processing Systems 25, pages 1223–1231. Curran Associates, Inc. [Google Scholar]
- Deng L. (2012). The mnist database of handwritten digit images for machine learning research. IEEE Signal Processing Magazine, 29(6):141–142. [Google Scholar]
- Dijkstra E. (1965). Solution of a problem in concurrent programming control. Communications of the ACM, 8:569. [Google Scholar]
- Fei-Fei L, Fergus R, and Perona P. (2007). Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. Computer Vision and Image Understanding, 106(1):59–70. Special issue on Generative Model Based Vision. [Google Scholar]
- Forero PA, Cano A, and Giannakis GB (2010). Consensus-based distributed support vector machines. J. Mach. Learn. Res, 11:1663–1707. [Google Scholar]
- Gazula H, Baker BT, Damaraju E, Plis SM, Panta SR, Silva RF, and Calhoun VD (2018). Decentralized analysis of brain imaging data: Voxel-based morphometry and dynamic functional network connectivity. Frontiers in neuroinformatics, 12. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Glorot X. and Bengio Y. (2010). Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, pages 249–256. [Google Scholar]
- Gollub RL, Shoemaker JM, King MD, White T, Ehrlich S, Sponheim SR, Clark VP, Turner JA, Mueller BA, Magnotta V, et al. (2013). The mcic collection: a shared repository of multi-modal, multi-site brain image data from a clinical investigation of schizophrenia. Neuroinformatics, 11(3):367–388. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, and Bengio Y. (2014a). Generative adversarial nets In Ghahramani Z, Welling M, Cortes C, Lawrence ND, and Weinberger KQ, editors, Advances in Neural Information Processing Systems 27, pages 2672–2680. Curran Associates, Inc. [Google Scholar]
- Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, and Bengio Y. (2014b). Generative adversarial nets In Ghahramani Z, Welling M, Cortes C, Lawrence ND, and Weinberger KQ, editors, Advances in Neural Information Processing Systems 27, pages 2672–2680. Curran Associates, Inc. [Google Scholar]
- Gould IC, Shepherd AM, Laurens KR, Cairns MJ, Carr VJ, and Green MJ (2014). Multivariate neuroanatomical classification of cognitive subtypes in schizophrenia: a support vector machine learning approach. NeuroImage: Clinical, 6:229–236. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hanlon FM, Houck JM, Pyeatt CJ, Lundy SL, Euler MJ, Weisend MP, Thoma RJ, Bustillo JR, Miller GA, and Tesche CD (2011). Bilateral hippocampal dysfunction in schizophrenia. Neuroimage, 58(4):1158–1168. [DOI] [PMC free article] [PubMed] [Google Scholar]
- He H, Bai Y, Garcia EA, and Li S. (2008). Adasyn: Adaptive synthetic sampling approach for imbalanced learning. In 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), pages 1322–1328. [Google Scholar]
- LeCun Y, Cortes C, and Burges C. (2010). Mnist handwritten digit database. AT&T Labs [Online]. Available: http://yann.lecun.com/exdb/mnist, 2. [Google Scholar]
- Lewis N, Plis S, and Calhoun V. (2017). Cooperative learning: Decentralized data neural network. In Neural Networks (IJCNN), 2017 International Joint Conference on, pages 324–331. IEEE. [Google Scholar]
- Liu C-L, Nakashima K, Sako H, and Fujisawa H. (2003). Handwritten digit recognition: benchmarking of state-of-the-art techniques. Pattern Recognition, 36(10):2271–2285. [Google Scholar]
- Liu Y, Hayes DN, Nobel A, and Marron JS (2008). Statistical significance of clustering for high-dimension, low–sample size data. Journal of the American Statistical Association, 103(483):1281–1293. [Google Scholar]
- Ming J, Verner E, Sarwate A, Kelly R, Reed C, Kahleck T, Silva R, Panta S, Turner J, Plis S, et al. (2017). Coinstac: Decentralizing the future of brain imaging analysis. F1000Research, 6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Plis SM, Sarwate AD, Wood D, Dieringer C, Landis D, Reed C, Panta SR, Turner JA, Shoemaker JM, Carter KW, et al. (2016). Coinstac: a privacy enabled model and prototype for leveraging and processing decentralized brain imaging data. Frontiers in neuroscience, 10:365. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Potkin S, Turner J, Brown G, McCarthy G, Greve D, Glover G, Manoach D, Belger A, Diaz M, Wible C, et al. (2008). Working memory and dlpfc inefficiency in schizophrenia: the fbirn study. Schizophrenia bulletin, 35(1):19–31. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Saha DK, Calhoun VD, Panta SR, and Plis SM (2017). See without looking: joint visualization of sensitive multi-site datasets. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI’2017), pages 2672–2678. [Google Scholar]
- Thompson PM, Stein JL, Medland SE, Hibar DP, Vasquez AA, Renteria ME, Toro R, Jahanshad N, Schumann G, Franke B, et al. (2014). The enigma consortium: large-scale collaborative analyses of neuroimaging and genetic data. Brain imaging and behavior, 8(2):153–182. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Viroli C. and McLachlan GJ (2017). Deep gaussian mixture models. CoRR, abs/1711.06929. [Google Scholar]
- Wang J. and Jo C. (2006). Performance of gaussian mixture models as a classifier for pathological voice. [Google Scholar]
- Wojtalewicz NP, Silva RF, Calhoun VD, Sarwate AD, and Plis SM (2017). Decentralized independent vector analysis. In Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference on, pages 826–830. IEEE. [Google Scholar]