Skip to main content
IEEE - PMC COVID-19 Collection logoLink to IEEE - PMC COVID-19 Collection
. 2021 Oct 12;26(1):172–182. doi: 10.1109/JBHI.2021.3119325

Mix-and-Interpolate: A Training Strategy to Deal With Source-Biased Medical Data

Yuexiang Li 1, Jiawei Chen 1, Dong Wei 1, Yanchun Zhu 2, Jianrong Wu 2, Junfeng Xiong 2, Yadong Gang 3, Wenbo Sun 3, Haibo Xu 3, Tianyi Qian 2, Kai Ma 1, Yefeng Zheng 1,
PMCID: PMC8908883  PMID: 34637384

Abstract

Till March 31st, 2021, the coronavirus disease 2019 (COVID-19) had reportedly infected more than 127 million people and caused over 2.5 million deaths worldwide. Timely diagnosis of COVID-19 is crucial for management of individual patients as well as containment of the highly contagious disease. Having realized the clinical value of non-contrast chest computed tomography (CT) for diagnosis of COVID-19, deep learning (DL) based automated methods have been proposed to aid the radiologists in reading the huge quantities of CT exams as a result of the pandemic. In this work, we address an overlooked problem for training deep convolutional neural networks for COVID-19 classification using real-world multi-source data, namely, the data source bias problem. The data source bias problem refers to the situation in which certain sources of data comprise only a single class of data, and training with such source-biased data may make the DL models learn to distinguish data sources instead of COVID-19. To overcome this problem, we propose MIx-aNd-Interpolate (MINI), a conceptually simple, easy-to-implement, efficient yet effective training strategy. The proposed MINI approach generates volumes of the absent class by combining the samples collected from different hospitals, which enlarges the sample space of the original source-biased dataset. Experimental results on a large collection of real patient data (1,221 COVID-19 and 1,520 negative CT images, and the latter consisting of 786 community acquired pneumonia and 734 non-pneumonia) from eight hospitals and health institutions show that: 1) MINI can improve COVID-19 classification performance upon the baseline (which does not deal with the source bias), and 2) MINI is superior to competing methods in terms of the extent of improvement.

Keywords: Source-biased data, interpolation, multi-source data, training strategy

I. Introduction

The outbreak of the coronavirus disease 2019, or COVID-19 in short, was declared a pandemic on March 11th, 2020 by the World Health Organization. Till March 31st, 2021, reportedly over 127 million people had been diagnosed as confirmed COVID-19 cases, and more than 2.5 million people died of this highly contagious disease.1 Early diagnosis of COVID-19 is crucial, both for timely treatment of individual patients and effective control of disease spread. On one hand, despite that the reverse transcription polymerase chain reaction (RT-PCR) assay of sputum or nasopharyngeal swab is considered as the gold standard for COVID-19 diagnosis, several studies have reported insufficient sensitivities (i.e., high false negative rates) of RT-PCR for effective early diagnosis of presumptive patients [1][3]. On the other hand, considering that most COVID-19 patients exhibit respiratory symptoms (mainly pneumonia) [4], the non-contrast thoracic computed tomography (CT) can be an alternative solution for confirming COVID-19 cases from suspected cohorts.

From thoracic CT exams, typical chest CT findings of COVID-19 reported in the literature [2], [5] include diffuse or focal ground-glass opacities, particularly bilateral and peripheral ground glass, as well as consolidative pulmonary opacities. In addition, Bernheim et al. [5] noted the progression of chest CT manifestations from early to late stages, characterized by greater lung involvement, crazing paving and reverse halo signs, etc. Moreover, both [3] and [2] reported superior sensitivities of non-contrast chest CT to RT-PCR. Under such circumstances, computer aided diagnosis (CAD) systems based on deep learning (DL) methodologies [6][8] can support radiologists in the triage, quantification and trend analysis of COVID-19 diagnosis, at least from two aspects: i) reducing the heavy workload of exam reading and increasing the throughput capacity with efficient, automated pipelines (a thin-slice CT sequence of 300 slices can take a radiologist 5–15 minutes to diagnose); and ii) reducing the chance of fatigue- and omission-induced, and lack-of-experience related (i.e., junior radiologists) misdiagnosis with objective and reliable CAD systems.

Facing the worldwide sudden outbreak of COVID-19, the research community of medical image analysis has responded quickly and proposed new methods/systems tailored for imaging-based diagnosis of COVID-19. For example, Han et al. [9] proposed a novel attention-based deep 3D multiple instance learning to classify CT scans into three categories, i.e., COVID-19, common pneumonia, and non-pneumonia, and achieved a high overall accuracy (97.9%) on a relatively small collection of data (460 CT scans). Similarly, both [10] and [11] proposed to discriminate between COVID-19 and community acquired pneumonia (CAP) from chest CT images and experimentally demonstrated moderate (88%) to high accuracies (95.5%), yet with much larger datasets (4,982 and 2,522 CT scans, respectively). Despite many studies have emerged to tackle various COVID-19 related tasks (more will be reviewed shortly), a practically relevant and important task regarding the unintended data bias is less explored. When collecting moderate to large scale training data from multiple sources, we notice that many sources provide data naturally dominated by a single class (Fig. 1). In reality, hospitals exclusively designated for accommodation of COVID-19 cases (such as the emergency specialty field hospitals built in Wuhan, China) provide mostly COVID-19 scans, whereas a physical examination institution provides mostly non-pneumonia cases, and the cases from conventional hospitals are usually CAP. As different sources are equipped with different CT scanners, scanning protocols and post-processing procedures, the generated CT scans are most likely to have implicit image characteristics representing the origins. Training with such biased data may lead to suboptimal and discriminatory outcomes, i.e., associating model predictions with data sources instead of pneumonia types. As far as the authors are aware of, despite its practical relevance, this problem has not been considered yet for imaging-based CAD of COVID-19.

Fig. 1.

Fig. 1.

The data bias problem: certain sources of data may comprise only a single class of data (e.g., Source 2 has only positive cases (COVID-19), whereas Sources 1 and 3 have only negative cases—non-pneumonia and community acquired pneumonia (CAP), respectively), introducing a source-related bias in network training. The data bias problem occurred in reality while developing the CAD for COVID-19. Hospitals exclusively designated for accommodation of COVID-19 cases (such as the emergency specialty field hospitals built in Wuhan, China) provide mostly COVID-19 scans, whereas a physical examination institution provides mostly non-pneumonia cases, and the cases from conventional hospitals are usually CAP. Models trained with such multi-source data without dealing with the source bias cannot generalize to the data collected from an unseen centre (Source 4).

In this work, we propose a training strategy, namely MIx-aNd-Interpolate (MINI), to deal with the data bias problem occurred in real-world COVID-19 diagnosis applications by expanding the sample space. Particularly, the volumes of the absent class for each data source are generated by combining samples from another source (domain) via mixing and trilinear interpolation. It is worthwhile to mention that our MINI is easy-to-implement and computation-friendly; therefore, researchers can respond rapidly to the source bias problem using our training strategy, which is especially important facing the rapid spread of COVID-19.

In summary, our contributions are three folds:

  • 1)

    Unlike existing literature, we address a practically relevant yet previously overlooked problem for training COVID-19 classification networks using real-world multi-source data, that is, certain data sources may comprise only a single class of data, introducing source bias in training.

  • 2)

    We propose MIx-aNd-Interpolate (MINI), an intuitive and easy-to-implement training strategy, to deal with the specific problem.

  • 3)

    We conduct thorough experiments to validate the effectiveness of MINI, study its behaviour and optimal configuration, and demonstrate its superiority to applicable competing methods.

II. Related Work

A. Imaging-Based Diagnosis of COVID-19

In the past few months, we have seen that a great number of imaging-based COVID-19 diagnosis systems have been proposed to use advanced deep learning (DL) techniques with various imaging modalities. Roy et al. [12] explored the application of DL methodologies for the classification and localization of COVID-19 markers using lung ultrasonography (LUS) images; Oh et al. [13] proposed a patch-based DL network architecture and employed majority voting for the final decision, to detect COVID-19 signs in chest X-rays (CXR). However, most methods were proposed for CT-based diagnosis, as CT images (especially those with thin slice below 2 mm) can capture more and finer details of the lung infection compared to LUS and CXR, thus having a higher clinical diagnostic value.

For CT images, rapid development of DL models in response to the sudden outbreak of COVID-19 is among the mostly visited topics. Gozes et al. [14] proposed to modify and adapt existing DL models that performed a similar task and combined them with initial clinical understanding of COVID-19 for rapid development of a new CAD system, at the early stage of the disease outbreak when both the data and annotation were scarce. The need for rapid model development was also considered in [15], which additionally addressed the data imbalance problem (i.e., much fewer COVID-19 cases compared to the available CAP and non-pneumonia cases) using a self-supervised dual-track ranking strategy. Besides rapid development of the models, fast screening in practical use has also drawn attention, e.g., Wang et al. [16] described a DL model that was able to process a whole CT volume—including both COVID-19 classification and lesion localization—in less than two seconds. Another frequently visited topic is to cope with the scarcity or difficulty in manual data annotation, for which both semi-supervised [17] and weakly supervised [12], [16] methodologies were proposed. Unlike all the related work described above, this work deals with a problem that has been overlooked so far, i.e., the data source bias problem, and proposes a simple yet effective approach (namely MIx-aNd-Interpolate, MINI) as a solution.

B. Unsupervised Domain Adaptation

To alleviate the problem caused by different imaging conditions of multiple centres, existing solutions applied the unsupervised domain adaptation technique to close the gap between the source and target domains. One common choice for domain adaptation is to directly align two different domains in the feature subspace [18], [19]. For example, Sun et al. [18] proposed the CORrelation Alignment (CORAL) method for unsupervised domain adaptation, which minimizes domain shift by aligning the second-order statistics of source and target distributions. In more recent studies, researchers tried to adopt generative adversarial networks (GANs) to reduce domain discrepancy [20][23] by optimizing an adversarial objective between the feature learning network and a domain discrimination network. Hoffman et al. [20] developed an method, called cycle-consistent adversarial domain adaptation (CyCADA), to guide transfer between domains according to a discriminatively trained network. The approach alleviated the divergence problem by enforcing consistency of the relevant semantics before and after adaptation. Meanwhile, the GAN-based domain adaptation is also widely used in the area of person re-identification (re-ID) [24], [25] and medical image processing, e.g., for color normalization of histopathological slices [26], intensity standardization of magnetic resonance images [27] and cross-modality adaptation [28]. However, most existing domain adaptation approaches focus on transferring knowledge between two domains (source and target), which are difficult to adapt to tasks with multi-source data (i.e., data from more than two centres) [29].

C. Domain Generalization

As an extension of domain adaptation, domain generalization (DG) aims to learn from multiple source domains to incorporate domain invariance into the model, in hopes that such invariance also holds in target domains. Most previous approaches attempted to learn a domain-invariant feature representation, typically through minimising the discrepancy between all source domains and assuming the resulting domain-invariant feature will generalize well for unseen target distributions [30][34]. More recently, meta-learning [35] starts to be used for solving DG problems. Bala et al. [36] learned an adaptive regularizer through meta-learning for cross-domain recognition. Li et al. [37] alternated domain-specific feature extractors and classifiers across domains via episodic training, but without using inner gradient descent update. Dou et al. [38] learned discriminative features that would allow for semantic coherence across meta-train and meta-test domains. However, meta-learning approaches are complicated to implement, and slow to train. Meanwhile, data augmentation strategies, such as gradient-based domain perturbation [39] or adversarially perturbed samples [40] demonstrated effectiveness for model generalization. These methods augment the source domain to a wider span of the

Algorithm 1 Volume Generation by MINI

  • 1:

    Input:

    Inline graphic Inline graphic: input CT samples from different domains (resampled to a uniform shape Inline graphic)

    Inline graphic Inline graphic: the hyperparameter controlling the number of slices taken for interpolation

  • 2:

    Output:

    Inline graphic Inline graphic: the generated CT volume

  • 3:

    Functions:

    Inline graphic Inline graphic: generate an all-zero volume of shape Inline graphic

    Inline graphic Inline graphic: randomly yield a floating point number in range Inline graphic

    Inline graphic Inline graphic: calculate the remainder

    Inline graphic Inline graphic: resize Inline graphic to shape Inline graphic using the designated interpolation method (‘trilinear’ or ‘nearest’)

  • 4:

    Volume Generation Procedure:

  • 5:

    Step 1 Initialization

  • 6:

    Inline graphic

  • 7:

    Inline graphic

  • 8:

    Inline graphic

  • 9:

    Inline graphic

  • 10:

    Inline graphic

  • 11:

    Step 2 Mixing

  • 12:

    for Inline graphic in range(Inline graphic):

  • 13:

    if Inline graphic:

  • 14:

    Inline graphic

  • 15:

    else:

  • 16:

    Inline graphic

  • 17:

    Inline graphic

  • 18:

    Step 3 Interpolation

  • 19:

    for Inline graphic in range(Inline graphic):

  • 20:

    Inline graphic

  • 21:

    Inline graphic

  • 22:

    Inline graphic

training data space for enlarging the possibility of covering the span of the data in the target domain. A recent method with the state-of-the art performance is JiGen [41], which leverages self-supervised signals by solving jigsaw puzzles.

Despite achieving promising improvements to the model generalization, both unsupervised domain adaptation and generalization approaches require the source and target domains to consist of the same categories, which cannot always be fulfilled in practical applications, e.g., when the problem of missing category occurs for multi-source data as formulated in this study.

III. Mix-and-Interpolate

Medical images from multicentres are often captured under different imaging conditions, making models trained in one domain (centre) frequently fail to generalize on another. Therefore, the generalization capacity of machine learning models to unseen domains draws increasing attention from the research community. To address the problem, recent studies have proposed various domain generalization methods [37], [38]. However, compared to the typical domain generalization, the challenge we face while developing the CAD system for COVID-19 diagnosis is more skewed. As mentioned earlier, many hospitals may provide only COVID-19 or non-pneumonia samples, which the current domain generalization methods fail to handle. Witnessed the fast outbreak of COVID-19, we propose a simple yet effective approach, namely MIx-aNd-Interpolate (MINI), to deal with such a training set containing biased multi-source data, which accordingly improves the generalization capacity of deep learning models.

A. Algorithm

The proposed MINI addresses the problem of biased multi-source data via extra data generation. Specifically, let Inline graphic (where Inline graphic, Inline graphic and Inline graphic are the shape of Inline graphic) denote a 3D CT sample in the training set. Its corresponding class label is denoted as Inline graphic. MINI aims to generate a new training sample (Inline graphic, Inline graphic) by combining two authentic training samples (Inline graphic, Inline graphic) and (Inline graphic, Inline graphic) from different domains (sources), i.e., CT volumes of different patients captured in different hospitals. The generation process of Inline graphic is summarized in Algorithm 1 and illustrated in Fig. 2. In short, MINI generates samples by mixing the volume-wise information of Inline graphic and Inline graphic via interpolation. To generate samples having closer intensity distributions to the original samples, we stretch the grayscale of Inline graphic to the average interval of Inline graphic and Inline graphic after the volume generation.

Fig. 2.

Fig. 2.

Illustration of the proposed MINI. The proposed approach consists of three steps—mix, trilinear interpolation and nearest interpolation. A hyperparameter Inline graphic is used to control the amount of information to fuse along the z-axis.

On one hand, the proposed MINI is simple, which costs a negligible computational overhead like existing data augmentation approaches, e.g., volume rotation and flipping. On the other hand, MINI-generated samples (Inline graphic) can effectively narrow down the gap between domains and improve the model generalization. Such an efficient method can be readily applied for improving the performance of any network architecture.

B. Labeling Strategy

After the generation of volume Inline graphic, we propose two strategies (hard and soft labeling) to yield the corresponding label, which are introduced below.

1). Hard Labeling

Taking binary classification as an example, this strategy assigns 0 (negative) or 1 (positive) to the generated volume, according to the original labels of Inline graphic and Inline graphic (i.e., Inline graphic and Inline graphic). Denoting the assigned label as Inline graphic, the rule of hard labeling can be formulated as:

1).

Specifically, the generated volume (Inline graphic) is treated as positive (1) if there is at least one positive sample in Inline graphic and Inline graphic. Otherwise, a negative label (0) is assigned to Inline graphic.

2). Soft Labeling

Consistent with the existing data augmentation approaches, the soft-labeling strategy is also adopted in our MINI. Particularly, the random factors Inline graphic and Inline graphic in Algorithm 1 are utilized to yield the soft label (Inline graphic) as a pair of labels Inline graphic.2 The soft labeling can reflect the similarity of the generated sample to negative and positive samples, respectively, when Inline graphic and Inline graphic are from different categories; otherwise, this strategy equals to the hard labeling strategy.

C. Analysis of Hyperparameter Inline graphic

As presented in Algorithm 1, our MINI performs information fusion along the z-axis via trilinear interpolation.3 Since the trilinear interpolation can only fuse the information from the adjacent slices, we propose an iterative interpolation strategy,4 as presented in Algorithm 1 (Lines 18–22), to involve more slices from Inline graphic and Inline graphic for the per slice generation of Inline graphic. The hyperparameter Inline graphic controls the number of slices from each training volume for the per slice generation; specifically, Inline graphic slices from Inline graphic and Inline graphic are utilized to generate a single slice of Inline graphic. The influence of this hyperparameter on performance will be empirically studied in the next section.

Additionally, as shown in Fig. 2, the proposed MINI adopts the neareast neighbor interpolation for shape unification, instead of the trilinear interpolation. The underlying reason is that the step of shape unification aims to reshape the compressed volume to the original size for neural network to process, which may not be sensitive to the choice of interpolation approach. To validate this claim, we conduct experiments and also present the results in the next section.

D. Comparison With Existing Approaches

We are aware that there are several existing approaches, e.g., Mixup [42], Cutout [43], and CutMix [44], highly related to our MINI. They adopt different ways to produce synthetic samples. Specifically, Mixup generates the synthetic sample by fusing two authentic ones via pixel-wise summation (e.g., Inline graphic). Compared to Mixup, Cutout is a simpler approach, which yields new samples by cropping part of the image content. CutMix can be seen as a hybrid method combining Mixup and Cutout. It crops part of Inline graphic and fill the content of Inline graphic to the cropped area. In this section, a comprehensive comparison (Table I) is presented to illustrate the advantage of the proposed MINI, compared to the existing approaches.

TABLE I. Comparison Between Existing Approaches and Our MINI.

Cutout CutMix Mixup MINI
Full volume region ×
Mixed label ×
Complete lesion × ×
Cross slice fusion × × ×

As shown in Table, only the Cutout approach encounters the problem of lacking usage of ‘"full volume region"’ and ‘"mixed label"’. In particular, lacking usage of ‘"full volume region"’ is caused by the discard operation in Cutout, which decreases the information contained in the newly generated samples, i.e., part of the volume (all black) provides no useful information at all. Moreover, the Cutout approach generates new samples without fusing information from two volumes. Hence, it assigns the original label to the generated samples, instead of mixing the labels from the fusing volumes (‘"mixed label"’).

Some existing approaches (Cutout and CutMix) face the problem of ‘"complete lesion"’, since they block/replace part of the volume content to generate new samples. When these approaches happen to crop and discard the lesion areas, the label of the generated volume may change (i.e., COVID-19 samples without lesion area become negative samples). However, these approaches still assign the original labels to the generated volumes. Training with such ‘"incorrectly-labeled"’ data, the performance of deep learning network may thereby degrade.

Last but not least, the existing approaches share a common shortage when applied to 3D volumes—the lack of cross-slice information fusion along the z-axis. Particularly, Mixup, Cutout, and CutMix perform slice-wise fusion using voxel-wise summation, region cropping, and region filling, respectively. As a result, the information cannot be transmitted and fused properly between adjacent slices. Current studies [45] demonstrate that the integration of spatial information plays an important role for 3D medical image processing. In this regard, our MINI adopts the trilinear interpolation to properly fuse per slice information during volume generation.

IV. Experiments

In this section, we evaluate our MINI for its capability to improve the model generalization on multi-domain data. We first conduct extensive experiments to evaluate the impacts of the hyperparameter Inline graphic, labeling strategies and interpolation approaches. Next, the proposed MINI is compared with the state-of-the-art approaches to validate its effectiveness. Finally, we also show that our MINI can effectively deal with an increasing number of domains/centres, i.e., a training set with more diverse samples.

A. Experimental Settings

1). Dataset

We collected 2,173 CT volumes from seven hospitals exclusively for model training, where each of them provided one of the three categories: COVID-19 (1,046), CAP (652) and non-pneumonia (475). The detailed information of the collected data is shown in Table II. In this study, we focus on the COVID-19 diagnosis task, which aims to separate COVID-19 (positive) patients from the negatives (i.e., CAP and non-pneumonia). The DL models trained with the multi-source data are then evaluated on a test set, consisting of 175 COVID-19, 134 CAP and 259 non-pneumonia CT samples, collected from the eighth hospital. This experimental setting enables us to validate the generalization of models trained with multi-source biased data on an unseen domain.

TABLE II. Detailed Information of Multi-Source Dataset.
Centre No. of CT Volumes No. slices Slice thickness (mm) Matrix size Pixel spacing (mm)
Hospital #1 Inline graphic 305 38–629 0.625–7.5 512×512 0.604–0.977
Hospital #2 Inline graphic 741 79–547 1.0–1.25 512×512 0.580–0.912
Hospital #3 Inline graphic 144 47–553 1.0–5.0 512×512 0.330–0.977
Training Hospital #4 Inline graphic 217 125–407 1.0–1.5 512×512 0.387–0.977
Hospital #5 Inline graphic 291 156–591 0.6–1.5 512×512 0.324–0.977
Hospital #6 Inline graphic 201 29–131 5.0 512×512 0.430–0.977
Hospital #7 Inline graphic 274 56–950 0.6–2.0 512×512 0.396–0.977
Test Hospital #8 Inline graphic 568 65–710 0.6–3.0 512×512 0.342–0.998

Inline graphic, Inline graphic, Inline graphic Indicate the COVID-19, CAP and non-pneumonia, respectively.

2). Evaluation Metric

The sensitivity (SEN), specificity (SPE), and F1 score are adopted as the metrics to evaluate the performance of COVID-19 diagnosis, which can be written as:

2).

where Inline graphic, Inline graphic, Inline graphic, and Inline graphic represent the numbers of true negatives, true positives, false positives, and false negatives, respectively. The average classification accuracy (ACC) and area under curve (AUC) are also adopted as metrics for the performance evaluation.5

3). Implementation

All experiments are implemented using PyTorch and according to the same training protocol. The 3D ResNet-18 [46] is adopted as the backbone, which is trained with a mini-batch size of 16. The initial learning rate is set to 0.001, which decreases with a gamma of 0.1 every 30 epochs of training. We train the network for 200 epochs. The Adam solver [47] is used for optimization.

It is worthwhile to mention that all the augmentation algorithms generate the equal numbers of samples during network training for a fair comparison. Concretely, the online data augmentation approach is adopted—the model decides whether to augment the input volume based on a random process. For the Mixup, Cutout, CutMix and our MINI, we first draw a random value (Inline graphic) according to a uniform distribution in [0, 1]. If Inline graphic is larger than 0.5, the model randomly selects another sample from the batch and performs corresponding augmentation approach with the selected and input volumes to generate a new sample for network training. For the conventional training strategy, an augmentation approach is randomly selected from the pool (i.e., torchvision.transforms, which includes rotation, flipping and elastic deformation, etc.) for new sample generation if Inline graphic. The input sample is maintained and fed to the neural network if Inline graphic. Hence, the total number of samples for network training is exactly the same (i.e., the size of training set) for different training strategies.

During inference, the well-trained network outputs the possibility for the positive class. If the possibility is larger than 0.5, the patient is identified as COVID-19. Note that the proposed MINI and existing approaches (e.g., Mixup, Cutout and CutMix) are used only for training. After that, the trained classification network can be used for COVID-19 classification without any data augmentation.

B. Empirical Experiments on Design Options

We conduct an ablation study to evaluate the influence of the hyperparameter Inline graphic and different labeling strategies, and present the evaluation results in this section.

1). Hyperparameter Inline graphic

The hyperparameter Inline graphic controls the amount of information fused along the z-axis—a larger Inline graphic leads to a fusion of more slices for the per slice generation. To this end, the proposed MINI with different values of Inline graphic is assessed and the corresponding COVID-19 classification performance is presented in Table III. Here, the soft-labeling strategy is used to yield the labels for generated samples.

TABLE III. Performance (%) Analysis of MINI With Different Values of Hyperparameter Inline graphic (SEN–Sensitivity, SPE–Specificity, ACC–Accuracy, AUC–Area Under Curve).
Inline graphic SEN SPE ACC AUC Inline graphic
2 84.3 87.3 86.4 92.7 79.3
3 86.6 87.3 87.1 92.8 80.6
4 85.4 87.3 86.7 92.7 79.9
5 84.3 86.2 85.7 92.7 78.4

As we can see, the best COVID-19 classification performance is achieved by setting Inline graphic to 3, i.e., fusing the information of four adjacent slices. The overall classification performance is observed to decrease as Inline graphic further increases. The underlying reason for the performance degradation may be the information loss while compacting the volume (Inline graphic) to a smaller shape Inline graphic.

2). Labeling Strategy

The results of MINI using the different labeling strategies are listed in Tabel IV. The 3D ResNet-18 trained without MINI is also involved for comparison.

TABLE IV. COVID-19 Classification Performance (%) of MINI Using Different Labeling Strategies With Inline graphic (SEN–Sensitivity, SPE–Specificity, ACC–Accuracy, AUC–Area Under Curve).
SEN SPE ACC AUC Inline graphic
w/o MINI 77.0 87.0 83.9 89.6 74.7
Hard labeling 88.1 85.6 86.3 92.6 79.9
Soft labeling 86.6 87.3 87.1 92.8 80.6

It can be observed that using either hard or soft labeling, our MINI substantially improves the overall COVID-19 classification performance, compared to the baseline (w/o MINI). Since the soft labeling can better reflect the information contained in the generated volumes, i.e., the similarity of Inline graphic to the negative and positive samples, it yields a larger improvement to the baseline (i.e., Inline graphic to Inline graphic), compared to the hard labeling.6 Nevertheless, the MINI with hard labeling yields a higher sensitivity than the one using soft labeling. The reason may be that the hard labeling strategy generates more positive samples (i.e., Inline graphic if Inline graphic or Inline graphic equals to 1) for the network to learn, which biases the model to make more positive predictions.

3). Interpolation Approach

As previously mentioned, our MINI adopts the trilinear interpolation to downsample the fused volume and the nearest neighbor to uniform the volume size. To analyze the influence caused by the selections, we evaluate the performance of the proposed MINI with different combinations of interpolation approaches. The widely-used areal interpolation [48] is also adopted for evaluation.

a) Downsampling:. We downsample the mixed volume to fuse the information of two authentic training samples. In this experiment, we maintain the nearest neighbor interpolation for upsampling and evaluate the performance of MINI with different interpolation approaches for information fusion. The experimental results are presented in Table V. The processing time for each interpolation approach is also evaluated. It can be observed that our MINI achieves the best COVID-19 classification performance using the trilinear interpolation, due to its better capacity of information fusion, with the similar processing time, compared to the others. Note that the processing time for each interpolation approach is tested with a single Tesla V100.

TABLE V. COVID-19 Classification Performance (%) of MINI Using Different Unterpolation Approaches for Information Fusion (SEN–Sensitivity, SPE–Specificity, ACC–Accuracy, AUC–Area Under Curve, P. T.–Processing Time for Interpolation, Neig.–Neighbor). Here, MINI Uses the Soft Labeling Strategy With Inline graphic.
SEN SPE ACC AUC Inline graphic P. T.
Nearest neig. 82.3 84.2 83.6 89.1 75.6 0.17 ms
Areal [48] 84.0 86.3 85.6 91.6 78.2 0.20 ms
Trilinear 86.6 87.3 87.1 92.8 80.6 0.22 ms

b) Shape Unification:. We also assess the performance of the proposed MINI with different interpolation approaches for shape unification. The experimental results are presented in Table VI. No significant difference between the accuracies achieved by different interpolation approaches is observed, which demonstrates that our MINI is insensitive to the choice of upsampling. Note that the trilinear interpolation is adopted for downsampling in this experiment.

TABLE VI. COVID-19 Classification Performance (%) of MINI Using Different Interpolation Approaches for Shape Unification (SEN–Sensitivity, SPE–Specificity, ACC–Accuracy, AUC–Area Under Curve). Here, MINI Uses the Soft Labeling Strategy With Inline graphic.
SEN SPE ACC AUC Inline graphic
Trilinear 86.8 86.3 88.0 92.2 80.4
Areal [48] 86.3 87.0 86.8 93.1 80.1
Nearest neighbor 86.6 87.3 87.1 92.8 80.6

C. Comparison with State-of-the-Art

In this section, the proposed MINI is compared with several state-of-the-art approaches, including Cutout, CutMix, and Mixup. For a fair comparison, we extend the typical 2D Cutout and CutMix to 3D versions, i.e., cropping and replacing sub-volume for Cutout and CutMix, respectively.

1). Visualization of Generated Volumes

The volumes generated by different approaches are illustrated in Fig. 3. A non-pneumonia sample (Inline graphic) and a COVID-19 sample (Inline graphic) are adopted for volume generation. As shown in Fig. 3, volumes yielded by Cutout and CutMix may encounter the problem of incomplete lesions (i.e., the lesion area may be cropped or replaced), which makes the generated volume (Inline graphic) contain inconsistent information with its label (Inline graphic). For example, the Cutout still assigns the positive label (1) to Inline graphic even if the lesion area is cropped. In contrast, Mixup and our MINI can alleviate the problem by excellently maintaining the lesions. By comparing the volumes generated by Mixup and MINI, we can observe that the proposed MINI prefers to preserve the primary information and discard unnecessary details. In our experiments, the neural networks are observed to be easier to train and converge with such volumes. As shown in Fig. 4, the model using MINI has a better convergence (i.e., lower training loss) than the one with Mixup.

Fig. 3.

Fig. 3.

Visualization of samples generated by different augmentation approaches. The samples generated by Cutout and CutMix face the risk of mistakenly discarding the lesion areas. Mixup and our MINI can alleviate the problem—maintaining the lesion areas via voxel-summation and trilinear interpolation, respectively.

Fig. 4.

Fig. 4.

Training curves for models using existing data augmentation approaches and our MINI. The model with MINI is observed to better converge (i.e., more stable with the lower loss values).

2). Quantitative Comparison

It is worthwhile to mention that we repeat this experiment five times to reduce the influence caused by random nature of network training—such a setting also enables us to conduct a statistical significance analysis of performance improvement. Hence, the results reported in the section are the average results of five repeated experiments with corresponding standard deviations. The performance of COVID-19 identification yielded by networks trained with different strategies are presented in Table VII. In general, the networks trained with generated volumes achieve consistent improvements for COVID-19 identification, compared to the baseline which is trained with only the original volumes. The inconsistent information contained in the volumes generated by Cutout and CutMix degrades their improvements. The network trained with our MINI with soft labeling achieves the best performance in most metrics (i.e., 87.3%, 87.1%, 92.8% and 80.6% for SPE, ACC, AUC and Inline graphic, respectively). Specifically, the Inline graphic of our MINI is Inline graphic and Inline graphic higher than the following-up (Mixup) and baseline, respectively.

TABLE VII. Comparison of COVID-19 Classification Performance (%) With Existing Augmentation Approaches (SEN–Sensitivity, SPE–Specificity, ACC–Accuracy, AUC–Area Under Curve, HL–Hard Labeling, SL–Soft Labeling).
SEN SPE ACC AUC Inline graphic
Baseline Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
Cutout [43] Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
CutMix [44] Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
Mixup [42] Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
Mixup w.HL Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
MINI w.HL Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic
MINI w.SL Inline graphic Inline graphic Inline graphic Inline graphic Inline graphic

Statistical Significance. A t-test validation is conducted using the results of repeated experiments to validate the statistical significance between our MINI and the following-up Mixup. A p-value of 0.039 is obtained, which indicates that the accuracy improvement (Inline graphic) produced by our MINI is statistically significant at the 5% significance level.

To compare the feature representations learned by frameworks with and without MINI, we visualize the embedded features of the test set using t-SNE [49] in Fig. 5. It can be observed that the negative cluster becomes more compact using our MINI (Fig. 5(b)), compared to the baseline (Fig. 5(a)), which is easier for the classifier to draw the decision boundary.

Fig. 5.

Fig. 5.

Comparison of t-SNE between models without (a) and with MINI (b). Our MINI is observed to compact the cluster of negative in the feature space.

3). Comparison Against Radiologists

We also conduct a human-machine confrontation to evaluate the potential of our deep learning framework for clinical applications. Since the labels for network training are provided by senior radiologists via comprehensive consideration of the RT-PCR results and clinical information, which is the upper limit of our model accuracy, a junior radiologist is invited to compete with our framework. A new test set, consisting of 150 negatives and 150 positives, is collected for human-machine confrontation. The experimental results are presented in Table VIII. It can be observed that the framework trained with our MINI outperforms the radiologist with all metrics, which demonstrates its potential for clinical application.

TABLE VIII. Comparison With Radiologists on an Independent Test Set With 300 Patients.
SEN SPE ACC Inline graphic
Radiologist 83.7 83.2 83.3 83.3
MINI (ours) 86.3 84.0 83.9 85.0

D. Performance With Different Numbers of Centres

To investigate the influence of multi-source biased data on the generalization capacity of deep learning models, we conduct an experiment evaluating the performance of the 3D ResNet-18 trained on data collected from different numbers of centres, while fixing the size of the training set to 1000 (500 COVID-19250 CAP and 250 non-pneumonia volumes). The training samples become more diverse as the number of centres increases. The detailed information about the training set composition with different numbers of centers is shown in Table IX.

TABLE IX. Detailed Information of Training Sets Containing Samples From Different Numbers of Centres.

No. of centres 3 4 5 6 7
Hospital #1 Inline graphic 250 250 250 250
Hospital #2 Inline graphic 500 250 250 250 250
Hospital #3 Inline graphic 50
Hospital #4 Inline graphic 125 125 100
Hospital #5 Inline graphic 250 250 125 125 100
Hospital #6 Inline graphic 125 125
Hospital #7 Inline graphic 250 250 250 125 125

Inline graphic, Inline graphic, Inline graphic Indicate the COVID-19, CAP And Non-Pneumonia Respectively.

The evaluation results are presented in Fig. 6, where Inline graphic is adopted as the evaluation metric for the performance of COVID-19 identification. It can be observed that as the diversity of the training set increases, the source-biased problem is alleviated, i.e., Inline graphic of the baseline (i.e., without MINI) increases from around 73% to around 76% when data from four more centres is included.

Fig. 6.

Fig. 6.

The Inline graphic of COVID-19 identification yielded by networks trained with data collected from different numbers of centres. Here, soft labeling is adopted to yield the labels for generated volumes.

Meanwhile, our MINI can effectively deal with the data source bias problem: an improvement of about Inline graphic is achieved upon the baseline with the training set consisting of only three centres; in addition, MINI continues to substantially improve the Inline graphic score when more centres are included (e.g., the Inline graphic score of the network trained with data from seven centres reaches 79% using MINI). Note that we use common augmentation approaches (e.g., vertical flipping) to generate the equal numbers of additional samples to that with MINI for a fair comparison. The experimental results demonstrate the effectiveness of our MINI approach to the data source bias problem encountered in the practical implementation of CAD systems for COVID-19 diagnosis.

V. Discussion

Although human observers may not be able to distinguish the subtle differences between CT volumes captured at different centres, such differences may be explicit to DL networks. To gain insights into the data source bias problem, we try to train a deep learning network to classify the sources of data. The same training and test sets are used to train and evaluate the deep learning network on data source classification (i.e., a single-label 7-class task). The experimental results show that the well-trained model can successfully identify the data sources, achieving an accuracy (ACC) of 96.40%. Therefore, the well-trained model can easily classify the positive/negative samples from a source biased dataset by their data originalities. Such a model often fails to generalize to an unseen domain (i.e., a new centre) with data from all the disease categories. Taking a broader view, it can be easily seen that the data source bias problem may also happen for DL-based classification of diseases beyond COVID-19 which involves multi-centre data, and our proposed MINI can be widely used in such scenario for handy performance improvement, thus making a broader impact.

As revealed by our empirical experiment, the data source bias problem can be mitigated by increasing the number of data sources. As shown in Fig. 6, the Inline graphic score without MINI increases from around 73% to around 76% while keeping the numbers of training data (both total and class-specific) unchanged. We conjecture that this is because the addition of data from more centres narrows the domain gap between centres, to a limited extent. Nonetheless, our proposed MINI can further narrow down the domain gaps and thus improves the classification performance for all configurations of centres. The experimental results imply that simply adding data from more centres (which is often expensive or even impractical) is not enough for dealing with the source bias problem, and the proposed MINI can be an effective solution despite the number of centres involved.

We study two labeling strategies for the proposed MINI in this work, i.e., hard versus soft labeling. As shown in Table IV, both strategies brought apparent general performance improvement upon the baseline, indicating effectiveness of MINI regardless of the labeling strategy. Meanwhile, hard labeling leads to a higher sensitivity than soft labeling, whereas the latter leads to better specificity and overall performance (indicated by higher AUC and Inline graphic) than the former. The underlying reason may be that the hard labeling strategy probably assigns more positive labels to the generated volumes (the generated volume is labeled positive as long as at least one of Inline graphic and Inline graphic is positive), resulting in elevated sensitivity. A similar trend can be observed in Table VII, where the hard labeling strategy still achieves the highest sensitivity (and comparable values for other metrics) when compared to competing methods, and the soft labeling achieves the highest specificity, the best overall performance, and the second best sensitivity among the competing methods. We believe the decision on which labeling strategy to use is up to the practical need: if the extremely high sensitivity is required (with reasonable overall performance at the same time), then the hard labeling strategy should be preferred; otherwise, the soft labeling strategy can be picked for its more balanced performance.

Although existing approaches (e.g., Mixup) can also alleviate the source bias problem, i.e., yielding improvements upon the baseline, as shown in Table VII, the proposed MINI approach outperforms them by around Inline graphic. The underlying reason is that our MINI takes the cross slice information into account. The voxel-wise summation adopted by Mixup may decrease the contrast of lesion areas against the background tissue while performing a fusion of COVID-19 and normal volumes. In contrast, our MINI integrates the information of multiple slices via trilinear interpolation, which assists to construct a 3D view of lesion area and accordingly reduce the loss of lesion-related information when fusing COVID-19 and normal volumes. To validate the importance of lesion area for COVID-19 identification, we invite experienced radiologists to provide the ground truth of lesion areas and visualize the activation maps of neural network for COVID-19 samples. The visualization results are presented in Fig. 7. The annotations provided by radiologists are marked with red rectangles. The visualization results demonstrate that the neural network identifies COVID-19 patients mainly based on the lesion area.

Fig. 7.

Fig. 7.

Visualization of the activation maps of neural network for COVID-19 samples. The annotations provided by radiologists are drawn with red rectangles.

As an immediate response to the COVID-19 pandemic, this work focuses on the differentiation of COVID-19 from both CAP and non-pneumonia cases. However, it would be more useful in clinical routine to develop a CAD system that can also distinguish non-pneumonia and CAP cases. In this regard, we train a multi-class classification framework and report its performance on the test set in Table X. The mean accuracy of the model using our MINI is 74.5%, which is Inline graphic higher than vanilla one.

TABLE X. Classification Accuracy (ACC %) of a Multi-Class Classification Framework Trained With and Without MINI.

Normal CAP COVID-19 Mean
w/o MINI 54.1 73.9 84.0 70.7
w MINI 72.6 61.2 89.7 74.5

Last but not least, we identify limitations of this work that warrant future work. Consistent to the previous works [42], [44], the proposed MINI is implemented to fuse the information of two volumes. However, we notice that our MINI can be further expanded to deal with multiple (more than two) samples, which results in a higher diversity of generated volumes. To this end, we will try to improve the performance of the proposed MINI along this direction in our future study. Meanwhile, interpretability of deep learning models [50] is drawing increasing attention of the community of medical image analysis, which can provide evidences to back up as well as boost users’ confidence in the networks’ predictions. Although the class activation maps in Fig. 7 provide some clues illustrating the mechanism of our model for COVID-19 identification, we plan to deeply explore along this direction and further enhance the model interpretability in the future.

VI. Conclusion

In this work, we looked into the data source bias problem when using multi-centre data for training deep neural networks for COVID-19 classification. Subject to practical conditions, a certain centre may provide samples of only a single class (e.g., COVID-19), while another centre may provide those of another class (e.g., non-pneumonia). Training with such data without any countermeasure may potentially bias the trained networks towards identifying sources of data instead of the pneumonia. Despite being practically relevant, this problem had been overlooked until this work, which presented MIx-aNd-Interpolate (MINI)—a conceptually simple, computationally efficient, and effective training strategy dealing with the source bias problem. Specifically, MINI generated volumes of the absent class by combining samples collected from different centres, which enlarged the sample space of the original source-biased dataset, and worked as a lightweight online data augmentation strategy. The experimental results on a large-scale real-patient dataset composed of samples collected from eight medical centres demonstrated 1) the effectiveness of the proposed MINI in dealing with the data source bias and thus improving classification performance upon the baseline, and 2) its superiority to several recently proposed strategies that can be employed against the source bias problem. In the future, we expect MINI to be broadly applied to training classification networks for other diseases beyond COVID-19.

Funding Statement

This work was supported by the Key-Area Research and Development Program of Guangdong Province, China under grant 2018B010111001, National Key R&D Program of China under grant 2018YFC2000702, and the Scientific and Technical Innovation 2030- “New Generation Artificial Intelligence” Project under grant 2020AAA0104100.

Footnotes

1

[Online]. Available: https://covid19.who.int/.

2

The optimization process of the mixed label Inline graphic is the same as Mixup [42].

3

Trilinear interpolation is a method of multivariate interpolation on a 3-dimensional regular grid. It approximates the value of intermediate voxel by taking the surrounding eight corner voxels into consideration. For detailed information, please refer to http://paulbourke.net/miscellaneous/interpolation.

4

A series of Inline graphics is generated (one smaller than the preceding one), and only the last one (in the loop) is retained (all preceding ones being overwritten).

5

The Inline graphic score, providing a comprehensive evaluation of classification performance, is taken as the primary metric, while the others are also listed for reference.

6

In the following experiments, we report the COVID-19 classification performance of MINI with Inline graphic and soft-labeling.

Contributor Information

Yuexiang Li, Email: vicyxli@tencent.com.

Jiawei Chen, Email: jiaweichen@tencent.com.

Dong Wei, Email: donwei@tencent.com.

Yanchun Zhu, Email: noonezhu@tencent.com.

Jianrong Wu, Email: edwinjrwu@tencent.com.

Junfeng Xiong, Email: francoxiong@tencent.com.

Yadong Gang, Email: gangyadong@hotmail.com.

Wenbo Sun, Email: sunwb3@mail2.sysu.edu.cn.

Haibo Xu, Email: xuhaibo1120@hotmail.com.

Tianyi Qian, Email: tianyiqian@tencent.com.

Kai Ma, Email: kylekma@tencent.com.

Yefeng Zheng, Email: yefengzheng@tencent.com.

References

  • [1].Ai T. et al. , “Correlation of chest CT and RT-PCR testing in cronavirus disease 2019 (COVID-19) in China: A report of 1014 cases,” Radiology, vol. 296, no. 2, pp. E32–E40, 2020, Art. no. 200642. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Fang Y. et al. , “Sensitivity of chest CT for COVID-19: Comparison to RT-PCR,” Radiology, vol. 296, no. 2, pp. E115–E117, 2020, Art. no. 200432. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Xie X., Zhong Z., Zhao W., Zheng C., Wang F., and Liu J., “Chest CT for typical coronavirus disease 2019 (COVID-19) pneumonia: Relationship to negative RT-PCR testing,” Radiology, vol. 296, no. 2, pp. E41–E45, 2020, Art. no. 200343. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [4].Huang C. et al. , “Clinical features of patients infected with 2019 novel coronavirus in Wuhan, China,” Lancet, vol. 395, no. 10223, pp. 497–506, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [5].Bernheim A. et al. , “Chest CT findings in coronavirus disease-19 (COVID-19): Relationship to duration of infection,” Radiology, vol. 295, no. 3, 2020, Art. no. 200463. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [6].Butt C., Gill J., Chun D., and Babu B. A., “Deep learning system to screen coronavirus disease 2019 pneumonia,” Appl. Intell., vol. 6, no. 10, pp. 1–7, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [7].Li L. et al. , “Artificial intelligence distinguishes COVID-19 from community acquired pneumonia on chest CT,” Radiology, vol. 296, no. 2, pp. 65–71, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [8].Zhang K. et al. , “Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography,” Cell, vol. 181, no. 6, pp. 1423–1433, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [9].Han Z. et al. , “Accurate screening of COVID-19 using attention based deep 3D multiple instance learning,” IEEE Trans. Med. Imag., vol. 39, no. 8, pp. 2584–2594, Aug. 2020. [DOI] [PubMed] [Google Scholar]
  • [10].Ouyang X. et al. , “Dual-sampling attention network for diagnosis of COVID-19 from community acquired pneumonia,” IEEE Trans. Med. Imag., vol. 39, no. 8, pp. 2595–2605, Aug. 2020. [DOI] [PubMed] [Google Scholar]
  • [11].Kang H. et al. , “Diagnosis of coronavirus disease 2019 (COVID-19) with structured latent multi-view representation learning,” IEEE Trans. Med. Imag., vol. 39, no. 8, pp. 2606–2614, Aug. 2020. [DOI] [PubMed] [Google Scholar]
  • [12].Roy S. et al. , “Deep learning for classification and localization of COVID-19 markers in point-of-care lung ultrasound,” IEEE Trans. Med. Imag., vol. 39, no. 8, pp. 2676–2687, Aug. 2020. [DOI] [PubMed] [Google Scholar]
  • [13].Oh Y., Park S., and Ye J. C., “Deep learning COVID-19 features on CXR using limited training data sets,” IEEE Trans. Med. Imag., vol. 39, no. 8, pp. 2688–2700, Aug. 2020. [DOI] [PubMed] [Google Scholar]
  • [14].Gozes O. et al. , “Rapid AI development cycle for the coronavirus (COVID-19) pandemic: Initial results for automated detection & patient monitoring using deep learning CT image analysis,” 2020, arXiv:2003.05037.
  • [15].Li Y. et al. , “Efficient and effective training of COVID-19 classification networks with self-supervised dual-track learning to rank,” IEEE J. Biomed. Health Informat., vol. 24, no. 10, pp. 2787–2797, Oct. 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [16].Wang X. et al. , “A weakly-supervised framework for COVID-19 classification and lesion localization from chest CT,” IEEE Trans. Med. Imag., vol. 39, no. 8, pp. 2615–2625, Aug. 2020. [DOI] [PubMed] [Google Scholar]
  • [17].Fan D.-P. et al. , “Inf-Net: Automatic COVID-19 lung infection segmentation from CT images,” IEEE Trans. Med. Imag., vol. 39, no. 8, pp. 2626–2637, Aug. 2020. [DOI] [PubMed] [Google Scholar]
  • [18].Sun B., Feng J., and Saenko K., “Return of frustratingly easy domain adaptation,” in Proc. AAAI Conf. Artif. Intell., 2016, pp. 2058–2065. [Google Scholar]
  • [19].Ganin Y. et al. , “Domain adversarial training of neural networks,” J. Mach. Learn. Res., vol. 17, no. 1, pp. 2096–2030, 2017. [Google Scholar]
  • [20].Hoffman J., Tzeng E., Park T., and Zhu J.-Y., “CyCADA: Cycle-consistent adversarial domain adaptation,” in Proc. Int. Conf. Mach. Learn., 2018, pp. 1989–1998. [Google Scholar]
  • [21].Sankaranarayanan S., Balaji Y., Castillo C. D., and Chellappa R., “Generate to adapt: Aligning domains using generative adversarial networks,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2018, pp. 8503–8512. [Google Scholar]
  • [22].Huang S., Lin C., Chen S., Wu Y., Hsu P., and Lai S., “AugGAN: Cross domain adaptation with GAN-based data augmentation,” in Proc. Eur. Conf. Comput. Vis., 2018, pp. 718–731. [Google Scholar]
  • [23].Chen M. H., Kira Z., AlRegib G., Woo J., Chen R., and Zheng J., “Temporal attentive alignment for large-scale video domain adaptation,” in Proc. IEEE Int. Conf. Comput. Vis., 2019, pp. 6320–6329. [Google Scholar]
  • [24].Deng W., Zheng L., Ye Q., Kang G., Yang Y., and Jiao J., “Image-image domain adaptation with preserved self-similarity and domain-dissimilarity for person re-identification,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2018, pp. 994–1003. [Google Scholar]
  • [25].Wei L., Zhang S., Gao W., and Tian Q., “Person transfer GAN to bridge domain gap for person re-identification,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2018, pp. 79–88. [Google Scholar]
  • [26].Zanjani F. G., Zinger S., Bejnordi B. E., van der Laak J. A. W. M., and de With P. H. N., “Stain normalization of histopathology images using generative adversarial networks,” in Proc. IEEE Int. Symp. Biomed. Imag., 2018, pp. 573–577. [Google Scholar]
  • [27].Gao Y., Liu Y., Wang Y., Shi Z., and Yu J., “A universal intensity standardization method based on a many-to-one weak-paired cycle generative adversarial network for magnetic resonance images,” IEEE Trans. Med. Imag., vol. 38, no. 9, pp. 2059–2069, Sep. 2019. [DOI] [PubMed] [Google Scholar]
  • [28].Zhang Z., Yang L., and Zheng Y., “Translating and segmenting multimodal medical volumes with cycle- and shape-consistency generative adversarial network,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2018, pp. 9242–9251. [Google Scholar]
  • [29].Peng X., Bai Q., Xia X., Huang Z., Saenko K., and Wang B., “Moment matching for multi-source domain adaptation,” in Proc. IEEE Int. Conf. Comput. Vis., 2019, pp. 1406–1415. [Google Scholar]
  • [30].Muandet K., Balduzzi D., and Schölkopf B., “Domain generalization via invariant feature representation,” in Proc. Int. Conf. Mach. Learn., 2013, pp. 10–18. [Google Scholar]
  • [31].Ghifary M., Kleijn W. Bastiaan, Zhang M., and Balduzzi D., “Domain generalization for object recognition with multi-task autoencoders,” in Proc. IEEE Int. Conf. Comput. Vis., 2015, pp. 2551–2559. [Google Scholar]
  • [32].Motiian S., Piccirilli M., Adjeroh D. A., and Doretto G., “Unified deep supervised domain adaptation and generalization,” in Proc. IEEE Int. Conf. Comput. Vis., 2017, pp. 5715–5725. [Google Scholar]
  • [33].Li Y. et al. , “Deep domain generalization via conditional invariant adversarial networks,” in Proc. Eur. Conf. Comput. Vis., 2018, pp. 624–639. [Google Scholar]
  • [34].Li H., Pan S. Jialin, Wang S., and Kot A. C., “Domain generalization with adversarial feature learning,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2018, pp. 5400–5409. [Google Scholar]
  • [35].Hospedales T., Antoniou A., Micaelli P., and Storkey A., “Meta-learning in neural networks: A survey,” 2020, arXiv:2004.05439. [DOI] [PubMed]
  • [36].Balaji Y., Sankaranarayanan S., and Chellappa R., “ MetaReg: Towards domain generalization using meta-regularization,” in Proc. Adv. Neural Inf. Process. Syst., 2018, pp. 998–1008. [Google Scholar]
  • [37].Li D., Zhang J., Yang Y., Liu C., Song Y.-Z., and Hospedales T. M., “Episodic training for domain generalization,” in Proc. IEEE Int. Conf. Comput. Vis., 2019, pp. 1446–1455. [Google Scholar]
  • [38].Dou Q., de Castro D. C., Kamnitsas K., and Glocker B., “Domain generalization via model-agnostic learning of semantic features,” in Proc. Adv. Neural Inf. Process. Syst., 2019, pp. 6450–6461. [Google Scholar]
  • [39].Shankar S., Piratla V., Chakrabarti S., Chaudhuri S., Jyothi P., and Sarawagi S., “Generalizing across domains via cross-gradient training,” 2018, arXiv:1804.10745.
  • [40].Volpi R., Namkoong H., Sener O., Duchi J. C., Murino V., and Savarese S., “Generalizing to unseen domains via adversarial data augmentation,” in Proc. Adv. Neural Inf. Process. Syst., 2018, pp. 5334–5344. [Google Scholar]
  • [41].Carlucci F. M., D’Innocente A., Bucci S., Caputo B., and Tommasi T., “Domain generalization by solving Jigsaw puzzles,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2019, pp. 2229–2238. [Google Scholar]
  • [42].Zhang H., Cisse M., Dauphin Y. N., and Lopez-Paz D., “Mixup: Beyond empirical risk minimization,” 2017, arXiv:1710.09412.
  • [43].DeVries T. and Taylor G. W., “Improved regularization of convolutional neural networks with cutout,” 2017, arXiv:1708.04552.
  • [44].Yun S., Han D., Oh S. J., Chun S., Choe J., and Yoo Y., “CutMix: Regularization strategy to train strong classifiers with localizable features,” in Proc. IEEE Int. Conf. Comput. Vis., 2019, pp. 6022–6031. [Google Scholar]
  • [45].Dou Q. et al. , “3D deeply supervised network for automated segmentation of volumetric medical images,” Med. Image Anal., vol. 41, pp. 40–54, 2017. [DOI] [PubMed] [Google Scholar]
  • [46].He K., Zhang X., Ren S., and Sun J., “Deep residual learning for image recognition,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2016, pp. 770–778. [Google Scholar]
  • [47].Kingma D. P. and Ba J., “Adam: A method for stochastic optimization,” 2014, arXiv:1412.6980.
  • [48].Comber A. and Zeng W., “Spatial interpolation using areal features: A review of methods and opportunities using new forms of data with coded illustrations,” Geography Compass, vol. 13, no. 10, 2019, Art. no. e12465. [Google Scholar]
  • [49].Maaten L. and Hinton G., “Visualizing data using t-SNE,” J. Mach. Learn. Res., vol. 9, pp. 2579–2605, 2008. [Google Scholar]
  • [50].Chakraborty S. et al. , “Interpretability of deep learning models: A survey of results,” in Proc. IEEE Smartworld, Ubiquitous Intell. Comput., Adv. Trusted Computed, Scalable Comput. Commun., Cloud Big Data Comput., Internet People Smart City Innov., 2017, pp. 1–6. [Google Scholar]

Articles from Ieee Journal of Biomedical and Health Informatics are provided here courtesy of Institute of Electrical and Electronics Engineers

RESOURCES