Abstract
Due to the shortage of COVID-19 viral testing kits, radiology imaging is used to complement the screening process. Deep learning based methods are promising in automatically detecting COVID-19 disease in chest x-ray images. Most of these works first train a Convolutional Neural Network (CNN) on an existing large-scale chest x-ray image dataset and then fine-tune the model on the newly collected COVID-19 chest x-ray dataset, often at a much smaller scale. However, simple fine-tuning may lead to poor performance for the CNN model due to two issues, first the large domain shift present in chest x-ray datasets and second the relatively small scale of the COVID-19 chest x-ray dataset. In an attempt to address these two important issues, we formulate the problem of COVID-19 chest x-ray image classification in a semi-supervised open set domain adaptation setting and propose a novel domain adaptation method, Semi-supervised Open set Domain Adversarial network (SODA). SODA is designed to align the data distributions across different domains in the general domain space and also in the common subspace of source and target data. In our experiments, SODA achieves a leading classification performance compared with recent state-of-the-art models in separating COVID-19 with common pneumonia. We also present initial results showing that SODA can produce better pathology localizations in the chest x-rays.
Keywords: COVID-19, medical image analysis, domain adaptation, open set domain adaptation, semi-supervised learning
1. Introduction
Since the Coronavirus disease 2019 (COVID-19) was first declared as a Public Emergency of International Concern (PHEIC) on January 30, 2020,1 it has quickly evolved from a local outbreak in Wuhan, China to a global pandemic, taking away millions of lives and causing dire economic loss worldwide. In the US, the total COVID-19 cases grew from just one confirmed on Jan 21, 2020 to over 1 million on April 28, 2020 in a span of 3 months. Despite drastic actions like shelter-in-place and contact tracing, the total cases in US kept increasing at an alarming daily rate of 20000–30000 throughout April, 2020. A key challenge for preventing and controlling COVID-19 right now is the ability to quickly, widely and effectively test for the disease, since testing is usually the first step in a series of actions to break the chains of transmission and curb the spread of the disease. COVID-19 is caused by the severe acute respiratory syndrome coronavirus 2 (SARS- CoV-2).2 By far, the most reliable diagnosis is through Reverse Transcription Polymerase Chain Reaction (RT-PCR)3 in which a sample is taken from the back of throat or nose of the patients and tested for viral RNA. Once the sample is collected, the testing process usually takes several hours and recent study reports that the sensitivity of RT-PCR is around 60–70 percent [1], which suggests that many people tested negative for the virus may actually carry it thus could infect more people without knowing it. On the other hand, the sensitivity of chest radiology imaging for COVID-19 was much higher at 97 percent as reported by [1], [2].
Due to the shortage of viral testing kits, the long period of waiting for results, and low sensitivity rate of RT-PCR, radiology imaging has been used as a complementary screening process to assist the diagnosis of COVID-19. Furthermore, radiology imaging can also provide more detailed information about the patients, e.g., pathology location, lesion size, the severity of lung involvement [3]. These insights can help doctors to timely triage patients into different risk levels, bring patients in severe conditions to ICU earlier, and saving more lives.
In recent years, with the rapid advancement in deep learning and computer vision, many breakthroughs have been developed in using Artificial Intelligence (AI) for medical imaging analysis, especially disease detection [4], [5], [6] and report generation [7], [8], [9], [10], and some AI models achieve expert radiologist-level performance [11]. Right now, with healthcare workers busy at front lines saving lives, the scalability advantage of AI-based medical imaging systems stand out more than ever. Some AI-based chest imaging systems have already been deployed in hospitals to quickly inform healthcare workers to take actions accordingly.4
Annotated datasets are required for training AI-based methods, and a small chest x-ray dataset with COVID-19 is collected recently: COVID-ChestXray [12]. Shortly after the COVID-19 outbreak, several works [13], [14], [15], [16] apply Convolutional Neural Networks (CNN) and transfer learning to detect COVID-19 cases from chest x-ray images. They first train a CNN on a large dataset like Chexpert [5] and ChestXray14 [4], and then fine-tune the model on the small COVID-19 dataset. By far, due to the lack of large-scale open COVID-19 chest x-ray imaging datasets, most works only used a very small amount of positive COVID-19 imaging samples [12]. While the reported performance metrics like accuracy and AUC-ROC are high, it is likely that these models overfit on this small dataset and may not achieve the reported performance on a different and larger COVID-19 x-ray dataset. Besides, these methods suffer a lot from label domain shift: these newly trained models lose the ability to detect common thoracic diseases like “Effusion” and “Nodule” since these labels do not appear in the new dataset. Moreover, they also ignored the visual domain shift between the two datasets. On the one hand, the large-scale datasets like ChestXray14 [4] and Chexpert [5] are collected from top U.S. health institutes like National Institutes of Health (NIH) clinical center and Stanford University, which are well-annotated and carefully processed. On the other hand, COVID-ChestXray [12] is collected from a very diverse set of hospitals around the world and they are of very different qualities and follow different standards, such as the viewpoints, aspect ratios and lighting, etc. In addition, COVID-ChestXray contains not only chest x-ray images but also CT scan images.
In order to fully exploit the limited but valuable annotated COVID-19 chest x-ray images and the large-scale chest x-ray image dataset at hand, as well as to prevent the above-mentioned drawbacks of those fine-tuning based methods, we define the problem of learning a x-ray classifier for COVID-19 from the perspective of open set domain adaptation (Definition 1) [17]. Different from traditional unsupervised domain adaptation which requires the label set of both source and target domain to be the same, the open set domain adaptation allows different domains to have different label sets. This is more suitable for our problem because COVID-19 is a new disease which is not included in the ChestXray14 or Chexpert dataset. However, since our task is to train a new classifier for COVID-19 dataset, we have to use some annotated samples. Therefore, we further propose to view the problem as a Semi-Supervised Open Set Domain Adaptation problem (Definition 2).
Under the given problem setting, we propose a novel Semi-supervised Open set Domain Adversarial network (SODA) comprised of four major components: a feature extractor
, a multi-label classifier
, domain discriminators
and
, as well as common label recognizer
. SODA learns the domain-invariant features by a two-level alignment, namely, domain level and common label level. The general domain discriminator
is responsible for guiding the feature extractor
to extract domain-invariant features. However, it has been argued that the general domain discriminator
might lead to false alignment and even negative transfer [18], [19]. For example, it is possible that the feature extractor
maps images with “Pneumonia” in the target domain and images with “Cardiomegaly” in the source domain into similar positions, which might result in the miss-classification of
. In order to solve this problem, we propose a novel common label discriminator
to guide the model to align images with common labels across domains. For labeled images,
only activates when the input image is associated with a common label. For unlabeled images, we propose a common label recognizer
to predict their probabilities of having a common label.
The main contributions of the paper are summarized as follows:
-
•
To the best of our knowledge, we are the first to tackle the problem of COVID-19 chest x-ray image classification from the perspective of domain adaptation.
-
•
We formulate the problem in a novel semi-supervised open set domain adaptation setting.
-
•
We propose a novel two-level alignment model: Semi-supervised Open set Domain Adversarial network (SODA).
-
•
We present a comprehensive evaluation to demonstrate the effectiveness of the proposed SODA.
2. Preliminary
2.1. Problem Definition
Definition 1. —
Unsupervised Open Set Domain Adaptation
Let x be the input chest x-ray image, y be the ground-truth disease label. We define
as a source domain with
labeled samples, and
as a target domain with
unlabeled samples, where the underlying label set
of the target domain might be different from the label set
of the source domain. Define
as the set of common labels shared across different domains,
and
be the sets of domain-specific labels which only appear in the source and the target domain. The task of Unsupervised Open Set Domain Adaptation is to build a model which could accurately assign common labels in
to samples
in the target domain, as well as distinguish those
belonging to
.
Definition 2. —
Semi-supervised Open Set Domain Adaptation
Given a source domain
with
labeled samples, and a target domain
consisting of
with
unlabeled samples and
with
labeled samples. The task of Semi-supervised Open Set Domain Adaptation is to build a model to assign labels from
to unlabeled samples in
.
2.2. Notations
We summarize the symbols used in the paper in Table 1.
TABLE 1. Notations.
| Symbols | Description |
|---|---|
|
set of labeled samples in the source domain |
|
set of unlabeled samples in the target domain |
|
set of labeled samples in the target domain |
|
set of labels for the source domain |
|
set of labels for the target domain |
|
set of common labels across domains |
|
set of domain-specific labels in the source domain |
|
set of domain-specific labels in the target domain |
|
set of all labels from all domains |
|
number of labeled samples in the source domain |
|
number of unlabeled samples in the target domain |
|
number of labeled samples in the target domain |
|
feature extractor |
|
multi-label classifier for
|
|
binary classifier for label (part of ) |
|
common label recognizer |
|
domain discriminator for common labels
|
|
general domain discriminator |
|
loss of multi-label classification over the entire dataset |
|
loss of over the entire dataset |
|
loss of over the entire dataset |
|
loss of over the entire dataset |
|
the coefficient of losses |
|
input image |
|
hidden features |
|
ground-truth label |
|
predicted probability |
|
predicted probability that belongs to source domain |
|
predicted probability that has common labels |
3. Methodology
3.1. Overview
An overview of the proposed Semi-supervised Open Set Domain Adversarial network (SODA) is shown in Fig. 1. Given an input image
, it will be first fed into a feature extractor
, which is a Convolutional Neural Network (CNN), to obtain its hidden feature
(green part). The binary classifier
(part of the multi-label classifier
) takes
as input, and will predict the probability
for the label
(blue part).
Fig. 1.
Overview of the proposed SODA. Given an input image
, the feature extractor
will extract its hidden features
(green part), which will be fed into a multi-label classifier
(blue part), a common label recognizer
(yellow part) and a domain discriminator
(red part) to predict the probability
of disease labels, the probability
that
is associated with a common label and the probability
that
belongs to the source domain.
,
and
denote the losses of image classification, common label classification and domain classification.
is the general domain discriminator, and
is the domain discriminator for images associated with a common label.
denotes the image classifier for the first label in the label set of the entire dataset
. Note that the gradients from
and
are not allowed to pass through
(grey arrows).
We propose a novel two-level alignment strategy for extracting the domain invariant features across the source and target domain. On one hand, we perform domain alignment (Section 3.2), which leverages a general domain discriminator
to minimize the domain-level feature discrepancy. On the other hand, we emphasize the alignment of common labels
(Section 3.3) by introducing another domain discriminator
for images associated with common labels. For labeled images in
and
, we compute loss for
and conduct back-propagation [20] during training only if the input image
is associated with a common label
. As for unlabeled data in
, we propose a common label recognizer
to predict the probability
that an image
has a common label, and use
as a weight in the losses of
and
.
3.2. Domain Alignment
Domain adversarial training [21] is the most popular method for helping feature extractor
learn domain-invariant features such that the model trained on the source domain can be easily applied to the target domain. The objective function of the domain discriminator
is:
|
where
denotes the predicted probability that the input image belongs to the source domain.
We use a Multi-Layer Perceptron (MLP) as the general domain discriminator
.
3.3. Common Label Alignment
In the field of adversarial domain adaptation, most of the existing methods only leverage a general domain discriminator
to minimize the discrepancy between the source and target domain. Such a practice ignores the label structure across domains, which will result in false alignment and even negative transfer [18], [19]. If we only use a general domain discriminator
in the open set domain adaptation setting (Definitions 1 and 2), it is possible that the feature extractor
will map the target domain images with a common label
, “Pneumonia,” and the source domain images with a specific label
, e.g., “Cardiomegaly,” to similar positions in the hidden space, which might lead to the classifier miss-classifying a “Pneumonia” image in the target domain as “Cardiomegaly”.
To address the problem of the miss-matching between the common and specific label sets, we propose a domain discriminator
to distinguish the domains for the images with a common label. For the labeled data from the source domain
and the target domain
, we know whether an image
has a common label or not, and we only calculate the loss
for
on the samples with common labels:
|
where
denotes the predicted probability that the input images is associated with a common label.
However, a large number of images in the target domain are unlabeled, and thus extra effort is required for determining whether an unlabeled image is associated with a common label. To address this problem, we propose a novel common label recognizer
to predict the probability
whether an unlabeled image has at least one common label. The probability
will be used as a weight in the loss function of
5:
|
In addition, we also use
to re-weigh unlabeled samples in
(Eq. (1)) to further emphasize the alignment of common labels:
|
Finally, the recognizer
is trained on the labeled set
via cross-entropy loss:
|
3.4. Overall Objective Function
The overall objective function of SODA is a min-max game between classifiers
,
and discriminators
,
:
|
where
,
,
and
are defined in Eqs. (5), (4), (2) and (3);
denotes the cross-entropy loss for multi-label classification;
denotes the coefficient of different loss functions.
4. Experiments
4.1. Experiment Setup
4.1.1. Dataset
Source Domain. We use ChestXray-14 [4] as the source domain dataset. This dataset is comprised of 112 120 anonymized chest x-ray images from the National Institutes of Health (NIH) clinical center. The dataset contains 14 common thoracic disease labels: “Atelectasis,” “Consolidation,” “Infiltration,” “Pneumothorax,” “Edema,” “Emphysema,” “Fibrosis,” “Effusion,” “Pneumonia,” “Pleural thickening,” “Cardiomegaly,” “Nodule,” “Mass” and “Hernia”. Target Domain The newly collected COVID-ChestXray [12] is adopted as the target domain dataset, which contains images collected from various public sources and different hospitals around the world. This dataset (by the time of this writing) contains 328 chest x-ray images in which 253 are labeled positive as the new disease “COVID-19,” whereas 61 are labeled as other well-studied “Pneumonia”.
4.1.2. Evaluation Metrics
We evaluate our model from four different perspectives. First, to test the classification performance, following the semi-supervised protocol, we randomly split the 328 x-ray images in COVID-ChestXray into 40 percent labeled set, and 60 percent unlabeled set. We report the AUC-ROC score for each label in the target domain. Second, we compute the Proxy-
Distance (PAD) [22] to evaluate models’ ability for minimizing the feature discrepancy across domains. Third, we use t-SNE to visualize the feature distributions of the target domain. Finally, we also qualitatively evaluate the models by visualizing their saliency maps.
4.1.3. Baseline Methods
We compare SODA with two types of baselines methods: fine-tuning based transfer learning models and domain adaptation models. For fine-tuning based models, we select the two most popular CNN models DenseNet121 [23] and ResNet50 [24] as our baselines. These models are first trained on the ChestXray-14 dataset and then fine-tuned on the COVID-ChestXray dataset. For domain adaptation models, we compare our model with two classic models, Domain Adversarial Neural Networks (DANN) [21] and Partial Adversarial Domain Adaptation (PADA) [25]. Note that DANN and PADA were designed for unsupervised domain adaptation, and we implement a semi-supervised version of them.
4.1.4. Implementation Details
We use DenseNet121 [23], which is pretrained on the ChestXray-14 dataset [4], as the feature extractor
for SODA. The multi-label classifier
is a one layer neural network and its activation is the sigmoid function. We use the same architecture for
,
and
: a MLP containing two hidden layers with ReLU [26] activation and an output layer. The hidden dimension for all of the modules:
,
,
and
is 1024. For fair comparison, we use the same setting of
,
and
for DANN [21] and PADA [25]. All of the models are trained by Adam optimizer [27], and the learning rate is
.
4.2. Classification Results
To investigate the effects of domain adaptation and demonstrate the performance improvement of the proposed SODA, we present the average AUC-ROC scores for all models in Table 2. Comparing the results for ResNet50 and DenseNet121, we observe that deeper and more complex models achieve better classification performance. For the effects of domain adaptation, it is obvious that the domain adaptation methods (DANN, PADA, and SODA) outperform those fine-tuning based transfer learning methods (ResNet50 and DenseNet121). Furthermore, the proposed SODA achieves higher AUC scores on both COVID-19 and Pneumonia than DANN and PADA, demonstrating the effectiveness of the proposed two-level alignment.
TABLE 2. Target Domain Average AUC-ROC Score.
4.3. Feature Visualization
We use t-SNE to project the high dimensional hidden features
extracted by DANN, PADA, and SODA to low dimensional space. The 2-dimensional visualization of the features in the target domain is presented in Fig. 2, where the red data points are image features of “Pneumonia” and the blue data points are image features of “COVID-19”. It can be observed from Fig. 2 that SODA performs the best for separating “COVID-19” from “Pneumonia,” which demonstrates the effectiveness of the proposed common label recognizer
as well as the domain discriminator for common labels
.
Fig. 2.

t-SNE visualization for DANN, PADA and SODA on the target domain.
4.4. Proxy
-Distance
Proxy
-Distance (PAD) [22] has been widely used in domain adaptation for measuring the feature distribution discrepancy between the source and target domains:
|
where
is the domain classification error (e.g., mean absolute error) of a classifier (e.g., linear SVM [28]).
Following [21], we train SVM models with different
and use the minimum error to calculate PAD. In general, a lower PAD means a better ability for extracting domain invariant features. As shown in Fig. 3, SODA has a lower PAD compared with the baseline methods, which indicates the effectiveness of the proposed two-level alignment strategy.
Fig. 3.
Proxy
-Distance.
4.5. Grad-CAM
Grad-CAM [29] is used to visualize the features extracted from all compared models. Fig. 4 shows the Grad-CAM results on seven different COVID-19 positive chest x-rays. These seven images have annotations (small arrows and box) indicating the pathology locations. We observe that ResNet50 and DenseNet121 can focus wrongly on irrelevant locations like the dark corners and edges. In contrast, domain adaptation models have better localization in general, and our SODA model gives more focused and accurate pathological locations than other models compared. In addition, we consult a professional radiologist with over 15 years of clinical experience from Wuxi People's Hospital and received positive feedback on the pathological locations as indicated by the Grad-CAM of SODA. We believe the features extracted from SODA can assist radiologists to pinpoint the suspect COVID-19 pathological locations faster and more accurately.
Fig. 4.

Grad-CAM [29] visualization for ResNet50, DenseNet121, DANN, PADA, and SODA.
5. Related Work
5.1. Domain Adaptation
Domain adaptation is an important application of transfer learning that attempts to generalize the models from source domains to unseen target domains [19], [21], [30], [31], [32], [33], [34], [35]. Adversarial training, inspired by the success of generative adversarial modeling [36], has been widely applied for promoting the learning of transfer features in image classification. It takes advantage of a domain discriminator to classify whether an image is from the source or target domains. Recently, researchers have started to study the open set domain adaptation problem, where the target domain has images that do not come from the classes in the source domain [17], [33]. Universal domain adaptation is the latest method that is proposed through using an adversarial domain discriminator and a non-adversarial domain discriminator to successfully solve this problem [33]. Although domain adaptation has been well explored, its application in medical imaging analysis, such as domain adaptation for chest x-ray images, is still under-explored.
5.2. Chest X-Ray Image Analysis
There has been substantial progress in constructing publicly available databases for chest x-ray images as well as a related line of works to identify lung diseases using these images. The largest public datasets of chest x-ray images are Chexpert [5] and ChestXray14 [4], which respectively include more than 200 000 and 100 000 chest x-ray images collected by Stanford University and National Institute of Healthcare. The creation of these datasets have also motivated and promoted the multi-label chest x-ray classification for helping the screening and diagnosis of various lung diseases. The problems of disease detection [4], [5], [6] and report generation using chest x-rays [7], [8], [9], [10] are investigated recently and have achieved much-improved results upon recently. However, there are very few attempts for studying the domain adaptation problems with the multi-label image classification problem using chest x-rays.
6. Conclusion
In this paper, in order to assist and complement the screening and diagnosing of COVID-19, we formulate the problem of COVID-19 chest x-ray image classification within a semi-supervised open set domain adaptation framework. We propose a novel deep domain adversarial neural network, Semi-supervised Open set Domain Adversarial network (SODA), which is able to align the data distributions across different domains at both domain level and common label level. Through evaluations of the classification accuracy, we show that SODA achieves better AUC-ROC scores than the recent state-of-the-art models. We further demonstrate that the features extracted by SODA is more tightly related to the lung pathology locations, and get initial positive feedback from an experienced radiologist. In practice, SODA can be generalized to any semi-supervised open set domain adaptation settings where there are a large well-annotated dataset and a small newly available dataset.
Acknowledgments
Jieli Zhou and Baoyu Jing contributed equally to this work.
Biographies

Jieli Zhou received the B.S. degree in mathematical sciences and the M.S. degree in computational data science from Carnegie Mellon University, Pittsburgh, PA, USA, in 2017 and 2018, respectively. He is currently working toward the PhD degree with the University of Michigan–Shanghai Jiao Tong University (UM-SJTU) Joint Institute, Shanghai Jiao Tong University, Shanghai, China. After graduation, he joined C3.ai as a data scientist and worked on a series of high-dimensional time-series modeling projects, such as predictive maintenance and anomaly detection. His research interests mainly include computational biology, medical image analysis, and time series analysis.

Baoyu Jing received the bachelor's degree from Beihang University, Beijing, China, in 2016 and the master's degree from the School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA, in 2018. He is currently working toward the PhD degree with Computer Science Department, University of Illinois at Urbana-Champaign. His research interests mainly include data mining, graph mining, and transfer learning.

Zeya Wang received the PhD degree in statistics from Rice University, Houston, TX, USA, in 2017. He received postdoctoral training with the University of Texas MD Anderson Cancer Center, Houston, TX, USA, where he conducted research to build statistical methods for biomedical studies. Since October 2017, he has been a scientist with Petuum Inc., Pittsburgh, PA, USA, for designing models to power large-scale industrial machine learning applications. His research interests include applied mathematics, statistical machine learning, biostatistics, and computer vision focusing on biomedical image analysis.

Hongyi Xin received the PhD degree from Computer Science Department, Carnegie Mellon University, where he worked on developing novel algorithms to improve the speed and the sensitivity of read mappers. He is currently a tenure-track assistant professor with Shanghai Jiao Tong University. He is jointly appointed by both the UM-SJTU Joint Institute and the Department of Automation. After graduation, he joined the School of Medicine, University of Pittsburgh, as a postdoc and switched focus to single-cell multiomics. He led the development of several single-cell multiomics analytical methods with multiple submissions both published in progress with UPMC Children's Hospital. His research interests include computer architecture, immunology, and cancer research.

Hanghang Tong received the MSc and PhD degrees majored in machine learning from Carnegie Mellon University in 2008 and 2009, respectively. Since August 2019, he has been an associate professor with Computer Science Department, University of Illinois at Urbana-Champaign. In August 2014, he was an assistant professor with the School of Computing, Informatics, and Decision Systems Engineering (CIDSE), Arizona State University. He was an assistant professor with Computer Science Department, City College, City University of New York, a research staff member with IBM T.J. Watson Research Center, and a postdoctoral fellow with Carnegie Mellon University. He has authored or coauthored more than 100 referred articles. His research focuses on large scale data mining for graphs and multimedia. He was the recipient of several awards, including the NSF CAREER Award (2017), the ICDM 2015 Highest-Impact Paper Award, four best paper awards (the TUP'14, the CIKM'12, the SDM'08, and the ICDM'06), five 'bests of conference' (the KDD'16, the SDM'15, the ICDM'15, the SDM'11, and the ICDM'10), and one best demo, honorable mention (the SIGMOD17). He is an associated editor for the SIGKDD Explorations (ACM), an action editor of Data Mining and Knowledge Discovery (Springer), and an associate editor for the Neurocomping (Elsevier), and was a program committee member in multiple data mining, databases, and artificial intelligence venues, including SIGKDD, SIGMOD, AAAI, WWW, and CIKM.
Footnotes
Note that gradients stop at
in the training period.
Contributor Information
Jieli Zhou, Email: zhoujieli777@hotmail.com.
Baoyu Jing, Email: baoyuj2@illinois.edu.
Zeya Wang, Email: zw17.rice@gmail.com.
Hongyi Xin, Email: hongyi.xin@sjtu.edu.cn.
Hanghang Tong, Email: htong@illinois.edu.
References
- [1].Ai T. et al. , “Correlation of chest CT and RT-PCR testing in coronavirus disease 2019 (COVID-19) in China: A report of 1014 cases,” Radiology, vol. 296, no. 2, E32–E40, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [2].Fang Y. et al. , “Sensitivity of chest CT for COVID-19: Comparison to RT-PCR,” Radiology, vol. 296, no. 2, E115–E117, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [3].Zhang K. et al. , “Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography,” Cell, vol. 181, no. 6, 1423–1433, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [4].Wang X., Peng Y., Lu L., Lu Z., Bagheri M., and Summers R. M., “ChestX-ray8: Hospital-scale chest X-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2017, pp. 2097–2106. [Google Scholar]
- [5].Irvin J. et al. , “Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison,” in Proc. AAAI Conf. on Artif. Intell., vol. 33, 2019, pp. 590–597. [Google Scholar]
- [6].Wang X., Peng Y., Lu L., Lu Z., and Summers R. M., “Tienet: Text-image embedding network for common thorax disease classification and reporting in chest x-rays,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2018, pp. 9049–9058. [Google Scholar]
- [7].Jing B., Xie P., and Xing E., “On the automatic generation of medical imaging reports,” 2017, arXiv:1711.08195.
- [8].Li Y., Liang X., Hu Z., and Xing E. P., “Hybrid retrieval-generation reinforced agent for medical image report generation,” in Proc. Adv. Neural Inf. Process. Syst., 2018, pp. 1530–1540.
- [9].Jing B., Wang Z., and Xing E., “Show, describe and conclude: On exploiting the structure information of chest x-ray reports,” in Proc. 57th Annu. Meeting Assoc. Comput. Linguistics, 2019, pp. 6570–6580. [Google Scholar]
- [10].Biswal S., Xiao C., Glass L., Westover B., and Sun J., “Clinical report auto-completion,” in Proc. Web Conf., 2020, pp. 541–550. [Google Scholar]
- [11].Lakhani P. and Sundaram B., “Deep learning at chest radiography: Automated classification of pulmonary tuberculosis by using convolutional neural networks,” Radiology, vol. 284, no. 2, pp. 574–582, 2017. [DOI] [PubMed] [Google Scholar]
- [12].Cohen J. P., Morrison P., and Dao L., “COVID-19 image data collection,” 2020, arXiv 2003.11597. [Online]. Available: https://github.com/ieee8023/covid-chestxray-dataset
- [13].Linda Wang Z. Q. L. and Wong A., “COVID-net: A. tailored deep convolutional neural network design for detection of COVID-19 cases from chest radiography images,” Sci Rep., vol. 10, 2020, Art. no. 19549. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [14].Li L. et al. , “Artificial intelligence distinguishes COVID-19 from community acquired pneumonia on chest CT,” Radiology, vol. 296, no. 2, E65–E71, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [15].Apostolopoulos I. D. and Mpesiana T. A., “COVID-19: Automatic detection from x-ray images utilizing transfer learning with convolutional neural networks,” Phys. Eng. Sci. Med., vol. 43, no. 2, pp. 635–640, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [16].Minaee S., Kafieh R., Sonka M., Yazdani S., and Soufi G. J., “Deep-COVID: Predicting COVID-19 from chest x-ray images using deep transfer learning,” 2020, arXiv:2004.09363. [DOI] [PMC free article] [PubMed]
- [17].Busto P. P. and Gall J., “Open set domain adaptation,” in Proc. IEEE Int. Conf. Comput. Vis., 2017, pp. 754–763. [Google Scholar]
- [18].Pei Z., Cao Z., Long M., and Wang J., “Multi-adversarial domain adaptation,” in Thirty-Second AAAI Conf. Artif. Intell., 2018, 3934–3941.
- [19].Wang Z., Jing B., Ni Y., Dong N., Xie P., and Xing E. P., “Adversarial domain adaptation being aware of class relationships,” 2019, arXiv:1905.11931.
- [20].Rumelhart D. E., Hinton G. E., and Williams R. J., “Learning internal representations by error propagation,” Inst. for Cognitive Science, Univ. California San Diego, La Jolla, CA, USA, Tech. Rep. 8506, 1985.
- [21].Ganin Y. et al. , “Domain-adversarial training of neural networks,” J. Mach. Learn. Res., vol. 17, no. 1, pp. 2096–2030, 2016. [Google Scholar]
- [22].Ben-David S., Blitzer J., Crammer K., and Pereira F., “Analysis of representations for domain adaptation,” in Proc. Adv. Neural Inf. Process. Syst., 2007, pp. 137–144. [Google Scholar]
- [23].Huang G., Liu Z., Van DerMaaten L., and Weinberger K. Q., “Densely connected convolutional networks,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2017, pp. 4700–4708. [Google Scholar]
- [24].He K., Zhang X., Ren S., and Sun J., “Deep residual learning for image recognition,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2016, pp. 770–778. [Google Scholar]
- [25].Cao Z., Ma L., Long M., and Wang J., “Partial adversarial domain adaptation,” in Proc. Eur. Conf. Comput. Vis. (ECCV), 2018, pp. 135–150. [Google Scholar]
- [26].Nair V. and Hinton G. E., “Rectified linear units improve restricted boltzmann machines,” in Proc. 27th Int. Conf. Mach. Learn. (ICML-10), 2010, pp. 807–814. [Google Scholar]
- [27].Kingma D. P. and Ba J., “Adam: A. method for stochastic optimization,” 2014, arXiv:1412.6980.
- [28].Cortes C. and Vapnik V., “Support-vector networks,” Mach. Learn., vol. 20, no. 3, pp. 273–297, 1995. [Google Scholar]
- [29].Selvaraju R. R., Cogswell M., Das A., Vedantam R., Parikh D., and Batra D., “Grad-cam: Visual explanations from deep networks via gradient-based localization,” in Proc. IEEE Int. Conf. Comput. Vis., 2017, pp. 618–626. [Google Scholar]
- [30].Ganin Y. and Lempitsky V., “Unsupervised domain adaptation by backpropagation,” in Proc. Int. Conf. Mach. Learn., 2015, pp. 1180–1189. [Google Scholar]
- [31].Tzeng E., Hoffman J., Saenko K., and Darrell T., “Adversarial discriminative domain adaptation,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2017, pp. 2962–2971.
- [32].Tzeng E., Hoffman J., Zhang N., Saenko K., and Darrell T., “Deep domain confusion: Maximizing for domain invariance,” 2014, arXiv:1412.3474.
- [33].You K., Long M., Cao Z., Wang J., and Jordan M. I., “Universal domain adaptation,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2019, pp. 2720–2729. [Google Scholar]
- [34].Jing B., Lu C., Wang D., Zhuang F., and Niu C., “Cross-domain labeled LDA for cross-domain text classification,” in Proc. IEEE Int. Conf. Data Mining (ICDM), 2018, pp. 187–196.
- [35].Wang D. et al. , “Coarse alignment of topic and sentiment: A unified model for cross-lingual sentiment classification,” IEEE Trans. Neural Netw. Learn. Syst., vol. 32, no. 2, pp. 736–747, Feb. 2021. [DOI] [PubMed] [Google Scholar]
- [36].Goodfellow I. et al. , “Generative adversarial nets,” in Proc. 27th Int. Conf. Neural Inf. Process. Syst., 2014, pp. 2672–2680. [Google Scholar]






























































