Abstract
Purpose: The rapid spread of the COVID-19 omicron variant virus has resulted in an overload of hospitals around the globe. As a result, many patients are deprived of hospital facilities, increasing mortality rates. Therefore, mortality rates can be reduced by efficiently assigning facilities to higher-risk patients. Therefore, it is crucial to estimate patients' survival probability based on their conditions at the time of admission so that the minimum required facilities can be provided, allowing more opportunities to be available for those who need them. Although radiologic findings in chest computerized tomography scans show various patterns, considering the individual risk factors and other underlying diseases, it is difficult to predict patient prognosis through routine clinical or statistical analysis.
Method: In this study, a deep neural network model is proposed for predicting survival based on simple clinical features, blood tests, axial computerized tomography scan images of lungs, and the patients' planned treatment. The model's architecture combines a Convolutional Neural Network and a Long Short Term Memory network. The model was trained using 390 survivors and 108 deceased patients from the Rasoul Akram Hospital and evaluated 109 surviving and 36 deceased patients infected by the omicron variant.
Results: The proposed model reached an accuracy of 87.5% on the test data, indicating survival prediction possibility. The accuracy was significantly higher than the accuracy achieved by classical machine learning methods without considering computerized tomography scan images (p-value <= 4E-5). The images were also replaced with hand-crafted features related to the ratio of infected lung lobes used in classical machine-learning models. The highest-performing model reached an accuracy of 84.5%, which was considerably higher than the models trained on mere clinical information (p-value <= 0.006). However, the performance was still significantly less than the deep model (p-value <= 0.016).
Conclusion: The proposed deep model achieved a higher accuracy than classical machine learning methods trained on features other than computerized tomography scan images. This proves the images contain extra information. Meanwhile, Artificial Intelligence methods with multimodal inputs can be more reliable and accurate than computerized tomography severity scores.
Keywords: Computer-aided diagnosis, Computerized tomography imaging, Deep neural networks, Omicron variant, Multimodal analysis
Highlights
-
•
A DNN model was proposed for predicting survival based on clinical features, axial CT scan images of lungs and the patients' planned treatment.
-
•
The proposed model achieves an accuracy of 87.3% on the test data, indicating survival prediction possibility.
-
•
The results suggest the proposed DNN model extracts and utilizes information not captured by classical machine learning methods.
1. Introduction
In late 2021, the Omicron variant of Coronavirus emerged in South Africa, leading to the spread of the disease to a wide range of countries. This unknown virus affected a significant population, overwhelming hospitals that lacked sufficient capacity to accommodate the high number of patients. Consequently, individuals at higher risk were unable to access hospital beds, Intensive Care Unit (ICU) rooms, or ventilator support, increasing mortality rates. However, due to vaccination efforts, many individuals are now considered outpatients and do not require hospitalization for treatment. Hospitals need a reliable method to determine which patients need hospitalization to optimize resources and reduce the mortality rate.
COVID-19 is not just a pulmonary infection. It can affect multiple organs, including the central or peripheral nervous system [1], [2], cardiovascular system [3], and cutaneous manifestation [4]. There are various symptoms reported, including fever in 85.6%-88.7% of patients [5], [6] to rare ones like viral myositis [7]. Also, radiologic findings and lung computerized tomography (CT) scan patterns vary from case to case, ranging from ground-glass opacity with the highest prevalence [5], [8], [9] to the rare ones like pneumothorax, or pneumomediastinum [10]. The ultimate outcome of COVID-19 is influenced by several factors, including individual risk factors, and underlying medical conditions.
COVID-19 patients demonstrate divergent clinical trajectories, characterized by a spectrum of severity in their prognostic outlooks. There is no single clinical, laboratory, imaging, or routine statistical analysis that can provide a reliable assessment of a patient's prognosis. The complex network interactions between these factors can cause clinicians to allocate resources in an inappropriate manner. Higher-risk patients can receive more intensive care if a fast and efficient method for this prediction is developed, while lower-risk patients may be managed outpatient or with minimal inpatient care. As a result of this method, hospital and health resource utilization will be maximized, and the number of mortalities resulting from the lack of required facilities will be reduced.
Artificial intelligence (AI) has made tremendous progress in recent years. AI-based systems have excelled over humans in many fields, such as games, and achieved competitive performance in others. Consequently, AI-based techniques have been used in a variety of real-life applications, such as autonomous vehicles and medical analysis. The machines are fast, fatigue-free, and produce consistent results regardless of the time of the day. In pandemics, they may provide assistance in reducing the load and managing the conditions [11], [12], [13]. Many studies have concentrated on utilizing AI-driven systems during the pandemic for different objectives. Several studies have focused on distinguishing COVID-19 patients from patients with other pulmonary diseases and healthy people from computerized tomography (CT) scan images of the chest [14], [15], [16], [13], [17], [18], [19]. Some of the introduced models can also calculate the percentage of infected regions of the lungs [13]. Another group of studies has focused on diagnosing COVID-19 from chest X-ray images to be utilized as a screening stage [10], [20], [21], [22], [23], [24], [25], [26], [27], [28], [29]. Other than single-person-based analysis, some studies have focused on large-scale predictions using the time series to help in large-scale management decisions, e.g., the number of COVID-19 cases and the mortality rate of COVID-19 [30], [31], [32], [33]. AI was also used in other applications, such as vaccine development and drug discovery for COVID-19 disease [34].
Recent studies have used AI to predict patients' outcomes using different features. In [35], they tried to predict short-term outcomes (favorable or adverse) from clinical information, including sex, age, WBC, Lymph, etc., into the volume of infected lungs of patients from their CT scans. They achieved 88% accuracy, 90% sensitivity, and 87% specificity by applying a support vector machine (SVM) algorithm on 106 patients. In [36], Li and his colleagues tried to predict COVID-19 patient mortality using clinical symptoms and CT scan images of 98 older patients. They achieved 87.5% sensitivity and 70.6% specificity. Furthermore, CT scans and their extracted features, clinical information, serology, and hematology tests of patients demonstrated promising results in predicting adverse outcomes in COVID-19 patients [37], [38], [39]. In previous studies, various methods have been used to predict mortality and hospitalization of patients, including the CT-Severity Index [40], [41], [42]. One of the drawbacks of this model is the unequal volume of the pulmonary lobes despite the equal pulmonary involvement score. Another problem in these studies is the prominent role of the interpreter. The radiologists' experience and accuracy are essential in determining the score. These factors may explain why these scoring and predictive systems are less commonly used in clinical practice.
Previous studies have addressed the question of predicting survival, however no studies have predicted survival in a particular condition. In this study, we attempted to predict outcomes based on patient health conditions at the time of hospitalization and specific treatment conditions (number of hospitalization days, number of ICU days, and number of ventilator days). Thus, it is possible to identify the minimum needed facilities for the patient to recover, or if even the best facilities are not sufficient to ensure survival. Axial CT scan images of the patient's lung were used along with clinical characteristics, including age and sex, and simple blood tests containing WBC, Hb, Plt, X.Neut, and X.Lymph. Using the above features, a deep network model was developed to predict the probability of survival. We also compared the results of the proposed method with simple machine learning methods trained on all the features of CT scan images to demonstrate their added value.
2. Materials and method
2.1. Dataset
A total of 640 patients (506 survived and 134 died) were studied at the Rasoul Akram Hospital in Tehran. We used 430 samples (336 survivors) for training, 68 samples (54 survivors) for validation, and 142 samples (116 survivors) for testing. Fig. 1a and Fig. 1b show CT scan images of two patients from this dataset. In addition to CT scan images of the patients, clinical information including age, sex, WBC, Hb, Plt, X.Neut, and X.Lymph, as well as treatment conditions including the number of hospitalization days, the number of days in the intensive care unit, and receiving ventilator support, were gathered from the medical records. The following data was collected for all patients: age, sex, WBC, Hb, Plt, and treatment conditions (X.Neut and X.Lymph were not recorded for less than 150 patients). A linear regression model was trained on fully known data in order to predict missing information. Training was conducted using the completed information.
Figure 1.
Two middle slices from the CT scan of lung of (a) a survivor infected with the Omicron variant (b) a patient who passed away after 4 days.
2.2. Workflow
By analyzing clinical data such as age and gender of patients, blood test results such as WBC, Hb, Plt, X.Neut, and X.LYmph, as well as axial lung CT scan images, the deep learning pipeline in Fig. 2 was used to predict the mortality status of COVID-19 patients upon arrival at the hospital. The model takes into account the conditions under which the patient is to be treated and predicts their mortality status. The number of hospitalization days, the number of days in the ICU, and the number of days receiving ventilator support are among these conditions. In the first step of the pipeline, CT scan images are preprocessed, and lobes are detected and extracted from images of slices using the method described in [13]. In order to obtain a general understanding of the condition of the lungs, a fixed number of slices is selected along with a sample height to cover different lung sizes with nearly equal distances. To extract features for each of the slices, images of the left and right lobes are fed independently into a deep convolutional neural network. In order to predict the patient's mortality, the extracted features, clinical features, blood test results, and treatment conditions are fed into another deep model. Finally, the results of different packs of slices are aggregated, and the highest probability of mortality in each group is assigned to the entire sample. In the preceding subsections, each part of the pipeline is described in detail.
Figure 2.
Flowchart of the pipeline for predicting survival of COVID-19 patients. The pipeline uses clinical information (age, sex), blood test results, and axial lung computerized tomography scan images as inputs. It also considers additional conditions such as hospitalization days, ICU days, and ventilator support. Computerized tomography scan images are preprocessed, and lobes are extracted. A fixed number of slices is chosen to cover different lung sizes. Deep convolutional neural networks are used to extract features from the left and right lobes of the chosen slices. Another deep model is then used to predict patient mortality using these features, clinical information, and treatment conditions. The results from different packs of slices are aggregated, and the maximum probability of mortality in the groups is assigned to the entire sample.
2.3. Choosing pack of slices for each sample
This step involves selecting a fixed number of evenly spaced slices along the lung's height for each sample. As a result of this sampling, GPU RAM is used less during training, allowing for larger batches to be used. Additionally, it ensures that the conditions for samples with different numbers of slices are equal, since the subsequent slices in the batch will have a nearly equal distance from one another. The use of different slices for each sample also prevents overfitting due to the same observations being made for each sample.
Some CT-scan images may contain fewer slices than the number selected. The problem is fixed by repeating some sample slices until they become equal to the selected number. By choosing this method over interpolating, artificial effects were avoided. Using the collection of points calculated by Equation (1), all slices are divided into partitions of nearly equal length. In this equation, refers to the number of slices of the sample, refers to the fixed number of slices chosen, and seq represents the sequence of numbers from the beginning to the end.
| (1) |
Each pack of slices is selected by adding an offset to the reference set of points. The offset can range from zero to the maximum partition length, which is the maximum difference between two consecutive points. The offset is selected at random during the training phase. During the test phase, all possible offsets are used to make the packs, and the mortality probability assigned by the network to the packs will be applied to the entire sample.
2.4. Extracting features for images of lobes
Fig. 3 illustrates the network structure used to extract features from lobe images. It is a part of the architecture employed in [13] for predicting diseased lung CT-scan images. Three consecutive slices of images related to one lobe are received by the network as a 256 x 256 x 3 array. The middle slice is used to calculate the features. Images related to the previous and next slices provide additional information regarding the continuity of the white material in the middle slice. The inputs are fed into a convolutional subnetwork (Table 1) with an output size of 32 x 32 x 256, corresponding to a 32 x 32 mesh of neurons with 256 features for each. In this network, each neuron's receptive field corresponds to a 36 x 36 patch in the input image, so the extracted features for each neuron are related to the 36 x 36 patch in the input image that the neuron is viewing. To add additional information from the vicinity of each patch, the output of the previous subnetwork is fed into a U-Net-style encoder-decoder (Table 1), in which features are extracted from a larger receptive field in the encoder part and the decoder part can increase the resolution to the input size. The distance between each patch and the lung peripheral is calculated as the minimum Manhattan distance between each pixel in the patch and the nearest peripheral pixel using the Breadth-First Search (BFS) algorithm. In order to determine the final features for each patch, we concatenate the features extracted for each patch, those extracted from its vicinity, and the minimum distance of each patch from the lung peripheral. The features for each patch are then fed into a fully connected subnetwork (Table 1) to calculate a bounded attention weight between 0 and 1 for that patch. Based on the computed weights, the final features for the lobe are calculated by using the weighted average of the patches in the lobe. Considering that this network operates only on patches, the final decision can easily be mapped to the responsible patches, and the logic can be readily understood. The network is trained in conjunction with the model for predicting mortality described in section 2.5. Section 2.6 discusses the training procedure in detail.
Figure 3.
The network structure used for extracting features from lobe images. The network calculates features for the middle slice after receiving images of one lobe in three consecutive slices. In order to extract features from a larger receptive field, a convolutional subnetwork and a U-Net encoder-decoder are applied. Additionally, the minimum distance between each patch and the lung peripheral is calculated. Final features are formed by concatenating the features from each patch, the surrounding vicinity, and the minimum distance. A bounded attention weight is then calculated for each patch based on these features. By taking the weighted average of the patch features, we determine the final features for the lobe. The network is trained in conjunction with a model for predicting mortality.
Table 1.
Layer specifications used in each subnetwork of the model for extracting features from lobe images.
| Subnetworks | Layers |
|---|---|
| Patch feature extractor | conv(64x3x3,s=1), relu, conv(64x3x3,s=1), relu, MaxPool(2x2,s=2), [Name of the output: PFE1]; conv(128x3x3,s=1), relu, conv(128x3x3,s=1), relu, MaxPool(2x2,s=2) [Name of the output: PFE2]; conv(256x3x3,s=1), relu, conv(256x3x3,s=1), relu, MaxPool(2x2,s=2) |
| Vicinity feature extractor (Encoder) | conv(64x1x1,s=1) [Name of the output: VFE1]; conv(64x3x3,s=2), conv(64x3x3,s=1), relu [Name of the output: VFE2]; conv(64x3x3,s=2), conv(64x3x3,s=1), relu |
| Vicinity feature extractor (Decoder) | TransposeConv(64x2x2,s=2), conv(64x3x3,s=1), relu; UNet concatenation with VFE2; TransposeConv(64x2x2,s=2), conv(64x3x3,s=1), relu; UNet concatenation with VFE1 |
| Bounded attention weights calculator | Conv(64x1x1,s=1), relu, Dropout(0.5), Conv(2x1x1,s=1), Softmax |
2.5. Predicting mortality
At this point, the features of the chosen slices' lobes have been calculated. As illustrated in Fig. 4, the left and right lobes of each slice are concatenated to the relative position of the slice in the lung's height, the clinical characteristics, the blood test results, and the conditions of interest for the patient's treatment. The expanded features of slices are fed sequentially to a slice iterator subnetwork (Table 2). According to the history of the observed slices, this subnetwork calculates the updated features based on the characteristics of subsequent slices. The updated features of the last slice are then fed into a fully connected subnetwork (Table 2) to determine the probability of the patient's mortality.
Figure 4.
The model used for predicting mortality for a specific treatment condition. The features extracted from the left and right lobes of the model along with clinical characteristics, blood test results, and treatment conditions are passed to a Long Short Term Memory (LSTM) network. Each time the LSTM network receives a slice from a computerized tomography scan image, it updates the features extracted from the image. To calculate the probability of mortality, the extracted features are passed to a multi-layer, fully connected neural network. The final decision is made after observing all slices.
Table 2.
Layer specifications for each subnetwork in the mortality prediction model.
| Subnetworks | Layers |
|---|---|
| Slice Iterator | LSTM(hidden=512, layers=5, (with learnable initial hidden and memory), Dropout(0.5) after each layer), BatchNorm1d(512), elu, Dropout(0.5) |
| Mortality decider | Linear(256), BatchNorm1d(256), elu, Dropout(0.5), Linear(64), Linear(2), Softmax |
For iterating over slices, this submodule uses a Long Short Term Memory (LSTM) model rather than a fully connected layer connecting all the slices' features. Fully connected layers would have had many more parameters, resulting in overfitting.
2.6. Training model
2.6.1. The loss function
Cross-entropy loss is commonly used for classification problems. Based on the predicted probability of mortality for each patient and the ground truth about the patient's status, the same loss was used as the primary loss. Deep networks have a known gradient flow problem, which increases with the depth of the network, and an LSTM network is considered a deep network because its depth affects the length of the sequence it receives. To solve the problem, an auxiliary loss was used to inject a gradient through the LSTM model's steps. Each slice's updated features were fed independently into the same decider network that was used for the last slice's features for calculating the auxiliary loss. For each of them, a mortality probability was calculated. The cross-entropy loss was then calculated using the ground truth label. We considered the auxiliary loss as the estimated likelihood for each slice and the average of each slice's losses. The LSTM model was also made to predict the mortality as soon as possible, not only based on all slices, which would result in a more robust prediction.
The loss calculated for the ith slice of the jth sample is illustrated in Equation (2). In this equation, is the cross-entropy loss estimated based on the ground truth of the jth sample, and the probability of mortality determined by the network based on the sequence of the slices from the beginning till observing the ith slice, . was considered 1 for mortality and 0 otherwise. The auxiliary loss, , was determined by averaging the cross-entropy losses of all the slices as in Equation (3). In this equation, stands for each batch sample's fixed number of slices. The primary loss was considered the cross-entropy loss of the last slice. The weighted sum of two loss terms calculated the total loss as in Equation (4). These factors were selected based on multiple trials. The total loss for each batch was calculated as the average loss of the samples in the batch.
| (2) |
| (3) |
| (4) |
2.6.2. Hyperparameters of training
In order to evaluate the model, a part of the dataset was separated as the test set. To help select hyperparameters and choose models, the remaining samples were also severed as the validation set. We trained the model using the remaining samples. The weights of the final model trained in [13] were utilized to initialize the model for extracting lobe features. During the training, the parameters of the model were frozen from the attention segment to prevent overfitting. In addition, the remaining parts of the model as well as the model used to calculate the probability of mortality were trained using the Adam optimization algorithm [43] with default parameters and an initial learning rate of 1e-4. To prevent the model from becoming biased toward the class with more samples, batches of size 8 with four positive and four negative samples were used to train the model. For each sample, forty slices were selected. A number of trials have been conducted in order to determine this number. The model was trained for 500 epochs, and the epoch model with the highest accuracy for the validation data was selected as the final model.
To expand the network's data for training, the treatment conditions were randomized in each batch. Considering the availability of more facilities, the randomized treatment conditions were superior to the original treatment. If a patient survives at a specific level of illness, he will also survive at a higher level. Nevertheless, due to the limited number of facilities, the randomized conditions were worse than the original treatment, meaning the patient had died due to receiving the worst treatment possible.
2.7. Evaluation metrics
We used the most common metric to assess the performance of a binary classifier. The survival distribution is highly skewed towards the negative class in the problem of survival prediction. Models were evaluated using sensitivity (true positive rate) and specificity (true negative rate) to show error rates for each class. For the final model selection and comparison, accuracy was selected as a metric for representing the error rate in the distribution. Furthermore, Receiver Operating Characteristic (ROC) curves were computed to demonstrate the models' general performance.
3. Experimental results and discussion
According to previous reports, the predominant CT imaging pattern in COVID-19 was GGO with occasional consolidation. The majority of patients in the gathered dataset exhibit the same pattern [44]. Bilateral lung involvement, mainly in the lower lobes, has been reported in all studies. In this study, right lung involvement was slightly more prevalent than left lung involvement, which contradicts some other studies. Pleural effusion was also significantly higher in cases with a severe form of the disease [45], [44].
3.1. Evaluations
The results of the trained model are presented in Table 3. A test accuracy of 87.5% (sensitivity of 53.8% and specificity of 95.8%) indicates that survival can be predicted from CT scan images and simple blood tests. In order to assess the added value of CT scan images, the proposed model's results were compared with classical machine learning classifiers trained over all features except CT scan images (ML-F1). As extra information, the percentage of the infected volume of the whole lung and the left and the right lobes, the volume of the entire lung and the left and the right lobes, and the average percentage of infected volume in the top 20% and also in the top 4 lobes were calculated by the infection mask and the lung mask of the model trained in [13] and added to the clinical feature set (ML-F2). With these features, machine learning models such as SVM, perceptron, random forest, and logistic regression classifiers were trained. A 5-fold cross-validation was performed over both the training and validation sets in order to choose the hyperparameters for these models. In this study, scikit-learn [46] was utilized to implement models and set hyperparameters. The hyperparameters for these models are presented in Table 4. Within the first feature set (ML-F1), the perceptron model performed best in terms of accuracy (accuracy of 72.5%, sensitivity of 61.5%, and specificity of 75%). Among the models trained using the 2nd feature set (ML-F2), Random Forest had the highest test accuracy (84.5%), sensitivity (24.9%), and specificity (97.3%). The significance of the differences in error ratios between the proposed deep pipeline and each of the other models was calculated using McNemar's test [47], which is the only test with acceptable Type I error for algorithms that can only be executed once, like deep models [48]. The proposed deep framework had a significantly lower error ratio than the other models (p-value of 0.00004 for the ML-F1 model and p-value of 0.0156 for the ML-F2 model). The difference between the ML-F1 model and the ML-F2 model was also significant (p-value = 0.006). Significant improvements in accuracy can be attributed to information related to lung volume and infection percentage. Furthermore, the proposed deep pipeline reaches a higher accuracy than both models, suggesting that CT scan images may provide more information about mortality. The comparison results are shown in Table 6.
Table 3.
Results of evaluating different models on test data.
| Accuracy (%) | Sensitivity (%) | Specificity (%) | AUC (%) | |
|---|---|---|---|---|
| ML-F1 | 72.9 | 61.5 | 75.0 | 74.55 |
| ML-F2 | 84.5 | 26.9 | 97.4 | 69.83 |
| Deep Model | 87.5 | 53.8 | 94.8 | 82.42 |
Table 4.
Machine learning methods and their hyperparameters.
| Model | Set of the hyperparameters (The unspecified parameters are the defaults used by scikit-learn) |
| SVM | kernel: [‘linear’, ‘poly’, ‘rbf’, ‘sigmoid’] |
| MLP | sizes of hidden layers: [(i, j) for i in range(1, 202, 10) for j in range(2, 6)] activation: [‘identity’, ‘logistic’, ‘tanh’, ‘relu’] solver: [‘lbfgs’, ‘sgd’, ‘adam’] learning rate: [‘constant’, ‘invscaling’, ‘adaptive’] |
| Random Forest | max depth: range [1-24] criterion: [‘gini’, ‘entropy’] N estimators: [20, 40, 60, 80, 100] |
| Logistic regression | C=[1.0, 3.6, 12.9, 46.4, 166.8, 599.5, 2154.3, 7742.6, 27825.6, 1.e05] penalty=[‘l2’, ‘l1’] |
Table 6.
Comparison of current and previous studies.
In order to demonstrate the validity of the model under the treatment conditions, the survivors should also survive under better treatment conditions, and the dead patients should die under worse treatments. The percentage of 4 random possible better conditions for the true negative patients that are also predicted as negative, as well as the percentage of 4 random possible worse conditions for the true positive patients which are also predicted as positive, is calculated for training, validation, and test sets, within the range of 0-60 days of hospitalization. This model almost always works correctly, except in some exceptional cases, as shown in Table 5.
Table 5.
Percentage of valid augmentations in true positive and true negative samples.
| Percentage of better conditions for true negative samples predicted as negative | Percentage of worse conditions for true positive samples predicted as positive | |
|---|---|---|
| Train | 100% | 100% |
| Validation | 100% | 100% |
| Test | 100% | 99.03% |
4. Limitations of the study
This study has the following limitations:
-
•
This study could benefit from more positive cases (the unsurvived patients). A larger dataset would have provided more information regarding the patterns that contribute to mortality. Because of the low mortality rate associated with COVID-19 disease in this study, collecting the dataset required more resources than were available.
-
•
More details from the treatment process, such as the medicines used during hospitalization, in addition to the clinical histories of the patients, such as previous diseases and medicines taken for other ailments, would have resulted in a more accurate mortality prediction model. However, this study demonstrated that there were common patterns resulting in mortality within the set of data used.
-
•
The results of this study would be enhanced by using a multi-center dataset, as this would ensure that they are not biased by the specific conditions at each site. Nevertheless, obtaining acceptable results from one center is a crucial step before investing resources in a larger multi-center study.
5. Conclusion
Based on the conditions at the time of the patient's admission to the hospital, this study proposed a deep learning model to predict the survival and severity of COVID-19 patients in order to guide treatment options ranging from outpatient care to ICU admission and ventilator support. Based on age, sex, and simple blood tests of WBC, Plt, Hb, X.Neut, and X.Lymph with axial CT scans of the lung, the proposed model achieved an accuracy of 87.3%. This model had a specificity of 94.8%, which is more significant than sensitivity, which means that the patients who survived were correctly predicted. Thus, each patient's survival probability can be predicted based on all the possible conditions, and the minimum facilities required for survival can be determined. As a result, limited facilities will be used more efficiently, and fewer patients will die as a result of a lack of resources. In previous studies, machine learning models such as support vector machines and deep neural networks have been used to predict mortality using clinical signs such as fever and cough severity. When hospitals are overloaded with patients, predicting mortality is useful. As part of this study, we also included treatment conditions in order to predict mortality based on the planned treatment. By using the proposed model, hospitals can manage their facilities and use them for patients whose survival depends on them. The proposed model was also compared with classical machine learning models trained over all the features, CT scan images and even the classical models trained over all the mentioned features were compared, as well as some features related to lung volume and the percentage of infected volume calculated with another deep model from CT scan images. Although the classical models that used the features extracted from CT scan images as input reached a higher accuracy than the first group, indicating that the percentage of infection and the size of the lung lobes affect survival. As a result of the proposed deep model's higher accuracy, it is evident that lung CT scan images contain more information about survival. It is crucial to study CT scan images of lungs better, as there may be signs in the lungs of the patients that prevent them from receiving more intense treatment, which may result in a better treatment for COVID-19 patients. Although CT severity scores are used to predict a patient's condition, they are not accurate or sufficient to assess the patient's condition. Therefore, AI can improve the accuracy and reliability of mortality assessment methods when combined with other clinical inputs. Artificial intelligence systems would perform better, be more reliable, and be more generalizable if they used a larger dataset with vast diversity, included the patient's history of diseases, and performed extra clinical tests.
Ethical approval
The study has been approved by the Iran University of Medical Sciences (IUMS) and the Ethical Code IR.IUMS.REC.1399.008 authorized the use of CT-scan images. The patients consented to have their internal scans (e.g. x-rays, MRIs, CTs, ultrasound) images published.
Abbreviations
Computerized tomography (CT) Long Short Term Memory (LSTM) Intensive Care Unit (ICU) Artificial Intelligence (AI) Area Under the Curve (AUC) Long Short Term Memory (LSTM) Support Vector Machine (SVM) Multi-Layer Perceptron (MLP)
CRediT authorship contribution statement
Mahyar Ghafoori: Conceptualization, Data curation, Formal analysis, Methodology, Supervision. Mehrab Hamidi: Methodology, Validation, Writing – original draft. Rassa Ghavami Modegh: Formal analysis, Methodology, Validation, Writing – original draft. Alireza Aziz-Ahari: Data curation, Supervision. Neda Heydari: Data curation, Resources. Zeynab Tavafizadeh: Data curation, Resources. Omid Pournik: Conceptualization, Methodology, Supervision. Sasan Emdadi: Data curation. Saeed Samimi: Data curation. Amir Mohseni: Data curation. Mohammadreza Khaleghi: Data curation, Resources. Hamed Dashti: Conceptualization, Methodology, Writing – original draft. Hamid R. Rabiee: Conceptualization, Formal analysis, Methodology, Supervision, Writing – review & editing.
Declaration of Competing Interest
The authors declare the following financial interests/personal relationships which may be considered as potential competing interests:
Hamid R. Rabiee reports financial support was provided by Iran National Science Foundation. If there are other authors, they declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Acknowledgements
This study was partially funded by IR National Science Foundation (INSF) Grant No. 96006077 and ISTI grant number 11/41701 (Hamid R. Rabiee was the recipient of both grants).
Contributor Information
Mahyar Ghafoori, Email: mahyarghafoori@gmail.com.
Mehrab Hamidi, Email: mehrabhamidi78@gmail.com.
Rassa Ghavami Modegh, Email: ghavami@ce.sharif.edu.
Alireza Aziz-Ahari, Email: azizahari.ar@iums.ac.ir.
Neda Heydari, Email: Nedaheydarii90@yahoo.com.
Zeynab Tavafizadeh, Email: zaynabtavafi@gmail.com.
Omid Pournik, Email: pournik.o@iums.ac.ir.
Sasan Emdadi, Email: sasan.emdadi@gmail.com.
Saeed Samimi, Email: sadsamm95@gmail.com.
Amir Mohseni, Email: amir.mohseni@sharif.edu.
Mohammadreza Khaleghi, Email: Mkh712003@yahoo.com.
Hamed Dashti, Email: dashtih@ce.sharif.edu.
Hamid R. Rabiee, Email: rabiee@sharif.edu.
Data availability
The datasets associated with this study have not been deposited into a publicly available repository. However, these datasets are available from the corresponding authors upon reasonable request.
References
- 1.Mao L., Wang M., Chen S., He Q., Chang J., Hong C., Zhou Y., Wang D., Li Y., Jin H., Hu B. 2020. Neurological manifestations of hospitalized patients with COVID-19 in Wuhan, China: a retrospective case series study. [Google Scholar]
- 2.Wu Y., Xu X., Chen Z., Duan J., Hashimoto K., Yang L., Liu C., Yang C. Nervous system involvement after infection with Covid-19 and other coronaviruses. Brain Behav. Immun. 2020;87:18–22. doi: 10.1016/j.bbi.2020.03.031. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Zheng Y.-Y., Ma Y.-T., Zhang J.-Y., Xie X. Covid-19 and the cardiovascular system. Nat. Rev. Cardiol. 2020;17:259–260. doi: 10.1038/s41569-020-0360-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Casas G., Català A., Hernández G.C., Rodríguez-Jiménez P., Fernández-Nieto D., Lario A.R.-V., Fernández I.N., Ruiz-Villaverde R., Falkenhain-López D., Velasco M.L., García-Gavín J., Baniandrés O., González-Cruz C., Morillas-Lahuerta V., Cubiró X., Nart I.F., Selda-Enriquez G., Romaní J., Fustà-Novell X., Melian-Olivera A., Riesco M.R., Burgos-Blasco P., Ortigosa J.S., Rodriguez M.F., García-Doval I. Classification of the cutaneous manifestations of COVID -19: a rapid prospective nationwide consensus study in Spain with 375 cases. Br. J. Dermatol. 2020;183:71–77. doi: 10.1111/bjd.19163. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Lovato A., De Filippis C. Clinical presentation of Covid-19: a systematic review focusing on upper airway symptoms. Ear, Nose, Throat J. 2020;99:569–576. doi: 10.1177/0145561320920762. [DOI] [PubMed] [Google Scholar]
- 6.Rodriguez-Morales A.J., Cardona-Ospina J.A., Gutiérrez-Ocampo E., Villamizar-Peña R., Holguin-Rivera Y., Escalera-Antezana J.P., Alvarado-Arnez L.E., Bonilla-Aldana D.K., Franco-Paredes C., Henao-Martinez A.F., Paniz-Mondolfi A., Lagos-Grisales G.J., Ramírez-Vallejo E., Suárez J.A., Zambrano L.I., Villamil-Gómez W.E., Balbin-Ramon G.J., Rabaan A.A., Harapan H., Dhama K., Nishiura H., Kataoka H., Ahmad T., Sah R. Clinical, laboratory and imaging features of COVID-19: a systematic review and meta-analysis. Trav. Med. Infect. Dis. 2020;34 doi: 10.1016/j.tmaid.2020.101623. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Zhang Q., Shan K.S., Minalyan A., O'Sullivan C., Nace T. A rare presentation of coronavirus disease 2019 (Covid-19) induced viral myositis with subsequent rhabdomyolysis. Cureus. 2020;12 doi: 10.7759/cureus.8074. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Salehi S., Abedi A., Balakrishnan S., Gholamrezanezhad A. Coronavirus disease 2019 (Covid-19): a systematic review of imaging findings in 919 patients. Am. J. Roentgenol. 2020;215:87–93. doi: 10.2214/AJR.20.23034. [DOI] [PubMed] [Google Scholar]
- 9.Caruso D., Zerunian M., Polici M., Pucciarelli F., Polidori T., Rucci C., Guido G., Bracci B., Dominicis C.D., Laghi A. Chest CT features of COVID-19 in Rome, Italy. Radiology. 2020;296:E79–E85. doi: 10.1148/radiol.2020201237. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Wang W., Gao R., Zheng Y., Jiang L. COVID-19 with spontaneous pneumothorax, pneumomediastinum and subcutaneous emphysema. J. Travel Med. 2020;27 doi: 10.1093/jtm/taaa062. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Zhou S.K., Greenspan H., Davatzikos C., Duncan J.S., Van Ginneken B., Madabhushi A., Prince J.L., Rueckert D., Summers R.M. A review of deep learning in medical imaging: imaging traits, technology trends, case studies with progress highlights, and future promises. Proc. IEEE. 2021;109:820–838. doi: 10.1109/JPROC.2021.3054390. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Modegh R.G., Salimi A., Ilami S., Dehqan A.H., Dashti H., Javanmard S.H., Ghanaati H., Rabiee H.R. The Science Behind the COVID Pandemic and Healthcare Technology Solutions. Springer; 2022. Covid-19 diagnosis with artificial intelligence; pp. 353–378. [Google Scholar]
- 13.Modegh R.G., Hamidi M., Masoudian S., Mohseni A., Lotfalinezhad H., Kazemi M.A., Moradi B., Ghafoori M., Motamedi O., Pournik O., Rezaei-Kalantari K., Manteghinezhad A., Javanmard S.H., Nezhad F.A., Enhesari A., Kheyrkhah M.S., Eghtesadi R., Azadbakht J., Aliasgharzadeh A., Sharif M.R., Khaleghi A., Foroutan A., Ghanaati H., Dashti H., Rabiee H.R. 2020. Accurate and rapid diagnosis of Covid-19 pneumonia with batch effect removal of chest ct-scans and interpretable artificial intelligence. [Google Scholar]
- 14.Li L., Qin L., Xu Z., Yin Y., Wang X., Kong B., Bai J., Lu Y., Fang Z., Song Q., et al. Using artificial intelligence to detect Covid-19 and community-acquired pneumonia based on pulmonary ct: evaluation of the diagnostic accuracy. Radiology. 2020;296:E65–E71. doi: 10.1148/radiol.2020200905. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Gunraj H., Sabri A., Koff D., Wong A. COVID-net ct-2: enhanced deep neural networks for detection of Covid-19 from chest ct images through bigger, more diverse learning. Front. Med. 2022;8:3126. doi: 10.3389/fmed.2021.729287. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Harmon S.A., Sanford T.H., Xu S., Turkbey E.B., Roth H., Xu Z., Yang D., Myronenko A., Anderson V., Amalou A., et al. Artificial intelligence for the detection of Covid-19 pneumonia on chest ct using multinational datasets. Nat. Commun. 2020;11:4080. doi: 10.1038/s41467-020-17971-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Wang X., Deng X., Fu Q., Zhou Q., Feng J., Ma H., Liu W., Zheng C. A weakly-supervised framework for Covid-19 classification and lesion localization from chest ct. IEEE Trans. Med. Imaging. 2020;39:2615–2625. doi: 10.1109/TMI.2020.2995965. [DOI] [PubMed] [Google Scholar]
- 18.Gunraj H., Wang L., Wong A. Covidnet-ct: a tailored deep convolutional neural network design for detection of Covid-19 cases from chest ct images. Front. Med. 2020;7 doi: 10.3389/fmed.2020.608525. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Jin C., Chen W., Cao Y., Xu Z., Tan Z., Zhang X., Deng L., Zheng C., Zhou J., Shi H., et al. Development and evaluation of an artificial intelligence system for Covid-19 diagnosis. Nat. Commun. 2020;11:5088. doi: 10.1038/s41467-020-18685-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Das D., Santosh K., Pal U. Truncated inception net: Covid-19 outbreak screening using chest x-rays. Phys. Eng. Sci. Med. 2020;43:915–925. doi: 10.1007/s13246-020-00888-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Basu S., Mitra S., Saha N. 2020 IEEE Symposium Series on Computational Intelligence (SSCI) IEEE; 2020. Deep learning for screening covid-19 using chest x-ray images; pp. 2521–2527. [Google Scholar]
- 22.Li X., Li C., Zhu D. 2020. COVID-mobilexpert: on-device Covid-19 screening using snapshots of chest x-ray. [Google Scholar]
- 23.Castiglioni I., Ippolito D., Interlenghi M., Monti C.B., Salvatore C., Schiaffino S., Polidori A., Gandola D., Messa C., Sardanelli F. Artificial intelligence applied on chest x-ray can aid in the diagnosis of Covid-19 infection: a first experience from lombardy, Italy. MedRxiv. 2020-04;2020 doi: 10.1186/s41747-020-00203-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Hu T., Khishe M., Mohammadi M., Parvizi G.-R., Karim S.H.T., Rashid T.A. Real-time Covid-19 diagnosis from x-ray images using deep cnn and extreme learning machines stabilized by chimp optimization algorithm. Biomed. Signal Process. Control. 2021;68 doi: 10.1016/j.bspc.2021.102764. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Wu C., Khishe M., Mohammadi M., Taher Karim S.H., Rashid T.A. Evolving deep convolutional neutral network by hybrid sine–cosine and extreme learning machine for real-time Covid19 diagnosis from x-ray images. Soft Comput. 2021:1–20. doi: 10.1007/s00500-021-05839-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Shankar K., Mohanty S.N., Yadav K., Gopalakrishnan T., Elmisery A.M. Automated Covid-19 diagnosis and classification using convolutional neural network with fusion based feature extraction model. Cogn. Neurodyn. 2021:1–14. doi: 10.1007/s11571-021-09712-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Shome D., Kar T., Mohanty S.N., Tiwari P., Muhammad K., AlTameem A., Zhang Y., Saudagar A.K.J. Covid-transformer: interpretable Covid-19 detection using vision transformer for healthcare. Int. J. Environ. Res. Public Health. 2021;18 doi: 10.3390/ijerph182111086. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Saffari A., Khishe M., Mohammadi M., Hussein Mohammed A., Rashidi S. Dcnn-fuzzywoa: artificial intelligence solution for automatic detection of Covid-19 using x-ray images. Comput. Intell. Neurosci. 2022 doi: 10.1155/2022/5677961. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Cai C., Gou B., Khishe M., Mohammadi M., Rashidi S., Moradpour R., Mirjalili S. Improved deep convolutional neural networks using chimp optimization algorithm for Covid19 diagnosis from the x-ray images. Expert Syst. Appl. 2023;213 doi: 10.1016/j.eswa.2022.119206. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Sah S., Surendiran B., Dhanalakshmi R., Mohanty S.N., Alenezi F., Polat K. Forecasting Covid-19 pandemic using prophet, arima, and hybrid stacked lstm-gru models in India. Comput. Math. Methods Med. 2022;2022 doi: 10.1155/2022/1556025. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Sharma N., Yadav S., Mangla M., Mohanty A., Satpathy S., Mohanty S.N., Choudhury T. Geospatial multivariate analysis of Covid-19: a global perspective. GeoJournal. 2021:1–15. doi: 10.1007/s10708-021-10520-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Dash S., Chakravarty S., Mohanty S.N., Pattanaik C.R., Jain S. A deep learning method to forecast Covid-19 outbreak. New Gener. Comput. 2021;39:515–539. doi: 10.1007/s00354-021-00129-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Satpathy S., Mangla M., Sharma N., Deshmukh H., Mohanty S. Predicting mortality rate and associated risks in Covid-19 patients. Spat. Inf. Res. 2021;29:455–464. [Google Scholar]
- 34.Hwang W., Lei W., Katritsis N.M., MacMahon M., Chapman K., Han N. Current and prospective computational approaches and challenges for developing Covid-19 vaccines. Adv. Drug Deliv. Rev. 2021;172:249–274. doi: 10.1016/j.addr.2021.02.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Matos J., Paparo F., Mussetto I., Bacigalupo L., Veneziano A., Bernardi S.P., Biscaldi E., Melani E., Antonucci G., Cremonesi P., Lattuada M., Pilotto A., Pontali E., Rollandi G.A. Evaluation of novel coronavirus disease (Covid-19) using quantitative lung CT and clinical data: prediction of short-term outcome. Eur. Radiol. Exp. 2020;4 doi: 10.1186/s41747-020-00167-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Li Y., Yang Z., Ai T., Wu S., Xia L. Association of “initial CT” findings with mortality in older patients with coronavirus disease 2019 (Covid-19) Eur. Radiol. 2020;30:6186–6193. doi: 10.1007/s00330-020-06969-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Zhang R., Ouyang H., Fu L., Wang S., Han J., Huang K., Jia M., Song Q., Fu Z. CT features of SARS-CoV-2 pneumonia according to clinical presentation: a retrospective analysis of 120 consecutive patients from Wuhan city. Eur. Radiol. 2020;30:4417–4426. doi: 10.1007/s00330-020-06854-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Yuan M., Yin W., Tao Z., Tan W., Hu Y. Association of radiologic findings with mortality of patients infected with 2019 novel coronavirus in Wuhan, China. PLoS ONE. 2020;15 doi: 10.1371/journal.pone.0230548. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Colombi D., Bodini F.C., Petrini M., Maffi G., Morelli N., Milanese G., Silva M., Sverzellati N., Michieletti E. Well-aerated lung on admitting chest CT to predict adverse outcome in Covid-19 pneumonia. Radiology. 2020;296:E86–E96. doi: 10.1148/radiol.2020201433. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Francone M., Iafrate F., Masci G.M., Coco S., Cilia F., Manganaro L., Panebianco V., Andreoli C., Colaiacomo M.C., Zingaropoli M.A., et al. Chest ct score in Covid-19 patients: correlation with disease severity and short-term prognosis. Eur. Radiol. 2020;30:6808–6817. doi: 10.1007/s00330-020-07033-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Feng Z., Yu Q., Yao S., Luo L., Zhou W., Mao X., Li J., Duan J., Yan Z., Yang M., et al. Early prediction of disease progression in Covid-19 pneumonia patients with chest ct and clinical characteristics. Nat. Commun. 2020;11:1–9. doi: 10.1038/s41467-020-18786-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Aziz-Ahari A., Keyhanian M., Mamishi S., Mahmoudi S., Bastani E.E., Asadi F., Khaleghi M. Chest ct severity score: assessment of Covid-19 severity and short-term prognosis in hospitalized Iranian patients. Wien. Med. Wochenschr. 2022;172:77–83. doi: 10.1007/s10354-022-00914-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Kingma D.P., Ba Adam J. A method for stochastic optimization. 2014. arXiv:1412.6980 arXiv preprint.
- 44.Yang R., Li X., Liu H., Zhen Y., Zhang X., Xiong Q., Luo Y., Gao C., Zeng W. Chest CT severity score: an imaging tool for assessing severe Covid-19. Radiology. 2020;2 doi: 10.1148/ryct.2020200047. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Sun D., Li H., Lu X.-X., Xiao H., Ren J., Zhang F.-R., Liu Z.-S. Clinical features of severe pediatric patients with coronavirus disease 2019 in Wuhan: a single center's observational study. World J. Pediatr. 2020;16:251–259. doi: 10.1007/s12519-020-00354-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Pedregosa F., Varoquaux G., Gramfort A., Michel V., Thirion B., Grisel O., Blondel M., Müller A., Nothman J., Louppe G., Prettenhofer P., Weiss R., Dubourg V., Vanderplas J., Passos A., Cournapeau D., Brucher M., Perrot M. Édouard duchesnay, scikit-learn: machine learning in python. J. Mach. Learn. Res. 2011;12:2825–2830. [Google Scholar]
- 47.McNemar Q. Note on the sampling error of the difference between correlated proportions or percentages. Psychometrika. 1947;12:153–157. doi: 10.1007/BF02295996. [DOI] [PubMed] [Google Scholar]
- 48.Dietterich T.G. Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput. 1998;10:1895–1923. doi: 10.1162/089976698300017197. [DOI] [PubMed] [Google Scholar]
- 49.Zhu J.S., Ge P., Jiang C., Zhang Y., Li X., Zhao Z., Zhang L., Duong T.Q. Deep-learning artificial intelligence analysis of clinical variables predicts mortality in Covid-19 patients. J. Am. Coll. Emerg. Physicians Open. 2020;1:1364–1373. doi: 10.1002/emp2.12205. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Li X., Ge P., Zhu J., Li H., Graham J., Singer A., Richman P.S., Duong T.Q. Deep learning prediction of likelihood of icu admission and mortality in Covid-19 patients using clinical variables. PeerJ. 2020;8 doi: 10.7717/peerj.10337. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Pourhomayoun M., Shakibi M. Predicting mortality risk in patients with Covid-19 using machine learning to help medical decision-making. Smart Health. 2021;20 doi: 10.1016/j.smhl.2020.100178. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Näppi J.J., Uemura T., Watari C., Hironaka T., Kamiya T., Yoshida H. U-survival for prognostic prediction of disease progression and mortality of patients with Covid-19. Sci. Rep. 2021;11:9263. doi: 10.1038/s41598-021-88591-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The datasets associated with this study have not been deposited into a publicly available repository. However, these datasets are available from the corresponding authors upon reasonable request.




