You are viewing a javascript disabled version of the site. Please enable Javascript for this site to function properly.
Go to headerGo to navigationGo to searchGo to contentsGo to footer
In content section. Select this link to jump to navigation

Severity detection of COVID-19 infection with machine learning of clinical records and CT images

Abstract

BACKGROUND:

Coronavirus disease 2019 (COVID-19) is a deadly viral infection spreading rapidly around the world since its outbreak in 2019. In the worst case a patient’s organ may fail leading to death. Therefore, early diagnosis is crucial to provide patients with adequate and effective treatment.

OBJECTIVE:

This paper aims to build machine learning prediction models to automatically diagnose COVID-19 severity with clinical and computed tomography (CT) radiomics features.

METHOD:

P-V-Net was used to segment the lung parenchyma and then radiomics was used to extract CT radiomics features from the segmented lung parenchyma regions. Over-sampling, under-sampling, and a combination of over- and under-sampling methods were used to solve the data imbalance problem. RandomForest was used to screen out the optimal number of features. Eight different machine learning classification algorithms were used to analyze the data.

RESULTS:

The experimental results showed that the COVID-19 mild-severe prediction model trained with clinical and CT radiomics features had the best prediction results. The accuracy of the GBDT classifier was 0.931, the ROUAUC 0.942, and the AUCPRC 0.694, which indicated it was better than other classifiers.

CONCLUSION:

This study can help clinicians identify patients at risk of severe COVID-19 deterioration early on and provide some treatment for these patients as soon as possible. It can also assist physicians in prognostic efficacy assessment and decision making.

1.Introduction

Coronavirus disease 2019 (COVID-19) is a deadly viral infection that has spread rapidly around the world since its outbreak in 2019. According to a report from the World Health Organization, as of 25 May 2022, there have been more than 524 million confirmed cases of COVID-19, including more than 6.28 million deaths 0. The severity of COVID-19 can be classified into the following categories: mild, ordinary, severe, and critical [2]. Patients with severe COVID-19 may suffer from massive alveolar damage and respiratory failure, leading to death [3]. Therefore, early classification of COVID-19 and effective targeted treatment for critically ill patients can reduce the risk of complications. Early and automatic diagnosis will help countries all over the world provide timely treatment and quarantine. Hospitals can also offer more professional treatment for severe COVID-19 patients.

Nucleic acid screening, clinical features, epidemiological features, and imaging findings are basic for diagnosing COVID-19 [4]. After comparing the diagnostic effects of RT-PCR tests and chest computed tomography (CT) on the initial negative to positive diagnosis, AI [5] concluded that chest CT detection is faster. Chest CT is recommended as a routine test for surveillance and diagnosis of COVID-19 due to imaging features such as ground-glass opacity and consolidation on Chest CT used to determine SARS-CoV-2 infection-associated pneumonia [6]. A chest CT also assists physicians in identifying the early stages of lung infection [8] and is beneficial in helping governments to establish greater public health surveillance and response systems [10]. A combined assessment using clinical records and imaging features allows for a more accurate early diagnosis of patients with COVID-19.

Although some studies have constructed machine learning (ML) prediction models, to our knowledge, the current diagnosis of COVID-19 does not achieve satisfactory accuracy. Terwangne et al. [11] used 295 RT-PCR-positive COVID-19 patients data to develop a model based on the Bayesian network to predict the severity grading of COVID-19 patients and finally obtained a AUC of 83.8%. Yao et al. [12] developed a model to predict the severity of COVID-19 based on the SVM algorithm using data from 137 COVID-19 patients (75 severe, 62 mild) and finally obtained an accuracy of 81.48%.

Liang et al. [13] developed a model for predicting the severity of COVID-19 patients based on the SVM algorithm using data from 172 patients (60 severe) and finally obtained an average accuracy of 91.83%. Liang et al. [14] used data from 1590 (131 severe) patients to develop a model based on the LR algorithm for predicting clinical risk scores for the occurrence of critical illness in hospitalized patients with COVID-19, ultimately obtaining an AUC of 88% on the validation set.

Zhu et al. [15] used data from 127 patients (16 severe groups) to develop a model to assess the severity of infection in COVID-19 patients based on the LR algorithm, ultimately obtaining an AUC of 90.0%.

The above studies related to the severity classification of COVID-19 were based on clinical features only. They did not consider the influence of CT image feature factors on the classification of mild to severe disease. Because of this, we use clinical examination data and CT radiomics features to build a prediction model for mild and severe COVID-19 patients. The study aims to segment the region of interest in the lung, extract CT radiomics features based on the segmentation results, and train a machine learning prediction model using the clinical records and the extracted image features. In the following sections, data collection and modeling methods will be described first, followed by developing COVID-19 patient severity detection models using traditional and ensemble machine learning algorithms.

2.Materials and methods

The clinical and chest CT data used for this retrospective study were collected by the Shanghai Public Health Clinical Center from 20 January to 29 May 29 2020. The Ethics Committee approved the retrospective study of the Shanghai Public Health Clinical Center. Figure 1 shows the overall flowchart we developed for detecting the severity of COVID-19. Firstly, SP-V-Net was used to segment the CT images to obtain the lung contours. Radiomics was used to extract the lung image features, which were combined with clinical features to build a mild-severe diagnostic model after feature screening.

The 427 clinical data collected were used for this study, including 387 patients with mild COVID-19 (mean age, 40.26 ± 15.16 years; range, 22–88 years; male 213, female 174), 40 subjects with severe COVID-19 (mean age, 60.58 ± 14.69 years; range, 23–73 years; male 28, female 12), and seven eventually died. The relevant scan parameters for chest CT are as follows: slice number, 49–90; matrix, 512 × 512; pixel size, 0.590–1.0 mm; slice thickness, 5 mm. In the next diagnostic task of COVID-19 severity, all subjects were divided into mild and severe COVID-19 groups for binary classification. The baseline characteristics of patients with mild and severe COVID-19 are shown in Table 1.

Table 1

Baseline characteristics of COVID-19 patients with mild and severe disease

FeaturesMildSevereP value
PO214.38± 4.4610.91± 5.540.084
Age40.26± 15.1660.58± 14.690.263
Lactate dehydrogenase (LDH)218.98± 68.56347.85± 138.830.000
C-reactive protein (CRP)15.16± 19.5755.91± 54.910.000
APTT38.45± 5.7341.86± 6.910.030
eGFR117.36± 24.56104.16± 28.390.150
Blood sodium139.91± 2.54137.12± 4.000.000
Total cholesterol4.25± 0.904.19± 0.980.907
PRO-BNP60.05± 103.20179.28± 363.390.000
PCO25.43± 0.625.37± 0.890.000
Urea4.40± 1.395.72± 3.790.000
White blood cells (WBC)5.48± 2.085.43± 2.150.204
D-dimer0.54± 1.412.19± 4.520.000
Lymphocyte count1.44± 0.571.03± 0.470.174

Figure 1.

Flowchart of our approach to building the COVID-19 severity diagnosis model.

Flowchart of our approach to building the COVID-19 severity diagnosis model.

By observing the basic situation of the clinical dataset, we found that some clinical features had missing data, and the number of missing data and the proportion of missing data for different features are shown in Table 2. To reduce the impact of missing data on the experimental results, we use the Multivariate Imputation by Chained Equation (MICE) [16] method to interpolate the missing data. MICE is a multiple interpolation method that works iteratively and can resolve the uncertainty of missing values by creating multiple interpolations. The data can be interpolated by variable by specifying an interpolation model for each variable. The MICE interpolation technique has better robustness and better accounts for uncertainty and is selected to treat missing data values.

Table 2

Features and missing values in the dataset

FeatureNumber of missing data values% of missing data values on the total
C-reactive protein (CRP)12228.18
PRO-BNP10023.09
PCT255.77
Total cholesterol153.46
PO2153.46
PCO2153.46
eGFR61.39
Urea51.15
Lactate dehydrogenase (LDH)51.15
Blood sodium40.92
APTT30.69
D-Dimer30.69
White blood cells (WBC)10.23
Tlymphocyte10.23

2.1CT radiomics features extraction

SP-V-Net [17] is a lung parenchyma segmentation model based on image deformation. The advantage of this model is that it uses 3D V-Net for end-to-end lung extraction and combines the spatial transform network (STN) module and prior shape knowledge to refine the V-Net output results so that the final segmentation results are closer to the ground-truth label. First, the threshold segmentation results were used as the lung lobes shape prior, combined with the gold standard data to train the SP-V-Net segmentation model. Second, lung lobes of each patient were segmented automatically by SP-V-Net; experienced operators confirmed the CT image segmentation results. Finally, we used the binary image of the segmentation result to multiply the original image to obtain all CT lung ROIs of 427 patients. Radiomics was used to extract the CT radiomics features from lung ROIs for machine learning; 120 features presented by Zwanenburg et al. [18] were measured. All of them may be related to COVID-19 classification are extracted for our analysis.

2.2Feature selection

The feature selection algorithm effectively reduces the feature number and also helps to improve accuracy in many cases [19]. Feature selection algorithms can effectively remove those unrelated features [20], which usually enhances the model’s generalization performance. We used the random forest to rank features and selected the top-ranked features. Due to the random nature of the Random Forest algorithm, we trained the model several times, choosing a certain number of features each time and using the intersection of the results of multiple experiments as the last selected features.

2.3Model selection

By comparing different classes of machine learning classifiers, we consider training the following classifier model to predict the severity of COVID-19.

AdaBoost (Adaptive Boosting) [21] is an iterative algorithm that trains different weak classifiers for the same training set by increasing the weights of misclassified data and decreasing the importance of correctly classified data. Finally, AdaBoost combines these weak classifiers linearly to form a robust classifier. GBDT (Gradient Boosted Decision Tree) [22] is an ensemble algorithm that produces a weak classifier in each iteration by multiple iterations. The total classifier is obtained by weighting and aggregating the weak classifiers, which improves prediction accuracy. MLP (Multilayer Perceptron) [23] is a feed-forward ANN containing at least three layers of neurons, trained with back-propagation supervised learning techniques, which can identify not linearly separable data. XGBoost (Extreme gradient boosting) [24] is a scalable tree-boosting machine learning system in all scenarios and can solve real-world scale problems using minimal resources. KNN (K-nearest neighbors) [25] is a simple and effective classification algorithm that performs classification by measuring the distance between different feature values. LR (Logistic Regression) [26] is a classical classification method in supervised learning and is often used to deal with regression problems in which the dependent variable is categorical. Logistic regression is often used to analyze medical research risk factors for a particular disease. NB (Naive Bayes) [27] is one of the most effective inductive learning algorithms in data mining and machine learning, and it has surprising performance in classification. RF (Random Forest) [28] is an ensemble algorithm that can solve the data imbalance problem by decision tree voting to get the final prediction and can be used for feature selection by providing the relative importance of different features in the classification process.

2.4Data sampling methods

Data imbalance is one of the current challenges in data analysis, which usually leads to over-fitting models. To further describe the data imbalance, we represent the minority class sample by using P for the minority class sample and N for the majority class sample, and define the following:

(1)
P={(x,y)|y=1},N={(x,y)|y=0}

When |N||P| represents the height imbalance problem. The imbalance ratio is used to describe the level of imbalance in the data set and is defined as the ratio of the number of majority samples to the number of minority samples.

(2)
Imblanced Ratio(IR)=n𝑚𝑎𝑗𝑜𝑟𝑡𝑦nmin𝑜𝑟𝑖𝑡𝑦=|N||P|

There was also a data imbalance in this study, with 387 cases of mild disease and 40 cases of severe disease in the data we collected, with a data imbalance ratio of 9.675. To further address the effect of data imbalance on the experimental results, we sampled the data in three different ways (Under-sampling, Over-sampling, and Combination of over- and under-sampling methods) for all the data separately.

2.4.1Under-sampling methods

  • ClusterCentroids [29] use KMeans to cluster each sample type separately, replacing the entire cluster of samples using the center of mass.

  • RandUnder (Random Under Sampling) [30] randomly selects samples from the majority class samples for rejection.

  • NearMiss [31] selected the most representative samples from most classes for training to alleviate the problem of information loss in random undersampling.

  • TomekLink [32] represents the nearest pair of samples between different categories, which are nearest neighbors of each other and belong to different classes.

  • ENN (Edited Nearest Neighbor) [33] traverses the samples of most classes, and if most of the k-nearest neighbor samples are not the same as their class, they are deleted.

  • RENN (Repeated Edited Nearest Neighbor) keeps repeating the deletion process of ENN until it can no longer be deleted.

  • CNN (Condensed Nearest Neighbor) [34] uses the nearest neighbor approach to iterate and determine whether a sample should be retained or rejected.

  • OSS (One Side Sampling) [35] rejects noisy samples by using multiple TomekLink iterations.

  • AllKNN [36] applies ENN multiple times and will change the number of nearest neighbors.

2.4.2Over-sampling methods

  • RandOver (Random Over-Sampling) randomly samples from categories with few samples, and then adds the sampled samples to the data set.

  • SMOTE (synthetic minority oversampling technique) [37] interpolates between a few classes of samples to generate additional samples.

  • BorderSMOTE (Borderline Synthetic Minority Oversampling Technique) [38] first distinguishes the minority class samples located at the border and performs KNN sampling for these samples.

  • KMeansSMOTE [39] first applies KMeans clustering and then oversamples using SMOTE.

  • SVMSMOTE [40] uses the SVM classifier to generate support vectors to generate new minority class samples, which are then synthesized using SMOTE.

  • ADASYN (adaptive synthetic sampling) [41] uses some mechanism to automatically determine how many synthetic samples need to be generated for each minority class sample.

2.4.3Combination of over- and under-sampling methods

  • SMOTETomek (SMOTE with tomek links cleaning) [42] combines over and under sampling using SMOTE and Tomek links.

2.5Model training and evaluation

3-fold cross-validation was used to prove the performance of the model on the training data set. The training data consist of 67% of the total data; the test data consist of the remaining 33%. Among all 427 sets of patient data, 285 (258 mild, 27 severe) data sets were used as training data, and 142 (129 mild, 13 severe) data sets were used as test data. Finally, we used training data to train eight different machine learning algorithms and the trained models to predict the test data. Accuracy, f1-score, AUC, and AUCPRC were used to analyze the model’s performance. Figure 2 shows the workflow of our method.

Figure 2.

Overall training flowchart of the COVID-19 mild-severe disease prediction model.

Overall training flowchart of the COVID-19 mild-severe disease prediction model.

To better measure the model’s effectiveness, we used different evaluation indicators to compare the multiple aspects. According to the confusion matrix, the following indicators were used to evaluate the performance of the model comprehensively:

TP, FP, TN, and FN stand for True Positive, False Positive, True Negatives, and False Negative respectively.

(3)
𝐴𝑐𝑐𝑢𝑟𝑎𝑐𝑦=TP+TNTP+FP+FN+TN
(4)
𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛=TPTP+FP
(5)
𝑟𝑒𝑐𝑎𝑙𝑙=TPTP+TN
(6)
F1-score=2×(𝑟𝑒𝑐𝑎𝑙𝑙×𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛)𝑟𝑒𝑐𝑎𝑙𝑙+𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛
(7)
G-mean=𝑟𝑒𝑐𝑎𝑙𝑙𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛
(8)
𝑀𝐶𝐶=TP×TN-FP×FN(TP+FP)(TP+FN)(TN+FP)(TN+FN)
(9)
𝐴𝑈𝐶𝑃𝑅𝐶=Area Under precision-recall Curve
(10)
𝐴𝑈𝐶=area under the curve

3.Results

3.1Feature selection

To further verify the effect of different features on machine learning prediction results, we performed 100 RandomForest feature sorting on clinical features, CT radiomics features, a mixture of clinical features, and CT radiomics features, respectively. The final mean value of 100 experimental results was calculated as the final feature importance ranking result. We further filtered the number of features that make different machine learning algorithms get optimal results based on the feature ranking results. The optimal number of feature combinations for different machine learning models is shown in Fig. 3.

Figure 3.

Number of features for each machine learning algorithm to obtain optimal prediction results.

Number of features for each machine learning algorithm to obtain optimal prediction results.

Figure 4.

Experimental results of AUCPRC obtained by various machine learning algorithms using different types of features.

Experimental results of AUCPRC obtained by various machine learning algorithms using different types of features.

The order of feature selection was sorted according to the score of the random forest algorithm from high to low. The feature importance scores of the clinical features and the CT extracted features are shown in Table 3. To get the best prediction results, we used gridSearchCV to optimize the parameters of each machine learning model. The optimal model parameters are shown in Table 4.

Table 3

Feature importance scores

Feature nameFeature importance score
PO20.0513
Age0.0264
original_firstorder_Entropy0.0211
original_glcm_SumEntropy0.0189
PCT0.0180
original_firstorder_RobustMeanAbsoluteDeviation0.0170
Lactate dehy drogenase0.0168
C-reactive protein0.0166
original_glrlm_GrayLevelNonUniformityNormalized0.0164
original_firstorder_Kurtosis0.0159
original_firstorder_InterquartileRange0.0150
original_glrlm_RunEntropy0.0150
original_glcm_JointEnergy0.0147
APTT0.0143
original_glcm_JointEntropy0.0142
original_firstorder_Skewness0.0140
original_firstorder_Uniformity0.0135
original_glszm_SmallAreaEmphasis0.0126
original_glcm_Id0.0126
eGFR0.0125
original_glcm_InverseVariance0.0125
Na0.0123
original_glszm_SizeZoneNonUniformityNormalized0.0119
Total cholesterol (TC)0.0115
original_glszm_ZoneEntropy0.0104
original_glszm_GrayLevelNonUniformityNormalized0.0098
original_glszm_ZonePercentage0.0095
original_firstorder_MeanAbsoluteDeviation0.0093
original_glcm_MaximumProbability0.0092
original_glcm_Idm0.0086

Table 4

The optimal model parameters

AlgorithmModel parametersValue
KNNn_neighbors11
p1
LRPenaltyL2
C1
class_weightBalanced
RFclass_weightBalanced
CriterionEntropy
max_depth14
min_samples_leaf1
n_estimators65
XGBlearning_rate1
n_estimators64
max_depth2
subsample0.8
reg_alpha0.1
reg_lambda1
GBDTboosting_typegbdt
colsample_bytree0.6
learning_rate1
n_estimators32
num_leaves50
objectiveBinary
reg_alpha0.1
reg_lambda10
subsample0.8
MLPhidden_layer_sizes3
ActivationIdentity
Solverlbfgs
learning_rateConstant
random_state2008

After parameter optimization and combining the optimal features, we trained the COVID-19 mild-severe prediction models with eight machine learning algorithms. Due to some imbalance in the experimental data, we finally chose the PRAUC value to verify the model’s goodness. Ten 3-fold cross-validation experiments were conducted separately, and the mean values were obtained as the final results of the experiments. Figure 4 shows the experimental results of AUCPRC obtained by each machine learning algorithm using different types of features. By comparing the experimental results, we can see that the final results obtained by different machine learning algorithms using the combined Clinical and CT radiomics features are higher than those obtained with clinical or CT radiomics features alone. The experimental results also further demonstrate the more accurate results of the machine learning classification model built using clinical and CT radiomics features. Therefore, to further verify the effect of data imbalance on the experimental results. We will select combined clinical and clinical feature data for further sampling experiments in the following experiments.

3.2Evaluation of diagnostic performance

In the following data sampling experiments, 3-fold cross-validation was used to divide the data, and then the divided data were combined with different sampling methods for the sampling experiments. The Under-sampling methods were used to sample 0.2, 0.4, 0.6, 0.8, and 1.0 times the majority of the data. The Over-sampling methods were used to sample 0.1, 0.2, 0.4, 0.6, and 0.8 times the majority of the data for the minority data. All sampling methods were performed on the training set data, and no processing was done on the test set data. Ten 3-fold cross-validation experiments were conducted separately, and the average of the ten experimental results was used as the final result. Figure 5 shows the results without the data sampling process. Figure 6 shows the optimal results after data sampling. By comparing Figs 5 and 6, it can be found that the model results still improve after sampling the data. RandomForest has the largest PRAUC value improvement of 3.7% after using the sampling method, and GBDT has the optimal PRAUC result of 0.697 in the used model after the sampling process.

Figure 5.

The optimal prediction results were obtained by each machine learning model trained using unsampled data.

The optimal prediction results were obtained by each machine learning model trained using unsampled data.

Figure 6.

The optimal prediction results were obtained by each machine learning model trained using sampled processed data.

The optimal prediction results were obtained by each machine learning model trained using sampled processed data.

Table 5

The optimal prediction results of each machine learning model are obtained by training with data processed in different data sampling methods

ModelCategoryMethodRadioAccuracyf1-scoreAUCAUCPRCMCCGM
NBORG0.8610.8870.8640.5750.4540.659
UndersamplingNearMiss200.9050.9360.8550.5880.5630.649
RandUnder800.8640.8890.8650.5770.4520.663
ClusterCentroids800.8770.9050.8630.5800.4690.661
OversamplingBorderSMOTE11000.8350.8520.8670.5790.4310.662
BorderSMOTE2400.8380.8580.8650.5790.4370.660
MLPORG0.9200.9680.8690.5490.5720.614
UndersamplingRandUnder800.9150.9590.8680.5520.5750.615
ClusterCentroids800.9160.9650.8690.5540.5670.614
ENN0.9070.9440.8700.5550.5840.625
OSS0.9170.9620.8700.5540.5690.610
KNNORG0.9280.9930.8430.5250.5510.578
UndersamplingRandUnder400.9260.9780.8640.5490.5590.591
ClusterCentroids800.9270.9910.8490.5260.5420.570
ENN0.9290.9820.8530.5410.5540.581
RENN0.9230.9690.8540.5370.5520.583
AllKNN0.9260.9780.8550.5480.5650.593
OSS0.9300.9910.8440.5320.5570.582
LRORG0.8360.8440.8640.5390.5420.593
UndersamplingENN0.8200.8230.8660.5440.5590.606
RENN0.8030.8040.8680.5680.5620.611
AllKNN0.8180.8180.8680.5530.5630.613
OversamplingKMeansSMOTE1000.8900.9200.8640.5580.5670.610
SVMSMOTE400.8570.8720.8630.5560.5630.608
RFORG0.9190.9650.8930.6090.6010.631
UndersamplingClusterCentroids600.9210.9650.9060.6290.5980.635
ENN0.9100.9480.8860.6170.6050.639
AllKNN0.9080.9400.8960.6130.6070.640
CNN0.9080.9370.8880.6340.6290.659
OSS0.9220.9660.9030.6290.6140.647
OversamplingRandOver200.9170.9560.9010.6290.6110.646
SMOTE600.8900.9110.9060.6320.6050.641
KMeansSMOTE200.9210.9620.8960.6200.5980.636
SVMSMOTE200.9140.9540.9000.6360.6130.645
ADASYN200.9080.9400.8980.6120.6110.645
Over+UndersamplingSMOTETomek600.8910.9160.9000.6320.6190.657
AdaBoostORG0.9250.9750.9280.6090.6050.303
UndersamplingRandUnder800.9240.9730.9290.6230.6100.303
ClusterCentroids800.9180.9680.9260.6110.6040.303
TomekLinks0.9280.9810.931 0.654 0.6330.303
OversamplingRandOver400.9220.9560.9280.6200.6070.303
SMOTE400.9110.9420.9170.6160.6170.303
SVMSMOTE400.9170.9510.9160.6230.6010.303
Over+UndersamplingSMOTETomek200.9200.9640.9220.6200.6010.303
XGBORG0.9270.9720.9210.6430.6230.661
UndersamplingClusterCentroids800.9260.9670.922 0.661 0.6290.667
OversamplingSVMSMOTE200.9240.9670.9170.6550.6320.667
GBDTORG0.9300.9780.9400.6760.6440.684
UndersamplingClusterCentroids800.9310.9750.9390.6800.6370.674
OSS 0.932 0.9740.9380.6790.6350.672
OversamplingRandOver400.9310.970 0.942 0.6940.6600.696
KMeansSMOTE400.9320.9710.9400.6800.6400.677
SVMSMOTE400.9310.9650.940 0.697 0.6540.686

To further verify the final results of each machine learning algorithm under different sampling methods and sampling ratios, we collated all the results after sampling, and the final experimental results are shown in Table 5. By analyzing the experimental results in Table 5, we can find that the final results obtained by different models combined with different sampling methods and sampling ratios are different. The experimental results of all eight machine learning experiments were further improved after using the data after data sampling and processing. RandomForest obtained the largest AUCPRC value improvement of 2.7% for the data after sampling with SVMSMOTE. GBDT got the largest AUCPRC value of 0.697 among the eight machine learning prediction models after sampling with SVMSMOTE. GBDT used the RandOver sampling method to make the final combined performance optimal among the eight different machine learning algorithms, the accuracy is 0.931, the AUC is 0.942, and the AUCPRC is 0.694. The results can also provide some reference value for data imbalance experiments combining different algorithms and sampling methods to optimize the experimental results.

4.Discussion

The purpose of this study was to develop a diagnostic model for predicting the severity of patients with COVID-19. Using clinical features and CT radiomics features, the optimal prediction accuracy of 0.932 and AUC value of 0.942 were obtained for the diagnostic model based on the GBDT algorithm after data sampling processing and feature selection. The model can assist clinicians in screening patients with severe COVID-19, providing more medical resources for these patients, and can also be used to improve patient prognosis decisions and assess prognostic treatment outcomes.

Several studies have been conducted to build diagnostic models for the severity of COVID-19 patients using machine learning algorithms, and details of the diagnostic models are shown in Table 6. These studies used patients’ clinical features to build diagnostic models and did not consider the impact of CT radiomics features on model prediction performance. In our study, to further validate the effect of CT radiomics features on the prediction model’s performance, we segmented the ROI on the chest CT by using SP-V-Net and extracted the CT radiomics features on the ROI. Experiments were performed separately using clinical features, CT radiomics features, and a mixture of both. Figure 4 shows that the COVID-19 mild-severe prediction model built by using a mixture of features of both has better performance.

Table 6

Comparison of machine learning-based methods for COVID-19 mild and severe diagnostic studies

StudyFeaturesML approachesSample sizePerformance
Terwangne et al. [11]clinicalBayesian295 patients (-)AUC of 83.8%
Yao et al. [12]clinicalSVM137 patients (75 severe)Accuracy of 81.48%
Zhao et al. [13]clinicalSVM172 patients (60 severe)Accuracy of 91.38%
Liang et al. [14]clinicalLR1590 patients (131 severe)AUC of 88%
Zhu et al. [15]clinicalLR127 patients (16 severe)AUC of 90.0%
Our methodclinical+CT radiomicsGBDT427 patients (40 severe)Accuracy of 0.931 and
AUC of 0.942

The data imbalance problem was widely present in many real-world decision problems, and in medical diagnosis models, the data imbalance problem can have a negative impact on experimental results. The problem of data imbalance in the studies of Liang et al. [14] and Zhu et al. [15] was not addressed. In our research, to reduce the impact of the data imbalance problem on the prediction results, we sampled the data using three different data sampling methods. By comparing Figs 5 and 6, it can be found that the prediction results of different models were still further improved after the data imbalance treatment. In this study, the optimal prediction results were obtained using integrated models such as GBDT, AdaBoost, and XGB combined with the data after sampling. These classifiers all used integrated learning techniques to improve the accuracy of individual classifiers and overall classifier performance. And related studies have demonstrated that the model prediction performance can be further improved by using integrated models when dealing with data imbalance problems [43].

By further feature ranking and screening of a mixture of CT radiomics features and clinical features, we found that PO2, age, PCT, LDH, and CRP were the five most important clinical risk factors with the most severe degree of prognostic observation for patients with COVID-19, and this result was also consistent with previous related studies. Studies have shown that older age, elevated PCT, LDH, and CRP are all important correlates of the severity of COVID-19 [17]. More important is the continued importance of image features extracted from chest CT to diagnose mild to severe disease in patients with COVID-19. The combination of clinical features and chest CT extracted features has a good performance in diagnosing the severity of COVID-19. COVID-19 patients have specific chest CT image features, including ground-glass opacities (GGO), multifocal patchy consolidation, or interstitial changes in the peripheral distribution [44]. The increase in lesion volume, gross glass volume, and other volumes also provided the possibility for the model to predict the severity of COVID-19.

Machine learning algorithms are now widely used in complementary medical diagnosis and are playing an increasingly important role. Catic et al. [45] built prenatal diagnosis classification models using Artificial Neural Networks (ANNs) artificial neural networks to help physicians in their daily work, eventually obtaining feedforward neural networks with an average accuracy of 89.6% and feedback of 98.8%. Begic et al. [46] applied machine learning algorithms to diagnose congenital heart defects, obtaining a diagnostic accuracy of 94.28% by making the model built by SVM. Stokes et al. [47] applied the trained machine learning model to diagnose and refer to bronchitis and pneumonia. They obtained a 93% AUC value model performance by using decision trees. In our study, the COVID-19 mild-severe diagnostic model built using the GBDT algorithm obtained optimal predictive performance with an accuracy of 93.2% and an AUC of 94.2%. Compared with the above three studies, the model we developed also has good performance in diagnosing mild and severe COVID-19, which can assist physicians in the early detection of severe COVID-19 patients and provide them with better medical resources, and this study has better clinical significance. ML can help improve disease-specific diagnostic systems’ reliability, performance and accuracy. The research and application of ML in the medical field are also increasing, and the related research will provide more convenience for doctors and patients.

The present study still has some limitations that need to be considered. First, the number of patients with COVID-19 is relatively small, limiting the accuracy of the predictive model. Second, the diversity of data in our study is limited, all subjects are Chinese COVID-19 patients, and the results may not be fully applicable to data from other countries. Third, the number of severe patients’ data is small, and the mild and severe patients’ data are imbalanced. We need to collect more COVID-19 patient data, especially severe patients. Further research is still required.

5.Conclusions

This study proves that the COVID-19 mild-severe prediction model based on features extracted from chest CT and clinical characteristics can effectively differentiate the severity of COVID-19 patients and can provide helpful insights for early diagnosis of patients with COVID-19 mild-severe disease. And the prediction models based on both chest CT features and clinical features had higher prediction performance than those built using the two different types of data alone. The results could help clinicians more effectively assess the severity of COVID-19 patients and stratify patients for treatment to reduce potential mortality and ease the burden of care.

Acknowledgments

This research was funded by Henan Science and Technology Development Plan 2022 (grant number 222102210219) and Zhengzhou University of Light Industry (grant number 2019ZCKJ228).

Ethics statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board (or Ethics Committee) of the Shanghai Public Health Clinical Center (YJ-2020-S035-01, approval on 22 February 2020).

References

[1] 

WHO Coronavirus (COVID-19) Dashboard Data [WHO (COVID-19) Homepage]. World Health Organization; 2022 [updated 2022 May 25; cited 2022 May 26]. Available from: https://covid19.who.int/.

[2] 

Jin X, Pang B, Zhang J, Liu Q, Yang Z, Feng J, et al. Core outcome set for clinical trials on coronavirus disease 2019 (COS-COVID). Engineering. (2020) ; 6: (10): 1147-52.

[3] 

Xu Z, Shi L, Wang Y, Zhang J, Huang L, Zhang C, et al. Pathological findings of COVID-19 associated with acute respiratory distress syndrome. The Lancet Respiratory Medicine. (2020) ; 8: (4): 420-2.

[4] 

Shi H, Han X, Jiang N, Cao Y, Alwalid O, Gu J, et al. Radiological findings from 81 patients with COVID-19 pneumonia in Wuhan, China: A descriptive study. The Lancet Infectious Diseases. (2020) ; 20: (4): 425-34.

[5] 

Ai T, Yang Z, Hou H, Zhan C, Chen C, Lv W, et al. Correlation of chest CT and RT-PCR testing for coronavirus disease 2019 (COVID-19) in China: A report of 1014 cases. Radiology. (2020) ; 296: (2): E32-E40.

[6] 

Zu ZY, Jiang MD, Xu PP, Chen W, Ni QQ, Lu GM, et al. Coronavirus disease 2019 (COVID-19): A perspective from China. Radiology. (2020) ; 296: (2): E15-E25.

[7] 

Lei J, Li J, Li X, Qi X. CT imaging of the 2019 novel coronavirus (2019-nCoV) pneumonia. Radiology. (2020) ; 295: (1): 18-.

[8] 

Kanne JP. Chest CT findings in 2019 novel coronavirus (2019-nCoV) infections from Wuhan, China: key points for the radiologist. Radiological Society of North America; (2020) . 16-7.

[9] 

Pan Y, Guan H. Imaging changes in patients with 2019-nCov. Springer; (2020) . 3612-3.

[10] 

Ng M-Y, Lee EY, Yang J, Yang F, Li X, Wang H, et al. Imaging profile of the COVID-19 infection: Radiologic findings and literature review. Radiology: Cardiothoracic Imaging. (2020) ; 2: (1): e200034.

[11] 

de Terwangne C, Laouni J, Jouffe L, Lechien JR, Bouillon V, Place S, et al. Predictive accuracy of COVID-19 world health organization (Who) severity classification and comparison with a bayesian-method-based severity score (epi-score). Pathogens. (2020) ; 9: (11): 880.

[12] 

Yao H, Zhang N, Zhang R, Duan M, Xie T, Pan J, et al. Severity detection for the coronavirus disease 2019 (COVID-19) patients using a machine learning model based on the blood and urine tests. Frontiers in cell and developmental biology. (2020) . 683.

[13] 

Zhao C, Bai Y, Wang C, Zhong Y, Lu N, Tian L, et al. Risk factors related to the severity of COVID-19 in Wuhan. International Journal of Medical Sciences. (2021) ; 18: (1): 120.

[14] 

Liang W, Liang H, Ou L, Chen B, Chen A, Li C, et al. Development and validation of a clinical risk score to predict the occurrence of critical illness in hospitalized patients with COVID-19. JAMA Internal Medicine. (2020) ; 180: (8): 1081-1089.

[15] 

Zhu Z, Cai T, Fan L, Lou K, Hua X, Huang Z, et al. Clinical value of immune-inflammatory parameters to assess the severity of coronavirus disease 2019. International Journal of Infectious Diseases. (2020) ; 95: : 332-9.

[16] 

Van Buuren S, Groothuis-Oudshoorn K. mice: Multivariate imputation by chained equations in R. Journal of Statistical Software. (2011) ; 45: : 1-67.

[17] 

Zhao C, Xu Y, He Z, Tang J, Zhang Y, Han J, et al. Lung segmentation and automatic detection of COVID-19 using radiomic features from chest CT images. Pattern Recognition. (2021) ; 119: : 108071.

[18] 

Zwanenburg A, Leger S, Vallières M, Löck S. Image biomarker standardisation initiative. arXiv preprint arXiv:161207003. 2016.

[19] 

Koller D, Sahami M. Toward optimal feature selection. Stanford InfoLab; (1996) .

[20] 

Zhu F, Li X, Tang H, He Z, Zhang C, Hung G-U, et al. Machine learning for the preliminary diagnosis of dementia. Scientific Programming. (2020) ; 2020.

[21] 

Bahad P, Saxena P, editors. Study of adaboost and gradient boosting algorithms for predictive analytics. International Conference on Intelligent Computing and Smart Communication 2019; Springer; (2020) .

[22] 

Lusa L. Gradient boosting for high-dimensional prediction of rare events. Computational Statistics & Data Analysis. (2017) ; 113: : 19-37.

[23] 

Neagoe V-E, Ciotec A-D, Cucu G-S, editors. Deep convolutional neural networks versus multilayer perceptron for financial prediction. 2018 International Conference on Communications (COMM); IEEE; (2018) .

[24] 

Chen T, Guestrin C, editors. Xgboost: A scalable tree boosting system. Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining; (2016) .

[25] 

Chen Z, Zhou LJ, Da Li X, Zhang JN, Huo WJ. The Lao text classification method based on KNN. Procedia Computer Science. (2020) ; 166: : 523-8.

[26] 

Luo C-L, Rong Y, Chen H, Zhang W-W, Wu L, Wei D, et al. A logistic regression model for noninvasive prediction of AFP-negative hepatocellular carcinoma. Technology in Cancer Research & Treatment. (2019) ; 18: : 1533033819846632.

[27] 

Battineni G, Sagaro GG, Nalini C, Amenta F, Tayebati SK. Comparative machine-learning approach: A follow-up study on type 2 diabetes predictions by cross-validation methods. Machines. (2019) ; 7: (4): 74.

[28] 

Pal M. Random forest classifier for remote sensing classification. International Journal of Remote Sensing. (2005) ; 26: (1): 217-22.

[29] 

Singh H, Kaur K. New method for finding initial cluster centroids in K-means algorithm. International Journal of Computer Applications. (2013) ; 74: (6).

[30] 

Tahir MA, Kittler J, Yan F. Inverse random under sampling for class imbalance problem and its application to multi-label classification. Pattern Recognition. (2012) ; 45: (10): 3738-50.

[31] 

Mani I, Zhang I, editors. kNN approach to unbalanced data distributions: a case study involving information extraction. Proceedings of workshop on learning from imbalanced datasets; ICML; (2003) .

[32] 

Tomek I. Two modifications of CNN. IEEE Trans Systems, Man and Cybernetics. (1976) ; 6: : 769-72.

[33] 

Wilson DL. Asymptotic properties of nearest neighbor rules using edited data. IEEE Transactions on Systems, Man, and Cybernetics. 1972(3): 408-21.

[34] 

Hart P. The condensed nearest neighbor rule (corresp).. IEEE Transactions on Information Theory. (1968) ; 14: (3): 515-6.

[35] 

Kubat M, Matwin S, editors. Addressing the curse of imbalanced training sets: one-sided selection. Icml; (1997) ; Citeseer.

[36] 

Tomek I. AN EXPERIMENT WITH THE EDITED NEAREST-NIEGHBOR RULE. (1976) .

[37] 

Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP. SMOTE: Synthetic minority over-sampling technique. Journal of Artificial Intelligence Research. (2002) ; 16: : 321-57.

[38] 

Han H, Wang W-Y, Mao B-H, editors. Borderline-SMOTE: a new over-sampling method in imbalanced data sets learning. International conference on intelligent computing; Springer; (2005) .

[39] 

Last F, Douzas G, Bacao F. Oversampling for imbalanced learning based on k-means and smote. arXiv preprint arXiv:171100837. (2017) .

[40] 

Nguyen HM, Cooper EW, Kamei K. Borderline over-sampling for imbalanced data classification. International Journal of Knowledge Engineering and Soft Data Paradigms. (2011) ; 3: (1): 4-21.

[41] 

He H, Bai Y, Garcia EA, Li S, editors. ADASYN: Adaptive synthetic sampling approach for imbalanced learning. 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence); IEEE; (2008) .

[42] 

Batista GE, Bazzan AL, Monard MC, editors. Balancing Training Data for Automated Annotation of Keywords: a Case Study. WOB; (2003) .

[43] 

Galar M, Fernandez A, Barrenechea E, Bustince H, Herrera F. A review on ensembles for the class imbalance problem: Bagging-, boosting-, and hybrid-based approaches. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews). (2011) ; 42: (4): 463-484.

[44] 

Chung M, Bernheim A, Mei X, Zhang N, Huang M, Zeng X, et al. CT imaging features of 2019 novel coronavirus (2019-nCoV). Radiology. (2020) ; 295: (1): 202-7.

[45] 

Catic A, Gurbeta L, Kurtovic-Kozaric A, Mehmedbasic S, Badnjevic A. Application of neural networks for classification of patau, edwards, down, turner and klinefelter syndrome based on first trimester maternal serum screening data, ultrasonographic findings and patient demographics. Bmc Medical Genomics. (2018) ; 11: (1): 19.

[46] 

Begic E, Gurbeta Pokvic L, Begic Z, Begic N, Dedic M, Mrsic D, Badnjevic AFROM. Heart murmur to echocardiography-congenital heart defects diagnostics using machinelearning algorithms. Psychiatria Danubina. (2021) ; 33: (suppl 13): 236-246.

[47] 

Stokes K, Castaldo R, Franzese M, Salvatore M, Fico G, Pokvic LG, Pecchia L. A machine learning model for supporting symptom-based referral and diagnosis of bronchitis and pneumonia in limited resource settings. Biocybernetics and Biomedical Engineering. (2021) ; 41: (4): 1288-1302.