1 Introduction

1.1 Background of COVID-19

Coronaviruses are a group of viruses that can cause illnesses such as the common cold, Middle East respiratory syndrome (MERS) and severe acute respiratory syndrome (SARS). The virus is known as severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2). The resulting disease is called emerging coronavirus disease 2019 (COVID-19). The first reported case of COVID-19 was in Wuhan city, China, in December 2019 [6]. Then, the virus spread from one country to another in a few months, and the number of deaths increased. In March 2020, the World Health Organization announced the Coronavirus (Covid 19) to become a global pandemic. According to WHO reports, the total number of infected people is 521 607 587, and dead people is 6 243 038 as of 5 May 2022 [30].

There are common symptoms that appear when infected with the virus, such as fever, cough, fatigue, sore throat, and shortness of breath. Some of them are less common such as myalgias, runny nose, headache, diarrhea, sweating, pleural effusion, and others [2].

The effects of the Corona pandemic extended to all aspects of social [11], economic [19], educational [24], and healthy [14] life as well.

The typical method to diagnose COVID-19 is reverse transcription polymerase chain reaction (RT-PCR). In the early days of the epidemic in China RT-PCR was only 30 to 70 percent sensitive, whereas chest CT was reportedly much more sensitive in that context. However, more recent data from the U.S. labs at the University of Washington suggest that second-generation COVID RT-PCR tests fare much better with 95% or more sensitivity [15].

RT-PCR is used to determine whether the person is infected or not, but it does not determine the severity and degree of infection. So, another way must be found to determine the severity of the infection. CT scans are another way to diagnose COVID-19, and they also are used to identify the severity of infection based on the findings in the image. The CT scan is noninvasive imaging technology. It uses X-ray to identify the density of tissues. Lungs which are considered soft tissue absorb X-rays lower than High dense tissues like bones. Coronavirus attacks the lungs and causes the formation of fibrotic tissues that can be noted in CT. Lung fibrosis causes stiffening and scarring of the lung tissue, which makes the dense appearance in the CT images [5].

At the beginning of the pandemic, it was necessary to determine the changes that occur in the lungs as a result of infection. So, many researchers focused on studying these changes and the changes that occur during the treatment period. These changes were used to determine the stages of infection and to determine the severity of infection. Based on the previous studies, the CT findings of COVID-19 are GGO, Consolidation, and Crazy Paving, [9, 20, 26, 29, 36].

1.2 Machine learning

In the fact, the best way to reduce the spread of the virus is through social distance, hand washing, and face masking. However, technology can reduce its spread through early identification. Artificial intelligence is used in wide applications in the medical field and proved its efficiency. Machine learning is a branch of AI and computer science that is concerned with designing algorithms to allow computers to have the ability to learn without programming rules for each issue [1, 10, 12, 32]. The advantages of using ML classifiers are:

  1. 1.

    Quickly identifies trends and patterns.

  2. 2.

    No human intervention is needed (automation).

  3. 3.

    Continuous Improvement.

  4. 4.

    Handle multi-dimensional and multi-variety data.

  5. 5.

    Wide Applications.

There are three types of machine learning:

  1. 1.

    Supervised Learning: The computer is trained by giving it input and output data, so it is called supervision, where learning is supervised by giving outputs to data. Through training, the computer builds relationships and patterns between inputs and outputs so that it can later predict new data outputs. Classification and recognition are applications of supervised learning [16].

  2. 2.

    Unsupervised Learning: The computer is trained by giving it input data only, so it is called unsupervised, where no output data are given. Through training, the computer builds relationships and patterns between the data itself so that it can predict new data outputs. Clustering is an application of unsupervised learning [16].

  3. 3.

    Reinforcement Learning: It is a self-learning system in which the learning process takes place through trial and error. It differs from supervised learning; in supervised learning, the training data has output data, which means it trains with correct answers. Whereas in reinforcement learning there is no output data, it trains by its experiences. Autonomous cars are an application of reinforcement learning [16].

1.3 Our contribution

The main objectives of our research are as follows:

  1. 1.

    Proposing a simple, accurate, and reliable method for detecting the degree of infection with COVID-19 based on CT scans.

  2. 2.

    Classifying the CT scans into four stages: Normal, Mild, Moderate, and Severe based on the CT findings, and the description file that is found with the dataset.

  3. 3.

    Use four machine learning models (DT, KNN, SVM, and Naïve Bayes) to train and test our model. These four machine learning algorithms are the most popular ones. These algorithms are based on simple concepts but efficient classification results; SVM depends on finding the best hyperplane that separates the stages. KNN attempts to predict the proper class for the test instance by computing the distance between the test instance and all labeled training examples. Naïve Bayes operates primarily on the probability principle, computing the odds of the test instance being assigned to each of the available class labels. The label of the test instance is determined by the largest value of these probabilities. DT based on Yes / No questions.

  4. 4.

    Extracting ratio of infection (ROI), statistical texture features (mean, standard deviation, skewness and kurtosis), GLCM and GLRLM texture features from the scans.

  5. 5.

    Comparing our model with state-of-arts and proving that it is better performance than theirs.

The rest of this research is as follow: Sect. 2 includes a literature review. Section 3 classifies our approach. Section 4 shows the experiments’ result and discussion. The conclusions and future works are expalied at Sect. 5.

2 Related work

Although the Corona pandemic did not appear for a long time, many researchers have taken care of it due to its danger to human life.

Many researchers employ artificial intelligence techniques to classify the CT images to whether they are infected with this virus or if they are healthy and some of them classify the infected cases in the severity stages as [4] propose a method to diagnosis and severity detection for CT Covid 19 images. Two public datasets are used that have the metadata of patients’ CT images. There are two stages COVID-19 diagnosis and severity detection. In the COVID-19 stage, the images classify into covid or non-covid images using transfer learning ( ResNet-50). The covid images are classified into low, medium or high severity in the second stage. The performance is 98.5% of accuracy for the Covid 19 diagnosis stage, and 97.3% of accuracy for the severity detection stage.

Tang et al. [28] assesses the severity of CT scans of COVID-19 into non-severe or severe stage using the machine learning algorithm (RF). Quantitative features are calculated from CT images (63 features) such as the infection volume/ratio to the whole lung volume, and the volume of GGO, then these features are used to assess the severity of the disease. This method achieves 87.5% of accuracy. And the highest quantitative feature related to the severity is the volume of GGO to the whole lung volume.

Yu et al. [34] proposes multiple classifiers to classify the CT images of Covid19 into severe or non-severe. To extract features, CT images are passed into different deep learning methods (DenseNet -201, ResNet-50, Inception-V3 and ResNet-101). Multiple classifiers are used to classify images into severe or non-severe images (KNN, Adaboost Decision Tree, Linear SVM, Cubic SVM and Linear Discriminate). Three validation strategies are used ( leave-one-out, ten-fold cross-validation and hold out). The accuracy of DenseNet-201 with Cubic SVM is 95.34% and 95.2% in leave one out and ten-fold cross-validation.

Irmak [13] proposes a fully automatic severity classification X-ray images of COVID-19 based on Convolution neural network CNN. The severity is divided into four stages (mild, moderate, severe and critical). The standard approach is used for COVID-19 lung severity scoring, which indicates that the scoring is based on the opacities and lung involvement. The hyperparameters of CNN are optimized by grid search optimization. The proposed model achieves 95.52% of accuracy.

Xiao et al. [31] use the different deep learning networks ( ResNet34, AlexNet, VGGNet, and DenseNet) to predict disease severity (severe/non-severe). The severity divides into four stages: mild, moderate, severe, and critical. The mild and moderate stages are combined in non-severe, and the severe and critical stages are combined in the severe stage. The ResNet34 achieves the highest accuracy compared to other networks, which achieves 97.4% for training and 81.9% for testing.

Amini and Shalbaf [3] uses CT scans to classify the severity of the COVID-19 into four stages normal, mild, moderate and severe. 28 statistical texture features are extracted which are skewness, kurtosis and variance GLCM with 23 features, GLRLM and GLSZM. The extracted features are used to determine the severity of the stages using machine learning algorithms (RF, KNN and LDA). The RF achieves the highest accuracy (90.95%) compared with others.

Zhu et al. [37] propose a deep learning convolution neural network to assess the severity of COVID-19 in Chest X-Ray images. The severity is divided into four stages mild, moderate, severe and critical. The total images are passed into CNN and classified into stages. The comparison between the deep learning method and chest radiologist scoring is done in terms of some metrics such as correlation coefficient and mean absolute error (8.5%). But the accuracy or ROC curve does not evaluate.

Fig. 1
figure 1

Our proposed methodology

3 Our proposed approach

3.1 The methodology

In our work, the proposed methodology consists of different steps of COVID-19 infection stage detection and classification. Figure 1 shows our proposed methodology.

3.2 Dataset

In our work, three CT scans datasets are used [21, 33, 35]. In Zaid [35] there are two types of images: 4001 positive CT (pCT) and 9979 negative CT (nCT) images. In our work, we use the 4001 positive CT (pCT) images but there are a lot of repetitions in the images, as well as we used the images that have full lungs by removing the upper and lower slices in the preprocessing step. The number of scans used from this dataset is 299 CT scans. In Yang et al. [33] there are 349 COVID-19 CT images and 463 non-COVID-19 CTs. We use all the COVID-19 CT images and 250 images from non-COVID-19 CTs. In Plameneduardo [21] there are 1252 CT positive scans and 1230 CT negative scans. We used 1108 positive CT scans. So, the total number of our dataset is 250 non-COVID-19 scans and 1801 COVID-19 scans.

Fig. 2
figure 2

Samples from the removing step (upper and lower parts are removed; full regions remain) and Samples from the cropping step

3.3 Image preprocessing

In our work, we do two preprocessing steps on the dataset: the removing step and the cropping step. Any CT image produces 20–30 slices from the upper part of the chest to the lower part. We removed the slices that contain the upper and lower part and keep the slices that contain the complete lung regions in the removing step. In the cropping step, we crop the lung regions only and delete any surrounded regions. This step is the most difficult because not all the images have the same position of the lung. So, we did multiple cropping steps to have the best results. Figure 2 shows some samples from the removing and the cropping step.

3.4 Feature extraction

Feature Extraction is one of the most important step in the COVID-19 classification and detection. In our work, we use 16 features to detect the stage of infection. These features are the ratio of white regions of lesions to white regions of the lung (Ratio of Infection) [22], global statistical texture features, GLCM and GLRLM texture features [3, 17, 25].

3.4.1 Ratio of infection feature

Fig. 3
figure 3

Ratio of Infection extraced feature step

Figure 3 shows our proposed approach to extracting Ratio of Infection feature [22]. In the masking lung step, we increase the intensity values of image by imadjust function in MATLAB. Then, we create a binary image using Otsu’s method which convertes all values above a global threshold with 1 s and all other values with 0 s. Next, we erode the binary image using disk structuring element to mask the lung. We complement the mask. Finally, we fill the holes in the lung area. After we masked the lungs, we calculate the white regions in the image.

In the masking lesions step. First, we feedback the segmented lung by multiplying the original image by the masked image. Then, we increase the intensity values of image by imadjust function in MATLAB. We erode the binary image using disk structuring element to mask the lung. We segment the lesions that are greater than the thresholding value (150). This value is determined by experiments. After we segmented the lesions, we calculated the white regions of lesions in the image. See Fig. 4.

We can calculate the Ratio of Infection by the following equation:

$$\begin{aligned} ROI= \frac{Lesions\,Area}{Lung\,Area} \end{aligned}$$
(1)
Fig. 4
figure 4

Samples from mild, moderate, and severe stages and the steps to extract features

3.4.2 Statistical texture features

four statistical texture features are extracted, which are mean, standard deviation, kurtosis and skewness. These features can be calculated by the following equations:

$$\begin{aligned} \overline{x}&=\frac{\sum _{a=1}^{M}xa}{M} \end{aligned}$$
(2)
$$\begin{aligned} s&=\sqrt{\frac{\sum _{a=1}^{M}(xa-\overline{x})^2}{M-1}} \end{aligned}$$
(3)
$$\begin{aligned} K&=\frac{\sum _{a=1}^{M}(xa-\overline{x})^4}{(M-1)\times s^4}-3. \end{aligned}$$
(4)
$$\begin{aligned} S&=\frac{\sum _{a=1}^{M}(x\small {a}-\overline{x})^3}{(M-1)\times s^3.} \end{aligned}$$
(5)

where \(\overline{x}\) is the mean, s is the standard deviation and M is the number of features.

3.4.3 Grey level co-occurrence matrix (GLCM) and grey level run length matrix (GLRLM) texture features

GLCM and GLRLM are types of texture features which can be extracted from the histogram of each image. For GLCM, there are four features are calculated: contrast, correlation, energy, and homogeneity. For GLRLM there are seven features are calculated: run percentage (RP), gray length non-uniformity (GLN), run length non-uniformity (RLN), short run emphasis (SRE), long run emphasis (LRE), low gray-level run emphasis (LGRE) and high gray-level run emphasis (HGRE). in each GLCM and GLRLM each feature is extracted in four orientations (0\(^\circ \), 45\(^\circ \), 90\(^\circ \), and 135\(^\circ \)), then the average is taken.

  1. 1.

    Gray level co-occurrence matrix (GLCM): Four GLCM features are calculated: contrast, correlation, energy, and homogeneity [3]. We can calculate them by the following equations:

    $$\begin{aligned}&\small {Contrast=\sum _{i}^{M}\sum _{j}^{N}(i-j)^2p[i,j]} \end{aligned}$$
    (6)
    $$\begin{aligned}&\quad {Correlation=\frac{\sum _{i}^{M}\sum _{j}^{N}(i-\mu )(j-\mu )p[i,j]}{\sigma ^2}} \end{aligned}$$
    (7)
    $$\begin{aligned}&\quad Energy=\sum _{i}^{M}\sum _{j}^{N}(p[i,j])^2 \end{aligned}$$
    (8)
    $$\begin{aligned}&\quad Homogeneity=\sum _{i}^{M}\sum _{j}^{N}\frac{p[i,j]}{1+|i-j|} \end{aligned}$$
    (9)

    where: M,N are the dimensions of the image p[ij] for i=1,2,3,\(\dots \),M. and j=1,2,3,\(\dots \),N. \(\mu \), \(\sigma \) are the mean and the variance of the image p[ij].

  2. 2.

    Gray level run length matrix (GLRLM): Seven GLRLM features are extracted: RP, GLN, RLN, SRE, LRE, LGRE and HGRE [3]. We can calculate them by the following equations:

    $$\begin{aligned}&\small {RP=\frac{\sum _{i\in Ng}\sum _{j \in Nr}(pij)}{N}} \end{aligned}$$
    (10)
    $$\begin{aligned}&\quad \small {GLN=\frac{\sum _{i\in Ng} [\sum _{j \in Nr}(pij)]^2}{\sum _{i\in Ng}\sum _{j \in Nr}(pij)}}{} & {} \end{aligned}$$
    (11)
    $$\begin{aligned}&\quad \small {RLN=\frac{\sum _{i\in Nr} [\sum _{j \in Ng}(pij)]^2}{\sum _{i\in Ng}\sum _{j \in Nr}(pij)}} \end{aligned}$$
    (12)
    $$\begin{aligned}&\quad \small {SRE=\sum _{i\in Ng}\sum _{j \in Nr}(pij)} \end{aligned}$$
    (13)
    $$\begin{aligned}&\quad \small {LRE=\frac{\sum _{i\in Ng}\sum _{j \in Nr}j^2 (pij)}{\sum _{i\in Ng}\sum _{j \in Nr}(pij)}} \end{aligned}$$
    (14)
    $$\begin{aligned}&\quad \small {LGRE=\frac{\sum _{i\in Ng}\sum _{j \in Nr}(pij/i)^2}{\sum _{i\in Ng}\sum _{j \in Nr}(pij)}} \end{aligned}$$
    (15)
    $$\begin{aligned}&\quad \small {HGRE=\frac{\sum _{i\in Ng}\sum _{j \in Nr}(i^2pij)}{\sum _{i\in Ng}\sum _{j \in Nr}(pij)}} \end{aligned}$$
    (16)

    where: Nr,Ng are the set of different run lengths and the set of different gray levels, respectively. N is the total number of the pixels in the image, calculated by the following equation:

    $$\begin{aligned} N= \sum _{i\in Ng}\sum _{j \in Nr}j(pij){} & {} \end{aligned}$$
    (17)

    pij is the (i,j)\(-th\) entry of the GLRLM.

3.5 Classification

To divide the dataset into training and testing, we used 8-folds cross-validation technique. four machine learning (DT, KNN, SVM and Naïve Bayes) are used to classify the images into four stages: normal, mild, moderate, and severe stage based on the features mentioned previously, see Fig. 5.

Fig. 5
figure 5

Classification steps

3.5.1 K-fold cross-validation

It is a method that uses to assess the machine learning models on a limited dataset. the dataset is split into k subsets randomly. One of the fold is used in the testing and the others are used in the training. this process is repeated k times. [25].

In our work, the best value of K fold is 8. The value of the fold is determined by trial and error. We try different value of the fold and notice anyone give us the lowest error in each stage. Figures 6, 7 and 8 show different values of the fold and the value of the error for each one in each stage (mild, moderate, and severe). From the figure, eight-fold cross-validation gives the lowest error in the mild, moderate, and severe stages.

Fig. 6
figure 6

The resulting of selecting the best number of folds for KNN,SVM, and Naïve Bayes in mild stage. The best value is 8

Fig. 7
figure 7

The resulting of selecting the best number of folds for DT, KNN,SVM, and Naïve Bayes in moderate stage. The best value is 8

Fig. 8
figure 8

The resulting of selecting the best number of folds for DT, KNN, SVM, and Naïve Bayes in sever stage. The best value is 8

3.5.2 Support vector machine (SVM)

One supervised machine learning technique that may be applied to classification or regression problems is SVM. SVM is a discriminative classifier that is officially described by a separating hyperplane. An ideal hyperplane that classifies new classes is produced by the algorithm when we provide labeled training data. In two dimension, This hyperplane divides a plane into two parts with each class lying on each side. The closest points to the hyperplane from two classes that provide the greatest margin, which are used to identify the optimum line. These points are referred to as support vectors [7, 23].

3.5.3 K-nearest neighbor (KNN)

An approach for supervised learning that may be applied to both classification and regression is K-Nearest Neighbors (KNN). KNN attempts to predict the proper class for the test instance by computing the distance between the test instance and all labeled training examples. Then choose the K spots that are closest to the test data [8, 23].

3.5.4 Naive Bayes

Nave Bayes is a type of supervised machine learning that may be used to solve classification challenges. It is presumed that the existence of one feature in a class is unrelated to the presence of any other feature. It operates primarily on the probability principle, computing the odds of the test instance being assigned to each of the available class labels. The label of the test instance is determined by the largest value of these probabilities [18, 23].

3.5.5 Decision tree

The Decision Tree algorithm is a supervised Machine Learning Algorithm that may be used for classification and regression. Every decision tree has two nodes: the decision node and the leaf node. There are several branches at decision nodes. The leaf nodes represent the result of the decisions made to get to the point. A decision tree algorithm is so named because it begins with a root node and then branches out into multiple branches. It just asks a question and accepts a Yes/No response [23, 27].

3.6 Fine-tuning of the models

Each classifier has parameters that we fine-tune through experiments. We choose the best ones that give us the best performance. For the Decision Tree, we try three types; Fine, Medium, and Coarse. Which differ in the number of splits. 100 splits for Fine DT, 20 splits for Medium DT, and 4 splits for coarse DT. From the experiments, we find that Fine DT is the best one, which gives us the best performance. For KNN, we try three types based on the distance function type; Euclidian, Cosin, and Minkowski. Table From the experiments, we find that Weighted KNN is the best one. See table 1 explains our KNN experiments. From the experiments, we find that Weighted KNN is the best one.

Table 1 KNN Types with their parameters

For SVM, we try four types of SVM; Cubic SVM with polynomial kernel function (order 3) and kernel scale 1, and three types of Gaussian SVM (Fine, Medium, and Coarse), all of them are Gaussian Kernel function, but they differ in kernel scale, which is 1, 4.1, and 16 for each one, respectively. From the experiments, we find that Cubic SVM is the best one.

3.7 Evaluating performance

We will investigate a number of measures to assess the efficacy of the suggested classification algorithms, including accuracy, sensitivity or recall, specificity, precision, F−score, and Receiver Operating Characteristic (ROC) area. Figure 9 depicts a 3\(\times \)3 confusion matrix. There are four indices in any confusion matrix which are:

Fig. 9
figure 9

The confusion matrices for ML models with our proposed feature extraction method

  1. 1.

    True positive (TP): the label belongs to the class, and it is correctly predicted.

  2. 2.

    True negative (TN): the label does not belong to the class, and it is correctly predicted.

  3. 3.

    False positive (FP): the label does not belong to the class, but classifier predicted as positive.

  4. 4.

    False negative (FN): the label belongs to the class, but the classifier predicted as negative.

The performance metrics can be calculated by the following equations:

$$\begin{aligned} Accuracy&=\frac{TP+TN}{TP+TN+FP+FN}\times 100. \end{aligned}$$
(18)
$$\begin{aligned} \small Sensitivity&=\frac{TP}{TP+FN}\times 100. \end{aligned}$$
(19)
$$\begin{aligned} \small Specificity&=\frac{TN}{TN+FP}\times 100. \end{aligned}$$
(20)
$$\begin{aligned} \small Specificity&=\frac{TP}{TP+FP}\times 100. \end{aligned}$$
(21)
$$\begin{aligned} F-score&=2\times \frac{Precision \times Recall}{Precision+Recall}\times 100. \end{aligned}$$
(22)

ROC curve is created by plotting the TP rate against the FP rate. The performance is measured by the area under the curve (AUC); the higher AUC is the best.

$$\begin{aligned} \small TP rate&=\frac{TP}{TP+FN}\times 100. \end{aligned}$$
(23)
$$\begin{aligned} \small FP rate&=\frac{FP}{FP+TN}\times 100. \end{aligned}$$
(24)
Fig. 10
figure 10

ROC Curves for ML models with our proposed feature extraction method

4 Experimental results and discussion

All experiments are written in MATLAB 2020a, they ran on the Windows 10 operating system of a desktop computer. The computer uses up to 8 GB of RAM, 1115 GB Hard Disk Drive, and Intel Core i3 @ 3 GHz.

The dataset (2051 scans) is classified into four stages normal with 250 CT scans, Mild with 1075 CT scans, Moderate with 434 CT scans and Severe with 292 CT scans, they rely on the description file that is found with the data set and rely on the CT findings. Then our methodology applies to each scan. 8-folds cross-validation is used to train and test our model. Figures 9, 10 and Table 4 show the results.

Figure 9 shows the confusion matrices for ML classification models DT, KNN, SVM and Naive Bayes. From the figure, the SVM model has the minimum misclassification scans (41 scans).

Figure 10 shows the ROC Curves and the AUC for each model. From the figure, the SVM has the maximum value of the AUC (99%).

Table 4 shows the performance metrics for each model in normal, mild, moderate and severe stage. From the table, SVM is the best performance in all stages. But KNN has the highest specificity in mild and moderate stages.

Table 3 shows the performance metrics for the classification models based on each feature alone (global statistical texture features, GLCM, GLRLM and ROI) features. From this step, we evaluate any feature effect on the performance of each model. From the table, we conclude that the ROI feature is the best feature and it achieves the best performance. The GLCM achieves the lowest performance (Table 2, Fig. 11).

Table 2 Performance metrics for ML models with our proposed model
Fig. 11
figure 11

Performance metrics for ML models with our proposed model

Table 3 The performance metrics for the classification models with Statistical, GLCM, GLRLM, and ROI Features

Figure 12 shows the average accuracy of four stages for four models based on each feature.

Fig. 12
figure 12

The average accuracy of four stages for four models based on each feature

4.1 Comparing with state-of-arts models results and discussion

Table 4 shows a comparison between our work and other studies in a classification of severity of COVID-19 CT scans. From it, our model achieves higher accuracy than other existed models. Our model is reliable, high speed, and robust. So, we can rely on it to determine the degree of infection and provide the appropriate dose during the treatment period.

Table 4 Performance metrics for ML models with our proposed model

5 Conclusions and future works

We developed a systematic technique for COVID-19 identification, lung and lesion segmentation, and patient severity rating using CT scans in this thesis. We presented and analyzed numerous cutting-edge segmentation networks in order to discover the top performing machine learning algorithms. The presented scheme and models obtained exquisite performance levels in segmentation, classification, and infection quantification. Nevertheless, we have proposed a simple, reliable and automatic model to detect the severity of COVID-19 based on the CT scans to three stages: mild, moderate, and severe stage. This model based on finding the ratio of infection ( ratio of the lesions area to the lung area), global statistical, GLCM, and GLRLM features. Three ML algorithms are used to classify CT scans (DT, KNN, SVM, and Naïve Bayes). From experiments, our model achieves 99.12% of accuracy for normal stage, 98.24% of accuracy for mild stage, 98.73% of accuracy for moderate stage and 99.9% of accuracy for severe stage by SVM model. Our model achieves better performance than state-of-arts models also. To summarize, computer-aided detection and quantification provides a reliable, simple, and cost-effective technique of diagnosing COVID-19 cases.

Our model based on CT scans only, as future works, we will consider other factors that effect on the COVID-19 severity such as background disease (like diabetes, etc.), blood test vital, and demographic. these features have useful information can be used to increase the accuracy of our model. As well as we will increase the number of datasets. We will add other machine learning models.