key: cord-0690591-yduljaan authors: Tello-Mijares, Santiago; Woo, Luisa title: Computed Tomography Image Processing Analysis in COVID-19 Patient Follow-Up Assessment date: 2021-04-29 journal: J Healthc Eng DOI: 10.1155/2021/8869372 sha: e00865ab90536d4b36d12282e4dbdf77282f922d doc_id: 690591 cord_uid: yduljaan The rapid worldwide spread of the COVID-19 pandemic has infected patients around the world in a short space of time. Chest computed tomography (CT) images of patients who are infected with COVID-19 can offer early diagnosis and efficient forecast monitoring at a low cost. The diagnosis of COVID-19 on CT in an automated way can speed up many tasks and the application of medical treatments. This can help complement reverse transcription-polymerase chain reaction (RT-PCR) diagnosis. The aim of this work is to develop a system that automatically identifies ground-glass opacity (GGO) and pulmonary infiltrates (PIs) on CT images from patients with COVID-19. The purpose is to assess the disease progression during the patient's follow-up assessment and evaluation. We propose an efficient methodology that incorporates oversegmentation mean shift followed by superpixel-SLIC (simple linear iterative clustering) algorithm on CT images with COVID-19 for pulmonary parenchyma segmentation. To identify the pulmonary parenchyma, we described each superpixel cluster according to its position, grey intensity, second-order texture, and spatial-context-saliency features to classify by a tree random forest (TRF). Second, by applying the watershed segmentation to the mean-shift clusters, only pulmonary parenchyma segmentation-identified zones showed GGO and PI based on the description of each watershed cluster of its position, grey intensity, gradient entropy, second-order texture, Euclidean position to the border region of the PI zone, and global saliency features, after using TRF. Our classification results for pulmonary parenchyma identification on CT images with COVID-19 had a precision of over 92% and recall of over 92% on twofold cross validation. For GGO, the PI identification showed 96% precision and 96% recall on twofold cross validation. e COVID-19 pandemic is an infectious disease that has affected millions of individuals all over the world, and it has caused thousands of deaths since December 19, 2019, according to the World Health Organization (WHO) [1] . On January 30, 2020, the WHO designated the outbreak of this novel coronavirus that had not been seen before in humans to be a "public health emergency of international concern." is was followed by the declaration of a pandemic on March 11, 2020 [1, 2] . COVID-19 presents a significant challenge to medical professions due to the widespread effect of this pandemic. Its influence on the practice of diagnosis and monitoring of ground-glass opacity (GGO) and including follow-up assessment, and in the evaluation of disease evolution [3, 4] . Different clinical studies have shown that chest CT imaging can be helpful in supporting early detection of COVID-19 [5] [6] [7] . Latif et al. [8] present a comprehensive review and they attempt to systematize ongoing data science activities in this area. Since the COVID-19 outbreak was identified in December 2019, there have been increasing efforts to develop different artificial intelligence methods to screen patients for COVID-19 based on medical images such as CT scans [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] . Wang et al. [9] used deep-learning methods to detect radiographical changes in COVID-19 patients. Chen et al. [10] proposed the UNet++ architecture to detect suspicious lesions on CT scans. Ophir et al. [11] used two-and threedimensional convolutional neural networks (CNNs) to calculate the corona score (which represents the evolution of the COVID-19 infection in the lungs). Li et al. [12] developed a neural network (COVNet) to extract visual features and modified a residual network with 50 layers to detect COVID-19 using CT scans. Mobiny et al. [13] developed a Detail-Oriented Capsule Networks (DECAPS) learning architecture that could identify finegrained and distinguishing image features to classify COVID-19 based on CT scans. Wang et al. [14] introduced a COVID-Net for the automatic interpretation of chest radiographs from COVID patients. Chaganti et al. [15] proposed a system that identifies suspicious lung regions in CTscans using a deep reinforcement learning algorithm, and they measured the abnormality and severity caused by COVID-19. Butt et al. [16] extracted features from CT scans using an architecture based on ResNet18, to identify segments of pathological lung regions that then serve as the input to a classifier for predicting COVID-19 disease. Al-Karawi et al. [17] showed the development and performance of machine learning schemes by an innovative frequency domain algorithm, which is called an FFT-Gabor scheme. He et al. [18] proposed a Self-Trans approach, which synergistically integrates contrastive self-supervised learning with transfer learning to learn powerful and unbiased feature representations for reducing the risk of overfitting. Chowdhury et al. used CNN to identify COVID-19 patients based on chest X-ray images [19] . Yang et al. developed a deep-learning-based CT diagnosis system (DeepPneumonia) to assist clinicians with identifying patients with COVID-19 [20] . Shan et al. used the "VB-Net" neural network to segment COVID-19 infection regions in CT scans [21] . Shen et al. proposed an infection-size-aware random forest (iSARF) method that can automatically categorize subjects into groups with different ranges of infected lesion sizes [22] . Polsinelli et al. [23] proposed a light CNN design based on the SqueezeNet model, as an efficient way to distinguish between COVID-19 CT images and other CT images (community-acquired pneumonia and/ or healthy images). Fan et al. [24] proposed a novel COVID-19 lung infection segmentation deep network (Inf-Net) to automatically identify infected regions from the chest CT scans. Amyar et al. [25] proposed a multitask deep-learning model to jointly identify COVID-19 patients and distinguish COVID-19 lesions from chest CT images. Wu et al. [26] developed a novel joint classification and segmentation (JCS) system to allow a real-time and explainable COVID-19 diagnosis. Here, we propose an effective and efficient method for segmenting and identifying GGO and PI in patients with typical COVID-19 CT images from the Zhao dataset [27] ( Figure 1 (a)-1(f )). Additionally, the ground truth was preset for pulmonary parenchyma (Figure 1 (g)-1(l)), GGO, and PI (Figure 1 (m) and 1(n)). We showed excellent results based on precision and recall using a set of real, practical, complex, and representative data. Our motivation is to contribute a robust two-level classifier system to help improve both the segmentation and classification of pulmonary parenchyma from CT scans of patients with COVID-19 based on GGO and PI identification and then have the medical expert perform the patient assessment and evaluation at a later time. Our work is a medical aid tool. e next section of this paper describes the proposed segmentation and the two-level cascade classification algorithm. e segmentation method was based on a mean shift for oversegmentation followed by morphologic and superpixel-SLIC clustering process for regrouping the pulmonary parenchyma CT images using a cluster feature extraction process and tree random forest (TRF) as the first objective. Next, watershed segmentation was applied to the mean-shift clusters only in the identified pulmonary parenchyma zone, followed by identification of the GGO and PI using the watershed cluster feature extraction. ey were then classified using the TRF method. Section 3 presents the classification results that were obtained, and Section 4 discusses the results and concludes the paper. Figure 2 illustrates the stages of the proposed method. e dataset was provided by Zhao (2020) [27] (Figure 1 ; images from the Zhao (2020) dataset [27] ). ey built a publicly available COVID-CT dataset that contained 349 CT images from patients who were positive for COVID-19. e images were not captured under the same conditions, and therefore, the resolution was different and the images were preprocessed. In this study, MATLAB was used to process all of the images, such as performing image preprocessing, segmentation, and feature extraction. is is discussed in detail in the sections below. Oversegmentation. Input grey images were first truncated by setting the two least significant bits of every pixel and channel to 0, thereby obtaining images with 64 levels. ese images were coded as 0, 4, 8, . . ., 252 for simplicity. is step assumes normalization of the CT input images, and it removed noise and perceptually irrelevant details. It also resized all the 200×300 pixels (Figure 3(a) ). is first segmentation stage consisted of oversegmentation. We propose the use of the mean-shift algorithm [28] to obtain an oversegmented image. Mean shift is a which was selected so that the chest was segmented into more than one region; this is required for the later regrouping process to be effective. One goal of this paper is to develop an automatic and robust mean-shift and superpixel-SLIC method for accurate segmentation of pulmonary parenchyma identification in CT imaging. e segmented image (Figure 3(b) ) contains a set of clusters that are each characterized by a vector of [L] values. To finally identify a single region in a candidate pulmonary parenchyma, we proposed an effective regrouping technique based on superpixel-SLIC approaches (see below, Figure 4 (a)). One of the original contributions of this paper is that our first segmentation stage is built on the mean-shift and superpixel-SLIC combination. In this paper, a superpixel-SLIC is a connected set of pixels that share similar properties. We obtain superpixel-SLIC using three-dimensional meanshift clustering that incorporates joint spatial and intensity features. e superpixel-SLIC expresses the local structure of the data in the three-dimensional feature space. Mean-shift clustering reduces data variation, thereby helping to produce accurate segmentation. By connecting superpixel-SLIC instead of pixels, the algorithm produces better results. e first objective was to regroup the pulmonary parenchyma ( Figure 4 (b)), the body, and the background. We used superpixel-SLIC approaches over the mean-shift L band to regroup the oversegmented image clusters (Figure 3 (b)), and we used the simple linear iterative clustering (SLIC) algorithm [30] . is algorithm groups pixels into regions with similar values based on their intensity similarity and proximity in the image joint spatial. Region size and high compactness, which are the two parameters of SLIC, were 100 and 20, respectively. Twenty-four iterations were used in the clustering phase of the algorithm. e next step was to obtain the average value of each new object in the superpixel-SLIC L image (Figure 4 (a)) using the oversegmented image. Applying morphology and superpixel-SLIC based on the oversegmented mean-shift clustered image has some advantages for regrouping (Figure 4(a) ). From the new image, which contained the grouped pulmonary parenchyma parts and the body and the background parts (Figure 4 (a)), we extracted the characteristics of the position, grey intensity, texture, and mainly the spatial-context-saliency to classify pulmonary parenchyma and nonpulmonary parenchyma (body and the background, Figure 4 (b)). In Figure 4 (c), 105 objects (numbers in red) and 276 connections between clusters (green lines) are shown. e blue dots are the Cartesian coordinates [x, y] of each superpixel-SLIC centroid, the pulmonary parenchyma is identified, and the spatial-context-saliency is shown by the green lines (see below). We obtained a grey intensity feature, and one way to distinguish between different textures is to compare L levels using first-order statistics. First-order statistics were calculated based on the probability of observing a particular pixel value at a randomly chosen location in the image. ey depend only on individual pixel values and not on the interaction of neighboring pixel values. e average (1) is the mean of the sum of all of the intensity values in the image. is was selected because of the specific intensity appearance of the chest, including average pixel values in L grey intensity space (µ SPc ) of the aforementioned clusters that were obtained from the statistical average of the superpixel-SLIC clusters image (Figure 4(a) ). We extracted the following four second-order statistics texture features: (1) contrast; (2) correlation; (3) energy; and (4) local homogeneity from the grey-level cooccurrence matrix (GLCM), which is the matrix M, as shown below (the measures are calculated for d � 0°). Haralick's GLCM [31] has been used successfully for biomedical image classification [32, 33] . We used the cooccurrence matrix for the entire L intensity. e contrast descriptor (cn SPc ) is a measure of local variation in the image. It has a high value when the region within the range of the window has high contrast. Correlation (cr SPc ) of the texture measures the relationship between the different intensities of grey. Mathematically, the correlation increases when the variance is low, meaning that the matrix elements are not far from the main diagonal. Energy (e SPc ) is the sum of the squared elements in the matrix the grey-level cooccurrence, which is also known as the uniformity or the second element of angular momentum. Local homogeneity (LH SPc ) provides information about the local regularity of the texture. When the elements of the cooccurrence matrix are closer to the main diagonal, the value of the local homogeneity is higher. In equation (3), μ and σ are the mean and standard deviation values of GLCM in each row (or column, because of symmetry), respectively. We used the original quantized grey image to analyze the original texture in one direction (d � 0°radians), which was used to construct the GLCM. Twelve spatial-context-saliency features, which were selected to obtain the characteristics of each object with respect to its neighbors, were based on the salience of each object obtained in a global way (i.e., the comparison of each object with respect to the other objects in the image) [34] . Journal of Healthcare Engineering ese features use element uniqueness Ui (intensity or texture) and element distribution Di (position). Based on Perazzi et al. [34] , we obtained spatialcontext-saliency features in a local way (i.e., by comparing each object with respect to its neighbors in the image) using the objects that were grouped by the superpixel-SLIC (Figure 4(a) ). We propose a four-element uniqueness Ui (using grey intensity and histograms). Each object i has a different set of j to m neighbors; ci and cj � {c1, c2, ... cm} represent the value of the intensity (average µ SPc ) and also represent the value of the texture (cn SPc , cr SPc , e SPc , and Figure 4 (c)) use the histogram values, and we obtained the Euclidean distance ED (6) and the Bhattacharyya distance BD (7) of each object hi with respect to its neighbor hj. Additionally, with the grey (average μ SPc ) and the texture values (cn SPc , cr SPc , e SPc , and HL SPc ), we obtained the average of the difference (8) and the energy (9) of each object ci with respect to its neighbors cj. We used the Gaussian weight of the position w ij (p) (10), and we effectively combined the local contrast estimation with the control of the radius of influence for the singularity operator. e local function w ij (p) yields a local contrast term, which tends to overemphasize object boundaries in the saliency estimation, as seen in [34] [35] [36] . Additionally, σ p controls the range of the uniqueness operator and is the variance of the distances of the position p i for object i to positions p j for object j. We subsequently obtained 19 characteristics of each object. Each superpixel-SLIC cluster that was characterized was then classified as being either pulmonary parenchyma or a nonpulmonary parenchyma in the first stage of the twolevel cascade classification scheme. e primary contribution of this work is the proposed segmentation and characterization method for the classification of pulmonary parenchyma on CT images (Figure 4(b) ). To classify segmented pulmonary parenchyma into CT images, we explored the use of classification approaches that were implemented in Weka (Waikato Environment for Knowledge Analysis) [37, 38] . For the first stage, which is for segmentation and identification of pulmonary parenchyma, we used TRF [39] ; the values that were used as parameters in the random forest algorithm are as follows: the number of iterations, 100; the number of trees to be generated, numTrees � 100; the maximum depth of the trees, unlimited maxDepth � 0; and the random number seed to be used, seed � 1. In the second stage in the next section, we identified the GGO and PI from the pulmonary parenchyma region (Figure 5(a) ). Our experimental results (see Section 3) were obtained for this two-level cascade classification scheme. e final proposed part of this second stage was the GGO and PI identification ( Figure 5(b) ) that was applied to the watershed transform to the mean-shift clusters only in the identified pulmonary parenchyma zone. Once the pulmonary parenchyma zone on the CT image was identified ( Figure 5(a) ), segmentation of the watersheds was initialized for this zone ( Figure 5(b) ). e concept of watersheds [40] in image processing is based on consideration of an image in three-dimensional space, with two spatial coordinates compared to the intensity. e value of the intensity is assumed to be the elevation information. Obtained pixels of watershed lines, which separate neighboring catchment basins and, consequently, separating different characteristic parts of the image, can cluster the oversegmentation that is obtained by the mean shift and focus on finding regions of interest (GGO and PIs). Because most of the PI usually has denser areas, with a pixel area intensity that is higher than that outside the area, a high gradient and well-delimited image across the infiltrate boundaries is expected. GGO is more textured, and the same high gradient and well-delimited image is expected. An additional characteristic that is considered on a CT scan that is positive for COVID-19 is GGO, which usually has a bilateral, peripheral, and subpleural distribution, and it is usually predominantly located in the posterior sectors. is is why more weight is given to watershed cluster peripherals for CT screenings that were submitted to our program. us, we selected ten specific blue points ( Figure 5(c) ), five left and five right delimitations of the pulmonary parenchyma borders. e efficient separation of the true GGO and PI regions from the total watershed regions requires the generation of meaningful features that have very good discriminative ability such as the position Cartesian coordinates [x, y] and the first-order statistical average (11) , which is the same as (1) that was applied to watershed clusters (μ w ). Four second-order statistics texture features, contrast, correlation, energy, and local homogeneity (12)- (15) , from the GLCM, which is the matrix M as shown below (the measures are calculated for d � 0°) and which is the same as (2)-(5), were applied to watershed clusters. We also integrated the entropy features of gradient X (16) and gradient Y (17) , and based on the entropy, we measured the randomness that can be used to characterize the texture of the gradient X and gradient Y in each watershed cluster. For the next ten Euclidean distance features d (18), we started from the observation of five left and five right delimitations of the pulmonary parenchyma borders (br [x,y], (18) ). ese are regions of interest for finding GGO and PI because they are close to the border regions (10 blue dots, Figure 5 (c)). is is the main reason why pulmonary parenchyma is first segmented and identified to obtain fiveposition Cartesian coordinates [x, y] on the left and right sides (10 blue dots, Figure 5 (c)). From these ten positions, we obtained the Euclidean distance features d (18) with the position Cartesian coordinates of each watershed cluster (wc [x,y], (18) ). To eliminate false-positive findings, we introduced seven global saliency features GS (19) , based on [41] . First, saliency was defined for each watershed cluster as the weighted sum of the cluster's contrast to all other watershed clusters in the segmented image ( Figure 5(c) ). e weights were set according to the spatial distances with farther regions being assigned smaller weights; salient feature contrast, which characterizes the intensity, the entropy of gradient X and gradient Y, and the texture intensity (contrast, correlation, energy, and local homogeneity) provided the seven global saliency features. For a watershed cluster wc k , we computed its saliency value by measuring its feature contrast in relation to all other clusters in the image, where w (wc i ) is the weight of cluster i and D wc (wc k , wc i ) is the feature distance metric between the two regions (20) . We obtained 30 characteristics from each cluster. e main contribution of this work is the proposed identification method for the classification of GGO and PIs ( Figure 5(b) ). To classify segmented watershed clusters into a normal or abnormal region (GGO and PIs), we have explored the use of the TRF classification approach [39] , as described in the last section, and the values used as parameters were numTrees � 100, maxDepth � 0, and seed � 1. Indicators. Several quality indicators have been obtained to quantitatively assess the PP and GGO-PI identification results and the performance of the TRF technique for the first and second stages, respectively. We divided them into final or external quality indicators, which evaluated the final identification results and are useful for external comparison with other studies, and internal quality indicators, which are useful to evaluate the internal behavior of the proposed classification options (% Training Set -% Test Set). For the external indicators, we assumed the following: let P be the number of pulmonary parenchyma clusters in the dataset and let TP, FP, and FN be the number of true positives, false positives, and false negatives, respectively. We then defined the following: sensitivity, recall, or true positive rate: TP Rate � TP /(TP + FN); precision or positive predictive value: PPV � TP /(TP + FP); and the F-Measure � (2 × recal1 × precision) /(recall + precision). Because the proposed algorithm first selects superpixel-SLIC clusters that are then characterized and separated into pulmonary parenchyma and nonpulmonary parenchyma, we can further evaluate the classification performance of the three selected classification schemes via the internal indicators, as follows: let N be the number of nonchest candidates resulting from the application of the proposed method to the complete dataset and let TN be the number of true negatives after classification. We can then define the fall-out or false-positive rate as FP Rate � FP /(FP + TN), and the area under the receiver operating characteristic curve as ROC area. e above indicators are also applied for ground-truth data for GGO and PI identification. A medical expert defined a region around the pulmonary parenchyma for a defined ground-truth comparison (see Figure 1 (g)-1(l)). e expert also selected and diagnosed the GGO and PI ground-truth comparison (see Figure 1 (m)-1(r)). us, good and precise segmentation and classification are both desired. e results of pulmonary parenchyma selection and the feature extraction phases, over the described dataset, are a collection of 43,928 candidate regions as follows: 28,080 nonpulmonary parenchyma (body and background) and 15,848 expected pulmonary parenchyma, which are each characterized by a 19-dimension feature vector for the first stage of the Two-Level Cascade Classification Scheme using TRF. For GGO and PI identification, a collection of 27,340 candidate regions was used as follows: 2,684 normal and 24,692 expected nonnormal (GGO and PIs), which were each characterized by a 30-dimension feature vector for the second stage and which also used TRF. Table 1 summarizes all of our quantitative results. In the most difficult but more realistic classification experiment, twofold cross validation (i.e., the dataset was divided into two equal parts, one used for training and the other for testing) was used. is finding confirms the reported advantages of TRF over other state-of-the-art classifiers. As expected, as the value of s increased in the s-fold cross Journal of Healthcare Engineering validation, our results improved until the full set case, in which TRF yielded full precision and recall. Our qualitative results for identification are shown in Figure 6 , with blue superimposed over correctly detected pulmonary parenchyma. ese results reveal that the proposed segmentation method can successfully identify pulmonary parenchyma regions. GGO and PIs are indicated with a superimposed red line. ere is a lack of publicly accessible datasets or evaluation scenarios that allow for a fair comparison among methods, and code for the reported methods is unavailable. us, we have chosen to present only our results on CT image classification. However, based on the radiographic changes in COVID-19 cases on CT images, many studies have hypothesized that deep-learning artificial intelligence methods could extract the specific graphic characteristics of COVID-19 and provide a clinical diagnosis before pathogenic molecular biological testing, which would save critical time for disease control [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [22] [23] [24] [25] [26] . e distinctive characteristics of COVID-19 are the bilateral distribution of irregular shadows with different degrees of radiopacity that are considered to be similar to sperylated or ground glass [2] . One study used a CNN to help with the detection and prediction of COVID-19 [9] . e classifying systems were effective and had an accuracy of 73.1%, specificity of 67%, and sensitivity of 74%. Another study's deep-learning model showed comparable performance to that of an expert radiologist for detecting COVID-19 pneumonia, and the approach provided a classification accuracy of about 95.24% [10] . Another study developed an artificial intelligence-based automated CT image analysis tool for the detection, quantification, and tracking of coronavirus [11] . ey achieved classification results for coronavirus versus noncoronavirus with an area under the curve (AUC) of 99.6%. Another study developed a deep-learning neural network model to extract visual features from volumetric chest CT examinations to detect COVD-19, and the AUC was 96% [12] . A novel learning architecture called Detail-Oriented Capsule Networks (DECAPS) was proposed for the automatic diagnosis of COVID-19 from CT scans. e model achieved 84.3% precision, 91.5% recall, and 96.1% AUC [13] . Another study introduced COVID-Net, a deep CNN design that was tailored to detect COVID-19 cases based on chest X-ray images [14] . For COVID-19 cases, it achieved a good accuracy of 93.3%, a sensitivity of 91.0%, and a high positive predictive value of 98.9%. Measures of disease severity and a method based on deep learning and deep reinforcement to compute them have been proposed [15] . e Pearson correlation coefficient between the method's prediction and ground truth for positive COVID-19 scans was calculated as 0.92 for the percentage of opacity (P < 0.001), 0.97 for the percentage of high opacity (P < 0.001), 0.91 for the lung severity score (P < 0.001), and 0.9 for the lung high opacity score (P < 0.001). Another study compared CNN models to classify CT samples between COVID-19, influenza viral pneumonia, and no infection [16] . We compared that study with one that combined two-and three-dimensional deeplearning models with the latest clinical understanding and achieved an AUC of 0.996 (95% confidence interval (CI): 0.989-1.00) for coronavirus versus noncoronavirus cases of thoracic CT studies. ey calculated a sensitivity of 98.2% and a specificity of 92.2%. An innovative frequency domain algorithm called the FFT-Gabor scheme was proposed, which had an average accuracy of 95.37%, sensitivity of 95.99%, and specificity of 94.76% [17] . Sample-efficient deep-learning methods that can achieve high diagnostic accuracy of COVID-19 from CT scans were developed [18] . e approach achieved an F1 of 0.85 and an AUC of 0.94 in diagnosing COVID-19 from CT scans. Another study classified chest X-ray images with a CNN into two different schemes (first normal and COVID-19; second normal, viral, and COVID-19) [19] . e classification accuracy, precision, sensitivity, and specificity of both schemes were 99.7%, 99.7%, 99.7%, and 99.55% for the first scheme and 97.9%, 97.95%, 97.9%, and 98.8% for the second scheme, respectively. A deep-learning-based CT diagnosis system (Deep-Pneumonia) was developed to identify patients with COVID-19 [20] . e model can accurately distinguish COVID-19 patients from others with an excellent AUC of 0.99 and recall (sensitivity) of 0.93. Another study developed a deep-learning-based segmentation system for quantitative COVID-19 infection assessment from chest CT scans [21] . A quantitative evaluation showed high accuracy for automatic delineation of infection regions (dice similarity coefficient � 91.6% ± 10.0%). e iSARF method was proposed, and the experimental results showed that it yielded a sensitivity of 0.907, specificity of 0.833, and accuracy of 0.879 [22] . ey proposed a light CNN that was design based on the SqueezeNet model to efficiently distinguish between COVID-19 CT images and other CT images [23] . It had 83.00% accuracy, 85.00% sensitivity, 81.00% specificity, and 81.73% precision and an F1score of 0.8333. Another study proposed a novel COVID-19 lung CT infection segmentation network called Inf-Net [24] . e quantitative results of the detected infection regions included a structure measure of 0.781, enhanced-alignment measure of 0.838, and mean absolute error of 0.082. A multitask deep-learning model jointly identified COVID-19 patients and segmented COVID-19 lesions from chest CT images [25] . e AUC curve was greater than 93% for the classification. ey developed a novel JCS system to make a real-time and explainable COVID-19 diagnosis [26] . It obtained an average sensitivity of 95.0% and a specificity of 93.0% on a classification test set. e chest CT scan is undoubtedly a tool that helps to obtain a presumptive diagnosis of COVID-19 disease, which provides radiographic images that are suggestive of a multiple focus on pneumonia and COVID-19 disease. e scans are accompanied by the patient's clinical context because the images are characteristic of the disease and can also occur in other types of pneumonia that is caused by microorganisms other than the SARS-CoV-2 virus. However, the chest CT is an imaging tool that provides radiographic information at the thoracic level and is undoubtedly a tool to evaluate the follow-up and prognosis of the disease. RT-PCR is a biomolecular test that is capable of detecting the DNA chain sequence of the SARS-CoV-2 virus. erefore, it is considered to be the gold standard for a precise diagnosis of COVID-19. Not many studies have focused on automatically analyzing GGO and PIs on CT scans with COVID-19. e final aim of these studies is to identify GGO and PIs on CT scans with COVID-19, for which GGO and PI identification are considered to be a first and crucial step for timely detection of images that are suggestive of pneumonia secondary to COVID-19. is is an important step because the prognosis and functionality of the patient after recovery from the disease depend on timely treatment. On the basis of these studies, we can confirm to some extent that our approach is valid based on our TRF classification results for GGO-PI. e GGO-PI identification showed 96% precision and 96% recall on twofold cross validation compared with other similar state-of-the-art studies on CT images. If the analysis is performed on chest CT scans of patients without COVID-19 (50 additional images), better representation with 96.4% precision and 97% recall is obtained. Identification of GGO and PI is required for the derivation of diagnostic conclusions and the characterization of the contents in the pulmonary parenchyma area. Automated identification of the GGO and PI in the CT images from these COVID-19 patients is a challenging issue for medical image processing. In this research, we effectively overcame the problem of locating and detecting GGO and PI, and we developed a method to identify pulmonary parenchyma in tomography images. Moreover, we propose a meaningful feature set for the detected clusters, which results in the efficient ability to distinguish between the true abnormal classes (GGO and PIs) using clustering algorithms. e main advantage of the proposed method is that it can be applied directly to the CT images that were obtained using an X-ray scan, without any observer interference, and it can accurately and in an automated fashion identify first the pulmonary parenchyma and later GGO and PI. Additional materials are available at this link: https://drive. google.com/drive/folders/18qAQqoEFn_ebOY92q-oXNOMZnUEzaufA?usp�sharing. is link contains all the quantitative classification results for download: Folder "Identification images of the lung parenchyma and images of ground-glass opacity and pulmonary infiltrate". Segmentation of the Pulmonary parenchyma identification images and ground-glass opacity and pulmonary infiltrate images (in bmp format). WEKA PP data features "PP.arff". 43928 × 20 features in Weka for TRF. Run "PP.arff" for obtained the classification results. GGO-PI data features "GGOPI.arff". 27340 × 31 features in Weka for TRF. Run "GGOPI.arff" for obtained the classification results. e authors declare that they have no conflicts of interest. World Health Organisation, Statement on the second meeting of the International Health Regulations World Health Organisation, WHO Director-General's Opening Remarks at the Media Briefing on COVID-19 -11 e role of chest imaging in patient management during the COVID-19 pandemic Review of artificial intelligence techniques in imaging data acquisition, segmentation, and diagnosis for COVID-19 Correlation of chest CT and RT-PCR testing for coronavirus disease 2019 (COVID-19) in China: a report of 1014 cases Sensitivity of chest CT for COVID-19: comparison to RT-PCR Imaging profile of the COVID-19 infection: radiologic findings and literature review Leveraging data science to combat COVID-19: a comprehensive review A Deep Learning Algorithm Using CT Images to Screen for Corona Virus Disease (COVID-19), medRxiv Deep Learning-Based Model for Detecting 2019 Novel Coronavirus Pneumonia on High-Resolution Computed Tomography: A Prospective Study Rapid AI development cycle for the coronavirus (COVID-19) pandemic: initial results for automated detection & patient monitoring using deep learning CT image analysis Using artificial intelligence to detect COVID-19 and community-acquired pneumonia based on pulmonary CT: evaluation of the diagnostic accuracy Radiologist-Level COVID-19 detection using CT scans with detail-oriented capsule networks Covid-net: A Tailored Deep Convolutional Neural Network Design for Detection of Covid-19 Cases from Chest Radiography Images Quantification of Tomographic Patterns Associated with COVID-19 from Chest CT Retracted article: deep learning system to screen coronavirus disease 2019 pneumonia Machine Learning Analysis of Chest CT Scan Images as a Complementary Digital Test of Coronavirus (COVID-19) Patients, medRxiv Sample-efficient Deep Learning for COVID-19 Diagnosis Based on CT Scans, medRxiv Can AI help in screening viral and covid-19 pneumonia? Deep Learning Enables Accurate Diagnosis of Novel Coronavirus (Covid-19) with CT Images, medRxiv Lung infection quantification of Covid-19 in CT images with deep learning Large-scale screening of Covid-19 from community acquired pneumonia using infection sizeaware classification A Light CNN for Detecting COVID-19 from CT Scans of the Chest Inf-net: automatic COVID-19 lung infection segmentation from CT images Multi-task Deep Learning Based CT Imaging Analysis for COVID-19: Classification and Segmentation, medRxiv An Explainable COVID-19 Diagnosis System by Joint Classification and Segmentation COVID-CT-dataset: A CT Scan Dataset about COVID-19 e estimation of the gradient of a density function, with applications in pattern recognition Mean shift: a robust approach toward feature space analysis SLIC superpixels compared to state-of-the-art superpixel methods Statistical and structural approaches to texture Nominated texture based cervical cancer classification Cirrhosis classification based on texture classification of random features Saliency filters: contrast based filtering for salient region detection Contrast-based image attention analysis by using fuzzy growing Fast high-dimensional filtering using the permutohedral lattice A machine learning workbench WEKA: a machine learning workbench Applying a machine learning workbench: experience with agricultural databases Random forests Digital Image Processing Global contrast based salient region detection is study was supported by Internal Grant "ITSLerdo #001-POSG-2020." is work was supported by Internal Grant no. 312304 from the F0005-2020-1 CONACYT Support for Scientific Research Projects, Technological Development, and Innovation in Health in the Face of Contingency by COVID-19.312304 from the F0005-2020-1