Deep Transfer Learning Approaches in Performance Analysis of Brain Tumor Classification Using MRI Images

Brain tumor classification is a very important and the most prominent step for assessing life-threatening abnormal tissues and providing an efficient treatment in patient recovery. To identify pathological conditions in the brain, there exist various medical imaging technologies. Magnetic Resonance Imaging (MRI) is extensively used in medical imaging due to its excellent image quality and independence from ionizing radiations. The significance of deep learning, a subset of artificial intelligence in the area of medical diagnosis applications, has macadamized the path in rapid developments for brain tumor detection from MRI to higher prediction rate. For brain tumor analysis and classification, the convolution neural network (CNN) is the most extensive and widely used deep learning algorithm. In this work, we present a comparative performance analysis of transfer learning-based CNN-pretrained VGG-16, ResNet-50, and Inception-v3 models for automatic prediction of tumor cells in the brain. Pretrained models are demonstrated on the MRI brain tumor images dataset consisting of 233 images. Our paper aims to locate brain tumors with the utilization of the VGG-16 pretrained CNN model. The performance of our model will be evaluated on accuracy. As an outcome, we can estimate that the pretrained model VGG-16 determines highly adequate results with an increase in the accuracy rate of training and validation.


Introduction
Tumor is also termed as neoplasm produced by uncontrolled growth of anomalous cells [1]. A brain tumor (cancer) is a mass of abnormal tissues found in the central spinal canal or brain [2], wherein few cells grow and spread uncontrollably, ostensibly unregulated by the natural process that controls normal cells. Our brain is encapsulated by the skull, which is very intransigent. e tumor grows rigorously inside such a restricted space and hampers the natural functioning of the brain. e main cause of deadly cancerous cells in the brain can be associated with substantial conditions like the disproportionate inhale of inorganic chemicals or ancestral disorders. Brain tumors can take the form of benign (noncancerous) and malignant (cancerous). When benign, precarcinoma, or malignant tumors grow, the pressure inside the skull beneath increases causing several complications for humans. is will result in traumatic brain damage and can be life-threatening. Brain tumor detection at early stages is very important in increasing the human survival rate. Several techniques [1] were suggested for the prediction of tumors in the brain. Tumors in the brain organ are categorized into three types that are generally recognized as glioma, meningioma, and pituitary tumors [3]. Medical image technology in the e-healthcare domain plays a vital role in today's emerging field. Research of medical experts faces a lot of obstacles in identifying deadly brain tumor cells. Diagnosis of brain tumors at early stages has become a crucial task in medical science as it is ranked tenth in the leading diseases. Cancerous tumors of the brain occur at different locations and have varying dimensions and sizes.
Generally, to yield images of the human body's soft tissue, various techniques are used for medical image analysis. MRI images are one of them that are used by medical experts. It is a noninvasive technique used for accurate image data analysis of human brain tumors in determining the patient's condition [4]. It is highly applicable due to tissue contrast normalization and image quality resolution. e MRI images provide the genetics, physiology, chemistry, and biological information about abnormalities present in the brain [5]. Tumors are classified into several forms based on their origin and the nature of the cell. e primary vertebrate cerebrum tumors appear in the central hemispheres area, whereas secondary brain tumors locate their path from other human organs to the brain. e most common primary brain tumors are glioma, meningioma, and pituitary tumors in Figure 1.
Gliomas usually arise from the internal (gluey) cell of the brain. is cell, named the glial cell, helps in the proper functioning of brain neurons. World Health Organization (WHO), grade II-III glioma as non-glioblastoma and grade IV defined as glioblastoma [6]. ere exist three forms of normal glial cells that produce deadly tumors [7]. Astrocytoma tumor will be produced by astrocyte cells (including glioblastomas), an oligodendrocyte will result from oligodendroglioma cells, and ependymoma occurs through ependymal cells. Tumors that include the combination of these distinctive glial cells are referred to as mixed glioma. Meningioma tumors arise from cerebral membranes that encapsulate the brain and spinal cord within the inner portion of the skull. Particularly, these passive growth tumors occur on the three membrane layers called meninges. Pituitary adenoma (tumors) arises abnormally in the pituitary gland region that exists inside the skull, between the brain and nasal passages. Meningioma and pituitary tumors are benign that do not escalate to other tissues, cells, and other parts of the human body. Glioma is a malignant tumor that spreads across the human body organs.
Image classification is an essential and extensive domain within the research area of deep learning and computer vision. Image classification refers to the process of assigning labels to an image into one of a number of predefined classes. e classification includes image data acquisition, image data preprocessing, image object detection, image segmentation, feature extraction, and image object classification. Image classification is the most significant and crucial task, especially in medical healthcare domains, including biomedical imaging, detecting, and diagnosing the disease accurately, which helps radiologists to improve diagnostic efficiency and provide a better path for surgical treatment.

Motivation.
e brain tumour is a life-threatening neurological disorder that occurs due to the uncontrollable growth of abnormal cells in the human nervous system. Over the last twenty years, the incidence of brain tumours has been increasing in all ages. It has been predicted recently as the third most extensive cancer recognized majorly in adults and teenagers. Around the world, according to International Agency for Research on Cancer (IARC), proportionately more than 136000 populations are investigated for brain tumors per year, with more than 87000 loss of survival rate in 2017 [8]. Medical experts put inconsistent effort to conquer the complications of brain tumors; According to the World Health Organization (WHO) statistics, it is estimated that 251,329 people die from cancerous brain disease approximately every year. erefore, accurate brain tumor diagnosis is an essential task for patient survival and providing effective medical treatment. Recently, many researchers proposed various multidisciplinary models, including medical science knowledge, mathematics, and computer science, to understand the disease and identify more adequate surgical methods. Various image modalities are referred for brain images data analysis, such as Magnetic Resonance Imaging (MRI), Single Photon Emission Computer Tomography (SPECT), BIOPSY scan, and computer tomography (CT). e most common imaging techniques are MRI and CT undertaken for brain scans to predict the existence of brain tumors and to identify their position for preferred professional treatment benefits. e proper treatment options depend on the brain tumor's location, size, shape, category, and grade. Medical treatment options also depend on whether or not the tumor is spreading and affecting the other organs of the body or within the part of the central nervous system (CNS).

Importance of MRI Biomedical Image Processing.
erefore, an MRI scan is considered the most efficient tool compared with other image modularities. A well-experienced radiologist can investigate the brain MRI scan and decide on treatment options based on the type of tumor [9].
MRI [10] is one of the essential imaging techniques used precisely in brain tumor detection. Brain tumor is one of the most deadly diseases occurring in the central nervous system of the human body. MRI medical imaging technique uses a very powerful magnet, magnetic field gradients, and computer-generated radio waves to scan and display internal organ images within the human body. MRI scan plays a vital role for healthcare professionals in examining brain tumors to provide better patient treatment. MRI is an analytical development in the medical field that produces high-resolution images to detect and diagnose tumor existence in the human brain organ. MRIs create more detailed high-quality images than CT scans. It uses magnetic radio waves to scan the organ within the body. erefore, an MRI scan is considered the most efficient tool compared with other image modularities. A well-experienced radiologist can investigate the brain MRI scan and decide on treatment options based on the type of tumor. Investigation of the MRI medical image by the radiologist is a time-consuming process, and the efficiency in decisionmaking confides their expertise in the medical field. erefore, a computerized brain tumor classification process helps radiologists in their investigation and reduces their interferences extremely.

Why Deep Learning Methods?
Machine learning techniques are the first automation classification methods in the image classification process for detecting deadly tumor diseases. Due to the need for preprocessing of image data and extraction of distinct features in MRI images, they take a long time to classify the tumor. e main disadvantage of machine learning methods is that classification process accuracy mainly depends on handcrafted features identified by feature extraction methods. Acquisition of data based on different algorithms is one of the significant challenges. e acquired data can be highly error-prone, as the data required for training the model must be clean, structured, and accurate. Also, the selection of machine learning algorithms for a particular problem is based on data collection; additionally, machine learning algorithms take more computational time to process a huge amount of data. In spite of machine learning methods limitations, some researchers proposed methods and achieved classification accuracy rates of 78% and 86% based on textual or statistical feature extraction methods.
Artificial intelligence and advanced deep learning models based on transfer learning and artificial neural networks achieved a higher classification accuracy rate in self-learning the image features to classify MRI images without the need for feature extraction methods suggested by [11,12]. Transfer learning is a predefined approach in deep learning or machine learning used to speed up deep learning training. e pretrained models can be used to solve complex problems with preexisting knowledge. ese models provide rapid advancement in the medical field to assist medical experts in investigating and diagnosing cancerous tumors at early stages. To solve complex issues in real-time functionalities and also in the processing of image feature extraction, classification, localization, and detection, CNN with a transfer learning approach is an important image processing technique worn for analysis of medical data imaging among various deep learning methods [13].

Literature Survey
In recent years, an enormous number of approaches to brain tumor classification on MRI brain images have been proposed based on deep transfer learning models. CNN was realized as the first real-world application in 1998 to observe handwritten digits [10]. Premamayudu Bulla et al. [14] developed a hybrid model based on CNN for classifying the tumor type in the brain. Hassan Ali Khan et al. [7] proposed an automated brain tumor detection mechanism applying CNN with transfer learning models on the MRI brain image dataset. e effect of MRI image data preprocessing steps analyzed by authors [15] improves the classification accuracy in predicting brain tumor disease. Researchers [16] focused on developing a new CNN-based model to classify the three forms of tumors that existed in brain MRI images. Toktam Hatami et al. [15] investigated presenting a CNN pretrained model with image segmentation techniques. Venkatesan Rajinikanth et al. [17] suggested a VGG-16 pretrained CNN model for the classification of multigrade brain tumors. ImageNet Large-Scale Visual Recognition Challenge (ILSVRC), a visual database project, was launched by ImageNet in 2010. is challenge provides a platform for many researchers to analyze the performance of proposed methodologies developed on the given image dataset and obtain a higher classification accuracy rate. Muhammed Talo et al. [4] proposed CNN architecture AlexNet to achieve good results on various tasks based on visual recognition. e major obstacle in the development of the deep neural network in the medical field was the lack of labeled image datasets. is is overcome by applying a data augmentation approach to increase the volume of data from available labeled image datasets for a better accuracy rate. CNN-based transfer learning models accomplished excellent performance in the medical healthcare field due to the capability of self-feature learning without any human expert; weight sharing provides an adequate impenetrable network to perform automatic prediction or detection of disease through MRI brain images. Since 2014, according to the statistics analysis gathered from researchers, the number of CNN-based publications in general in the medical image appliances and specifically for Journal of Healthcare Engineering brain tumor disease has been rapidly expanded. In today's scenario, brain MRI image datasets are publicly accessible for medical scientists [18]. is assisted medical investigators in proposing further new automation classification models for prediction or detection purpose used for medical applications. H. Mohsen et al. [19] developed computer-aided diagnosis (CAD) system for automatic detection of brain tumors in brain MRI images. e proposed model utilizes sequential minimal optimization (SMO) algorithm for training Support Vector Machine (SVM) to classify glioblastoma, sarcoma or metastatic bronchogenic carcinoma malignant tumors. S. T. Kebir et al. [20] designed an automatic and a supervised MRI Brain abnormalities detection methodology. e proposed method undergoes with three parts such as construction of CNN deep learning network, segmentation using K-Means algorithm and classification of normal or abnormal brain cases.
e main purpose of this paper is to adopt a deep transfer learning approach with CNN pretrained models such as VGG-16, Inception-v3, and ResNet-50 to perform brain tumor disease detection. Based on training time and epoch number, this work presents the overall classification accuracy rate of three pretrained architectures [21]. To minimize the computation time in prediction of disease, we explore the significance of epoch numbers. Lastly, this work explains the comparative performance analysis of deep transfer learning-based pretrained models on a classification of brain tumor disease through MRI. e drawback of the previous work relies on detection of tumors from either glioma, meningioma, or pituitary brain tumor types in predicting the tumor [22]. e limitations of the existing approaches were not performed without specifying the tumor type and its grade level. e CNN with pretrained models is implemented using the Keras and modularities TensorFlow due to its high performance in detecting cancerous tumor cells.

Dataset Collection and Preprocessing.
e dataset acquired in this study is a collection of images of brain MRI scans. ere exist around 256 raw MRI images of different dimensions (width * height), usually measured in terms of pixel values. e sample MRI brain images are gathered from the Kaggle dataset. e collected images are in Joint Photographic Experts Group (JPEG) format [23]. e image database is categorized into two segments, Yes and No, based on the existence of the tumor in an MRI brain image. Benign tumors are around 158 brain MRI images and the remaining 98 brain MRI images are malignant tumors. Generally, in our work, we split our dataset into three segments required for training, testing, and validation. e sample MRI brain image dataset is shown in Figure 2.
e steps involved to preprocess Kaggle's brain image dataset [24] before applying CNN pretrained models are as follows: (1) Import the packages needed.
(2) Import the two data folders with "Yes" and "No." e MRI images available in the dataset were preprocessed in the following manner shown in Figure 3.
Intensity normalization is a major step in brain MRI analysis. During image data acquisition, a variety of scanners would be used for scanning the images at different timeperiod, which may result in large variations in pixel intensity [25,26]. erefore, overall pixel values in multiple images are normalized into the same statistical distribution required for better analysis of MRI images.
Due to MRI machinery limitations, anomalies could be found in MRI brain images. e abnormalities such as poor quality image resolution, distortion, inhomogeneity, misinterpretation, and motion heterogeneity are produced by limitations in MRI image processing. ese inaccurate analyses of scanned images could result in false positives while investigating the brain MRI image.
is poor diagnosis further affects patient treatment options.
e CNN-pretrained models require the brain MRI to be resized with a 224 × 224 × 3 dimension [11], so the dataset MRI images are reformatted to a specific dimension.
Firstly, the input MRI images are cropped to include the brain portion only from MRI brain images with open-source computer vision (CV). Many of the researchers developed various computer-aided diagnosis (CAD) models on publicly available small datasets [27]. But, we know that state-ofthe-art deep transfer learning models require a large volume of data to obtain a better optimized classification accuracy. erefore, a huge voluminous dataset is required to train the CNN model to minimize the probability of overfitting problems. In order to overcome limited datasets, this study utilized a transfer learning preprocessing technique called data augmentation. Among several techniques, namely, scaling, cropping, resizing, flipping, rotation, and perspective transformation, a specific technique is applied based on the requirement.
It is a method used for generating artificial image data for training from the original dataset [28]. Artificial data augmentation is useful to increase the performance of the proposed diagnosis model by producing new and distinct data examples required to train the model. e experimental results signify that the deep learningbased CAD model trained with artificial augmented data works better and gives accurate results to train the model as compared with the real image dataset. Figure 4 shows the augmented MRI brain image dataset [29]. e data augmentation methods, namely, affine image transformation approach, and pixel-level image transformation methods are applied to produce the new training examples to train the model.

Convolution Neural Network (CNN).
Advanced artificial intelligence models based on a deep transfer learning approach are highly efficient for medical disease prediction problems based on image classification [30]. Deep neural networks based on CNN or ConvNets are networks that include some mathematical or linear operations called "convolution." A convolution neural network (CNN) is an architecture of several levels observed like multiple hidden layers, pooling layers, and output layers or fully connected (FC) layers. ese hidden layers are named a sequence of convolution layers with filters (kernels) that performs image classification operation in predicting patients' diseases [31]. e pooling layer reduces consecutively spatial size representation and decreases hyperparameters. Expensive computation in ConvNet is reduced by the pooling layer. us, the overfitting problem in ConvNet is resolved. CNN is one of the categories of artificial neural networks that do not require handcrafted feature extraction in advance to classify the MRI brain images. Finally, the network also consists of ReLU-activated segments of 3096, 3096, and 1200, respectively. e mathematical formula for expressing ReLU is as follows: e ReLU plays a significant role in improving the neural networks by speeding up the training phase. Also, the computational steps of ReLU are simple and easy with negative values set to zero and having no exponentials, division, or multiplication operations on the training dataset. e input MRI images are resized into 256 × 256 to overcome the overfitting problem. CNN models involve a series of layers with filters to perform the task of dimension reduction and feature extraction.
erefore, CNNs are termed as "feature extractor" that can extract significant features, analyze them on their own, and classify features efficiently [32]. To generate the result, the FC layer is used with the required set of classes to perform a nonlinear transformation on the extracted features, further operating as an image classifier to analyze and classify MRI brain images. e general CNN structure is shown in Figure 5. e feature vectors are extracted from fully connected layers of CNN and further passed as an input to 1000 units of softmax layer for classification.
Mathematically, the equation for the softmax activation function is as follows: Here, "z" is the softmax function input vector consisting of "n" features of "n" target values (outcomes), "zi" is the ith item of the input vector which can take the values either positive "+" or negative "−," "e zj " is the standard exponential function implied on xi, and "Σ e z j " is a normalization term to obtain the valid probability distribution. e implementation of our proposed model is done in Python programming language and executed in Anaconda Jupiter Notebook, Google Colab. e method is trained with 120 epochs with the training, testing, and validation dataset.

Transfer Learning Approach.
Advanced deep learning models [7] based on transfer learning demands huge volumes of data for training and executing optimally state-ofthe-art high processing resources and sufficient time to train the model. Transfer learning finds a solution to another task Journal of Healthcare Engineering by utilizing knowledge gained from previously learned task [27,33]. It is an approach of utilizing pretrained CNN models based on parameter fine-tuning for computing identical tasks. In deep learning, the transfer learning approach is used with already available pretrained models trained on a huge voluminous dataset like ImageNet for image classification instead of developing the CNN detection model [34] from scratch to classify the image. e healthcare data acquisition is usually a small sample size since it is very difficult to extract. Further, it is processed with the pretrained model ImageNet, mainly used for natural images. As there are a lot of differences between natural images and MRI images, we need to fine-tune our model with transfer learning methods with CNN-based pretrained hyperparameter models [5].

Augmentation Augmentation
is approach is used when the dataset size is small for training and extensively adopted for computer vision-based problems. Transfer learning generally works on acquiring the ImageNet dataset as an input. e convolution layers freeze, and weights with a particular dataset are transferred to other CNN models and then generate the final image classification as an output [22]. erefore, in a nutshell, it is called an improved learning model for new tasks by transforming knowledge gained from existing tasks. e general working principle of transfer learning is as shown in Figure 6.

CNN-Based Deep Learning Model.
e proposed work incorporates the process of automatic detection and diagnosis model for the prediction of brain tumors in MRI images. e steps involved in the proposed approach are as illustrated in Figure 7. e first step incorporates the acquisition of data in the form of MRI brain images from the Kaggle dataset [35]. Each individual image is further gone with various preprocessing techniques, namely, resizing, cropping, and pixel-level data augmentation. Later, preprocessed images were processed with the training and testing phase through CNN pretrained models for predicting the disease in the brain.
Our proposed research study makes use of CNN with pretrained architectures to train the model to predict the disease. e CNN deep learning brain tumor-based classification is split into two divisional phases, training phase and testing phase. e performance evaluation of the model is carried out in two phases based on the splitting of the acquired dataset as training and testing phases. e MRI brain images are categorized as tumorous and nontumorous MRI images. In our study, we considered three types of brain tumors like glioma, meningioma, and pituitary brain tumors.
In this proposed model, a pretrained CNN architecture is employed for the classification that uses many labeled images for training the model obtained from large-scale datasets like the ImageNet and Kaggle.
For our research work, three pretrained CNN architectures like VGG-16, Inception-v3, and ReseNet50 are used for classification purposes. e block diagram in Figure 8 represents the methods adopted in our proposed approach. e dense layer plays a significant role in classifying the brain MRI image based on outputs from the convolution layers [29]. It feeds outputs from the previous layer to all its neurons; each individual neuron provides one output to the next layer.

CNN-Based Pretrained Models for Image Classification.
CNN-pretrained architectures applied in this work are VGG-16, Inception-v3, and ResNet50 to classify MRI brain tumor problems using a transfer learning approach.

VGG-16 CNN Model.
In our experimentation, we applied the VGG-16-pretrained CNN model. Because of the small image dataset and to avoid overfitting problems, our model is fine-tuned by freezing some of the convolution (Conv) layers. VGG-16 is a CNN model with sixteen convolution layers developed in 2014 by researchers [15]. e model accepts brain MRI images as an input with a dimension of 224 × 224 × 3. It incorporates Conv layers with kernels (filters) of fixed 3 × 3 filter size and 5 maxpooling layers of dimension 2 × 2 in size within the network [36]. It also includes extensively ReLU activation functions and 2 fully connected layers with a softmax output layer. VGG-16 model is a broad network that contains approximately 138 million hyperparameters. It stacks multiple convolutional layers to construct deep neural networks that improve the ability to learn invisible handcrafted features. e hyperparameters are essential as they control the overall behavior of the model. Its aim is to minimize the predefined loss function and to obtain better results for a specific dataset. e hyperparameters to tune are the neuron and epoch counts, softmax activation function, learning rate, and optimizer. e number of convolution layers is the second step of the hyperparameter to undergo tuning.
us, with an increase in ConvNet depth, the capacity to learn hidden features increases with a lower cost. In Figure 9, the VGG-16 ConvNet architecture is shown.

Inception-v3 CNN Model. Inception CNN network named GoogleNet was introduced and developed by Google
Brain Team in 2014, trained on ImageNet database. Inception was a pretrained network of 22 layers including 5M parameters with a kernel (filter) dimension of 1 × 1, 3 × 3, and 5 × 5 to capture handcrafted features at different scales, including the max-pooling layer. e reason to use filters with 1 × 1 kernels is to preserve computation time with less impact on network performance. Later in 2015, Google extended up the Inception network to Inception-v3 [5], in which Conv layers are rescaled to decrease with hyperparameters. Inception-v3 is a 48-layer deep neural network, which classifies MRI brain images into thousand image classes and also avoids overfitting problems. e model has the self-learning capability to extract various images' handcrafted features and analyze them for brain image classification.
e Inception-v3 network [3] requires an input image of dimension 299 × 299. Inception-v3 utilizes batch normalization, RMSprop, and image distortion for better performance in computer vision problems. Figure 10 represents the basic architecture of the Inception-v3 model.  [12] for image recognition and classification. In Residual Network, the term residual is termed as feature subtraction. Instead of learning features, we learn from those features that are subtracted and learned from the layer's input. When compared with conventional deep CNN, ResNet50 is comparatively simpler to train.

ResNet50 CNN
e pretrained model is trained on the ImageNet database [1]. ese networks also overcome degrading image classification accuracy problems. It includes skip connections in addition to enormous batch normalization. ese skip connectivities are specified as gated recurrent units or gated units. ResNet50 network establishes a direct connection between nth layer input to some (n + x)th layer, which propagates additional layers to be stacked in order to construct a deep neural network.
is Residual Network has a lower time complexity when compared with VGG16 or VGG19 models. In our experiment, we adopted a pretrained ResNet50 model and calibrated it to suit the input image dataset. e pretrained architecture of ResNet50 is shown in Figure 11.     e sensitivity and specificity are associated with classification accuracy, which uses the terms true negative (TN), true positive (TP), false negative (FN), and false positive (FP). In the field of deep learning, one of the important techniques for summarizing the performance of image classification is the confusion matrix. It is also known as an error matrix, a tabular design that allows performance visualization of the predictions and truth labels in the classification model. In predictive analysis, each row of a confusion matrix represents the instances in an actual class, and each column represents an instance in the predicted class and vice versa. e diagonal elements of the error matrix represent correct classification predictions, and nondiagonal elements represent the incorrect image classification prediction. e output of the confusion matrix is in terms of true positive (tp), true negative (tn), false negative (fn), and false positive (fp). e above terminologies are used to   Table 2.

Performance Result Analysis and Discussion.
is segment discusses the performance evaluation of pretrained CNN architectures in findings of the MRI brain tumor image classification. e accuracy, precision, recall, and F1-score along with the training time of the model are shown in Table 3. e pretrained CNN classification architectures VGG-16, Inception-v3, and ResNet50 are implemented on the brain MRI dataset containing 253 brain MRI images, of which 154 images are of the affected cancerous cells. e VGG-16 has a total of 138 million hyperparameters, all the convolution kernels are of the size 3 * 3, and maxpool kernels are of the size 2 * 2. is reduces the number of trainable datasets by 44.9% out of 100%. e reduced trainable data means a faster learning rate and overfitting problems.
e Inception-v3 consists of various Inception modules; each module that exhibits the four operations in parallel helps in depth reduction of convolution layers. ResNet50, a pretrained CNN model, helps to train the model easily with an enormous number of convolution layers without increasing the training error rate. e original image size was in various dimensions, on an average larger than 1000 × 1000 pixels. It is then resized to a dimension of 160 × 160 pixels and forwarded as an input image to a VGG-16 classifier as in Figure 12. e performance of classifiers is measured in terms of metrics like accuracy, loss, validation_loss, and validation_accuaracy. e experiments were performed on a Kaggle kernel, a cloud-based processing platform for artificial intelligence and data science projects using GPU virtual machines and RAM that empowers collective and reproductive analysis. e following represents graphs of each test predictive analysis along with 74 epochs, where the x-axis indicates an epoch and the y-axis represents accuracy and loss levels. e performance evaluation of the pretrained models is visualized in tabular form for the classification of the MRI image. Each row represents predicted classification target values, while columns represent the true classes from the training dataset.

VGG-16 CNN Architecture.
e predicted accuracy with VGG-16 along with accuracy loss, validation loss, and validation accuracy is as shown in Figures 13 and 14, respectively.
In VGG-16, training and validation accuracy represents a positive pattern and tends to rise up to 0.96% at each epoch interval. Similarly, there is a stable increase in validation loss with 0.95% at each epoch interval, and validation loss tends to decrease till the final epoch. e training accuracy leads to how the proposed model is able to classify the two images during training on the trained dataset, while validation accuracy deals with classifying the images with the validation dataset. Meanwhile, the training loss implies how well the model is fitting the training dataset, while the validation loss indicates how well the model fits the newly acquired MRI    e performance of prediction made by the VGG-16 classification model is shown in the confusion matrix as illustrated in Figure 15, where each row indicates the actual class and each column indicates the prediction class. is is a sample to identify how well the model is predicted from the testing data, classifying 17 out of 19 nontumorous images and 27 out of 31 tumorous images.

Inception-v3 CNN Architecture.
In Inception-v3, training accuracy shows that the graph tends to rise while validation accuracy is unstable. e highest increase in validation accuracy occurs after the 8th epoch reaching 0.86% from the 4th epoch of 0.78%; validation loss shows a graph with unstable results, while training accuracy tends to rise with stable results until the last epoch as in Figures 16 and 17. e confusion matrix result of the Inception-v3 model predicting from testing data and classifying 18 images out of 19 images as false positive is shown in Figure 18.

ResNet50 CNN Architecture.
In the ResNet50 test, the resulting training accuracy is increasingly higher and stable from the 2nd epoch to the last epoch, reaching accuracy above 0.95%, while the validation accuracy that is unstable tends to rise at the 5th epoch and poor performance conditions after the 8th epoch at 0.78% cannot predict the new class of data accurately illustrated in Figures 19 and 20, respectively. e performance results of the ResNet50 model are expressed in a confusion matrix predicting tumors from the testing data and classifying 11 MRI brain images out of 31 images as a false negative, which is the drawback of this model as in Figure 21.

Conclusion
is research presents a comparison of deep pretrained convolution neural networks based on transfer learning architectures, which naturally classify brain MRI images as benign and malignant tumors. e image classification process is performed using transfer learning pretrained CNN architectures, namely, VGG16, Inception-v3, and ResNet50, which acts as a classifier in prediction analysis in brain tumor detection. Data preprocessing, data augmentation methods, and hypothesized hyperparameter integral tuning are added in each testing step. Pretrained CNN architecture is said to give better performance results if, for each epoch, training accuracy and validation accuracy increase. e architecture is predicted to undergo overfitting problems, specifically when validation accuracy decreases while training accuracy increases. e model suffers an issue of overfitting when it cannot make proper predictions for a new dataset and is limited to focusing on particular training data. With the evaluation performance of each architecture, it shows that all architectures gain training accuracy of more than 0.9000%, with the highest validation accuracy reaching 0.8826%. From the performance analysis of various pretrained CNN models, namely, VGG-16, Inception-v3, and ResNet50, we conclude that we found that VGG-16 gives better accuracy on the trained and tested dataset. And also validation accuracy is much nearer to the accuracy with less loss and validation loss. e proposed study focused on only three pretrained models, namely, VGG-16, Inception-v3, and ResNet50, for image classification based on various tumor types. e research work can further be extended for classification with others like VGG-19, MobileNet, and EfficientNet-pretrained CNN-based models in the predictive analysis of brain tumors.
Data Availability e dataset used in our research work is from the Kaggle dataset, which is openly available for experimentation.

Conflicts of Interest
e authors declare that they have no conflicts of interest.