Submitted Successfully!
Thank you for your contribution! You can also upload a video entry or images related to this topic.
Check Note
2000/2000
Ver. Summary Created by Modification Content Size Created at Operation
1 -- 2020 2023-04-28 11:44:11 |
2 update references and layout -19 word(s) 2001 2023-04-28 11:49:22 |
Data Augmentation-Based Brain Tumor Detection
Edit
Upload a video

Brain tumor is a severe cancer and a life-threatening disease. Thus, early detection is crucial in the process of treatment. Recent progress in the field of deep learning has contributed enormously to the health industry medical diagnosis. Convolutional neural networks (CNNs) have been intensively used as a deep learning approach to detect brain tumors using MRI images. Due to the limited dataset, deep learning algorithms and CNNs should be improved to be more efficient.

data augmentation brain tumor deep learning
Information
Contributors MDPI registered users' name will be linked to their SciProfiles pages. To register with us, please refer to https://encyclopedia.pub/register : , , , , , ,
View Times: 62
Revisions: 2 times (View History)
Update Date: 28 Apr 2023
Table of Contents

    1. Introduction

    In 2020, it was estimated that 308,102 people were diagnosed with a primary brain or spinal cord tumor in the world [1]. Brain tumors are the 10th leading cause of death worldwide [2]. It is caused by tissue abnormality that develops within the brain or the central spine. As a result, it disrupts the proper operation of the brain. The causes of brain tumors are unknown; nevertheless, the risk can be enhanced by exposure to radiation and family history [3]. Consequently, detection and identification of brain tumors at an early phase is key to successful treatment. Indeed, it plays an indispensable role in improving treatment and ensuring a higher gain of survival possibility. There are several medical imaging techniques and diagnostic methods used to acquire information about tumors, such as Computed Tomography (CT) scans and Magnetic Resonance Imaging (MRI) scans that can distinguish between normal and abnormal cells that grow in the brain [4]. The medical science field has, in the past few years, seen striking progress leading to accurate classification of brain tumors thanks to AI and deep learning. CNN is used in image processing techniques to segment, identify, and classify MRI images as well as to classify and detect brain tumors. These image processing techniques can be based on the image content analysis described in [5][6][7], which plays a dynamic role in various computer vision applications. Recent advances in AI, and in particular in machine learning and deep learning, have contributed to the development of autonomous objects, such as robots, drones, and cars. This has allowed it to become the most important innovation driving force in the fields of technology and industry. The last few years have been marked by the growing interest in the healthcare sector and diseases detection to enhance the implementations of E-Health services. Deep Learning has recently become an active field of interest that attracts researchers, mostly in the field of medical sciences. It has significantly impacted the study of diseases in numerous ways: in the detection, prediction, and diagnosis of diseases. In [8][9], the author’s proposed solutions and new techniques to impact image reconstruction and recognition performance. Computer science scientists have developed many deep learning algorithms to detect and diagnose diseases such as cancer, lung diseases, diabetes, heart diseases, Alzheimer’s disease, hepatitis, liver disease, among others. The attentiveness to deep learning is raised to convolutional neural networks (CNN), a powerful way to learn useful representations mainly of images and other structured data. Convolutional neural networks (CNN) are deep artificial neural networks majorly used in image classification, image segmentation, and objection detection. CNN has shown significant advantages in image recognition [10][11]. Currently, it is attracting interest in a variety of domains and has achieved a huge advancement in various fields. Recently, new technologies have also taken an interest in other medical fields, such as neurosurgery. In [12][13], authors showed that Augmented Reality (AR) and mobile devices could help in the operating room. In [14], authors developed a new approach based on deep learning techniques to classify White Blood Cells for disease diagnosing. Experimental results showed that the classification of the modified images is more significant than the classification of the original ones. Authors, in [15][16], proposed to identify and classify liver diseases by using a deep supervised learning method based on CNN architecture. A classification framework was proposed in [15] and consists of improving the processing images and a segmentation of the liver lesions. In [16], the authors developed a two-step classification approach. The first step is the collection of a sufficient number of isolated training samples. The second step is to train two CNN with the same architecture but employing different optimization algorithms. The architectures described in [15][16] have reached a classification accuracy of 95%. Recently, with the COVID-19 pandemic, the world is facing a virus with unknown behavior. Therefore, several studies have been initiated to detect people attacked by this virus [17]. In [18], the author introduced a study to identify the presence or absence of malaria parasites in the blood smears of people by using a deep learning algorithm. The Convolutional Neural Network algorithm has successfully achieved an accuracy rate of 96%. As for Ghulam [19], he suggested a study based on deep learning to develop an accurate classification model to classify Breast Cancer into eight subtypes. In [20], authors stated a deep learning survey for detecting lung disease.

    2. Related Works

    In [21][22], the authors provided an overview of some potential clinical use cases using deep learning techniques by defining the steps to undertake a deep learning project in radiology. The main idea of these two papers is to discuss opportunities and challenges for incorporating deep learning in the radiology practice of the future. The effectiveness of existing applications in radiology are not yet encouraging to say that the DL can replace a radiologist in all of his diagnostic work. However, radiologists and DL can help each other to give better results. Hence, several works have been done on the classification and segmentation of the brain using MRI images. El Abbadi et al. proposed a new method using SVD as a classifier to classify brain tumors. At the first level, the algorithm had been trained with normal brain MR images. Then, at the second level, it became capable of classifying the brain images into healthy and non-healthy images. The accuracy of this method reached up to 97%. In [23], Sheikh Basheera et al. focused on brain tumor classification in MRI images using a classifier based on Convolutional Neural Networks (CNN). The main idea of the proposed approach is based on two steps. The first one is the tumor region segmentation using an ICA mixture mode model (Independent Component Analysis). The second step is the extraction of deep features. In [24], Muhammad Sajjad et al. proposed a novel convolutional neural network (CNN) based multi-grade brain tumor classification system. The first step consists of segmenting the tumor regions from an MR image using a deep learning technique. After that, they employed extensive data augmentation to train the system effectively. Finally, a pre-trained VGG-19 CNN model is fine-tuned using augmented data for brain tumor grade classification. Sunanda Das et al. [25] trained a CNN model with an image processing technique to identify various brain tumor types and achieved 94.39% accuracy with an average precision of 93.33%. In [26], Muhammed Talo et al. used deep transfer learning to classify normal and abnormal brain MR images automatically. The proposed model that used ResNet34 has achieved a 5-fold classification accuracy of 100% on 613 MR images. Ahmet Inner et al., in [27], used the ResNet50 pre-trained model, and they removed the last 5 layers of the model, then they added 8 new layers. Then, comparing its accuracy with other pre-trained models such as GoogleNet, AlexNet, and ResNet50. The modified ResNet50 model showed effective results by achieving 97.2% accuracy. He obtained a 90% accuracy in the classified images as normal and abnormal in his proposed machine learning method. The authors in [28], proposed a modified AlexNet for the detection and classification of brain tumor images and obtained 91.6% of average classification accuracy. Another approach based on a modified ResNet50 model for brain tumor detection was developed in [29]. The proposed architecture is based on the ResNet50 model with a modified layer model including five convolutional layers and three fully connected layers. In [30], researchers proposed a brain tumor detection and classification. The main idea of their approach is to use a biologically inspired orthogonal wavelet transform and deep learning techniques. Techniques of graph theory were used [31] to detect abnormalities in brains. A VGG16 architecture was the main model to classify brain images in [32].
    Limited datasets are a particularly common challenge in medical image analysis. Most computer vision tasks could use more data and data augmentation is one of the techniques often used to enhance the performance of computer vision systems. To overcome this limitation, many approaches based on deep learning have been proposed and detailed in the literature. One of the first applications of data augmentation was proposed in LeNet-5 [33] to classify the handwritten digit. In 2012, Krizhevsky et al. [34] boosted image classification by the data augmentation techniques on the ImageNet dataset. The goal of the proposed approach is to increase the dataset size. The authors used in their experiments random cropping patches from the original images, flipping them horizontally, and changing the pixel intensity. Experimental results showed that the data augmentation reduced the error rate by over 1%. After the appearance of several research works using different data augmentation techniques, researchers can categorize them into two main categories [35]. (1): Traditional transformations, which are based on the combination of the affine image transformation and color modification. (2): Generative Adversarial Networks (GANs), a tool based on an unsupervised generation of new images using min-max strategy [36]. GANs were introduced in 2014 in [37] and it consists of generating a new dataset. The new dataset is indistinguishable from the original one. In [38], authors combined data augmentation with min-max normalization to increase the contrast of tumor cells.

    3. A Taxonomy of Deep Convolutional Neural Networks

    3.1. LeNet

    The LeNet model is a classic CNN model proposed by Yann LeCun et al. [39]. It has a wide range of applications in image classification [40][41][42]. The LeNet-5 usually uses the ReLU function or the Sigmoid function as an activation function. It consists of an input layer, two convolutional layers, two pooling layers, two fully connected layers, and an output layer.

    3.2. AlexNet

    This architecture was developed by Alex Krizhevsky, Ilya Sutskever, and Geoff Hinton, and it is considered the first convolutional network to popularize it in the field of computer vision [34]. The AlexNet architecture consists of five convolutional layers (conv), three pooling layers (Pool) which are followed by three fully connected layers (FC). Compared to LeNet, this network is much bigger and deeper.

    3.3. GoogleNet

    In 2015, Google released GoogleNet, a deep neural network, which is a convolutional neural network that is 22 layers deep. Parallelization was introduced in this architecture. Indeed, it is characterized by an inception block that comprises a 1 × 1, 3 × 3, and a 5 × 5 convolution filter in addition to a 3 × 3 max-pooling layer [43].

    3.4. ResNet

    He et al. initialized ResNet models that rely on deep architectures that have demonstrated convincing precision and convergence behaviors of high quality. ResNet was conceived through numerous stacked residual units and evolved using different numbers of layers: 18, 34, 50, 101, 152, and 1202. The main disadvantage of this network is that it is very expensive to evaluate due to a large number of parameters [44].

    3.5. VGGNet

    VGGNet is an abbreviation of Visual Geometry Group; it is a convolutional neural network architecture proposed by Karen Simonyan and Andrew Zisserman of the University of Oxford in 2014 [45]. Its main contribution was to show that the depth of the network is a critical component to achieve better recognition or classification accuracy in CNNs.

    3.6. DenseNet

    In 2017, Huang et al. developed DenseNet [46]. DenseNet uses dense connections between layers via dense blocks [47][48][49][50]. DenseNet basically connects every layer to every other layer. This is extremely powerful. The entry of a layer in DenseNet is the concatenation of feature maps from previous layers. By connecting in this way, DenseNet requires fewer parameters than an equivalent traditional CNN, as there is no need to learn redundant feature maps.

    3.7. SqueezeNet

    SqueezeNet was designed as a more compact replacement for AlexNet. It is a smaller network that has almost 50 times fewer parameters than AlexNet, but it runs 3 times faster [51]. To reduce the size of the model, SqueezeNet was designed with three strategies:
    • Reduction of the filter size with the use of 1 × 1 filter instead of 3 × 3.
    • Reduction of the input channels to 3 × 3 filters.
    • Downsampling at the end of the array so that the convolutional layers have large activation maps.

    3.8. MobileNet

    MobileNet is an architecture of CNN. It is efficient for mobile and embedded vision systems [52]. Its model is designed to be used in mobile applications and it is the first mobile computer vision model based on TensorFlow. In MobileNet, the convolution is replaced by a “Depthwise Separable Convolution” which is carried out in two stages:
    • Depthwise Convolution or Convolution in depth.
    • Pointwise Convolution or Point Convolution.
    The Depthwise Convolution applies a filter to each channel, unlike conventional convolution, which applies a filter to all channels. The Pointwise Convolution consists of combining the outputs of the Depthwise Convolution. It is also called 1 × 1 convolution.

    References

    1. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA A Cancer J. Clin. 2021, 71, 209–249.
    2. Toufiq, D.M.; Ali Makki Sagheer, H. A Review on Brain Tumor Classification in MRI Images. Turk. J. Comput. Math. Educ. (TURCOMAT) 2021, 12, 1958–1969.
    3. Magadza, T.; Viriri, S. Deep Learning for Brain Tumor Segmentation: A Survey of State-of-the-Art. J. Imaging 2021, 7, 19.
    4. Chauhan, S.; More, A.; Uikey, R.; Malviya, P.; Moghe, A. Brain tumor detection and classification in MRI images using image and data mining. In Proceedings of the 2017 International Conference on Recent Innovations in Signal Processing and Embedded Systems (RISE), Bhopal, India, 27–29 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 223–231.
    5. Wang, X.; Wang, Z. The method for image retrieval based on multi-factors correlation utilizing block truncation coding. Pattern Recognit. 2014, 47, 3293–3303.
    6. Unar, S.; Wang, X.; Wang, C.; Wang, Y. A decisive content based image retrieval approach for feature fusion in visual and textual images. Knowl.-Based Syst. 2019, 179, 8–20.
    7. Wang, X.y.; Chen, Z.f.; Yun, J.j. An effective method for color image retrieval based on texture. Comput. Stand. Interfaces 2012, 34, 31–35.
    8. Wang, C.; Wang, X.; Xia, Z.; Ma, B.; Shi, Y.Q. Image description with polar harmonic Fourier moments. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 4440–4452.
    9. Wang, C.; Wang, X.; Xia, Z.; Zhang, C. Ternary radial harmonic Fourier moments based robust stereo image zero-watermarking algorithm. Inf. Sci. 2019, 470, 109–120.
    10. Bhoi, A.K.; Mallick, P.K.; Liu, C.M.; Balas, V.E. Bio-Inspired Neurocomputing; Springer: Berlin/Heidelberg, Germany, 2021.
    11. Jyotiyana, M.; Kesswani, N. A Study on Deep Learning in Neurodegenerative Diseases and Other Brain Disorders. In Rising Threats in Expert Applications and Solutions; Springer: Berlin/Heidelberg, Germany, 2021; pp. 791–799.
    12. Montemurro, N.; Condino, S.; Cattari, N.; D’Amato, R.; Ferrari, V.; Cutolo, F. Augmented Reality-Assisted Craniotomy for Parasagittal and Convexity En Plaque Meningiomas and Custom-Made Cranio-Plasty: A Preliminary Laboratory Report. Int. J. Environ. Res. Public Health 2021, 18, 9955.
    13. Condino, S.; Montemurro, N.; Cattari, N.; D’Amato, R.; Thomale, U.; Ferrari, V.; Cutolo, F. Evaluation of a wearable AR platform for guiding complex craniotomies in neurosurgery. Ann. Biomed. Eng. 2021, 49, 2590–2605.
    14. Yildirim, M.; Cinar, A.C. Classification of White Blood Cells by Deep Learning Methods for Diagnosing Disease. Rev. D’Intell. Artif. 2019, 33, 335–340.
    15. Hassan, T.M.; Elmogy, M.; Sallam, E.S. Diagnosis of focal liver diseases based on deep learning technique for ultrasound images. Arab. J. Sci. Eng. 2017, 42, 3127–3140.
    16. Arjmand, A.; Angelis, C.T.; Christou, V.; Tzallas, A.T.; Tsipouras, M.G.; Glavas, E.; Forlano, R.; Manousou, P.; Giannakeas, N. Training of deep convolutional neural networks to identify critical liver alterations in histopathology image samples. Appl. Sci. 2020, 10, 42.
    17. Tabrizchi, H.; Mosavi, A.; Szabo-Gali, A.; Felde, I.; Nadai, L. Rapid COVID-19 diagnosis using deep learning of the computerized tomography Scans. In Proceedings of the 2020 IEEE 3rd International Conference and Workshop in Óbuda on Electrical and Power Engineering (CANDO-EPE), Budapest, Hungary, 18–19 November 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 000173–000178.
    18. Sandhya, Y.; Sahoo, P.K.; Eswaran, K. Malaria Disease Detection Using Deep Learning Technique. Int. J. Adv. Sci. Technol. 2020, 29, 7736–7745.
    19. Murtaza, G.; Shuib, L.; Mujtaba, G.; Raza, G. Breast cancer multi-classification through deep neural network and hierarchical classification approach. Multimed. Tools Appl. 2020, 79, 15481–15511.
    20. Kieu, S.T.H.; Bade, A.; Hijazi, M.H.A.; Kolivand, H. A Survey of Deep Learning for Lung Disease Detection on Medical Images: State-of-the-Art, Taxonomy, Issues and Future Directions. J. Imaging 2020, 6, 131.
    21. McBee, M.P.; Awan, O.A.; Colucci, A.T.; Ghobadi, C.W.; Kadom, N.; Kansagra, A.P.; Tridandapani, S.; Auffermann, W.F. Deep learning in radiology. Acad. Radiol. 2018, 25, 1472–1480.
    22. Mazurowski, M.A.; Buda, M.; Saha, A.; Bashir, M.R. Deep learning in radiology: An overview of the concepts and a survey of the state of the art. arXiv 2018, arXiv:1802.08717.
    23. Basheera, S.; Ram, M.S.S. Classification of brain tumors using deep features extracted using CNN. In Journal of Physics: Conference Series; IOP Publishing: Secunderabad, India, 2019; Volume 1172, p. 012016.
    24. Sajjad, M.; Khan, S.; Muhammad, K.; Wu, W.; Ullah, A.; Baik, S.W. Multi-grade brain tumor classification using deep CNN with extensive data augmentation. J. Comput. Sci. 2019, 30, 174–182.
    25. Das, S.; Aranya, O.R.R.; Labiba, N.N. Brain tumor classification using convolutional neural network. In Proceedings of the 2019 1st International Conference on Advances in Science, Engineering and Robotics Technology (ICASERT), Dhaka, Bangladesh, 3–5 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–5.
    26. Talo, M.; Baloglu, U.B.; Yıldırım, Ó.; Acharya, U.R. Application of deep transfer learning for automated brain abnormality classification using MR images. Cogn. Syst. Res. 2019, 54, 176–188.
    27. Çinar, A.; Yildirim, M. Detection of tumors on brain MRI images using the hybrid convolutional neural network architecture. Med. Hypotheses 2020, 139, 109684.
    28. Khawaldeh, S.; Pervaiz, U.; Rafiq, A.; Alkhawaldeh, R.S. Noninvasive grading of glioma tumor using magnetic resonance imaging with convolutional neural networks. Appl. Sci. 2017, 8, 27.
    29. Sharma, A.K.; Nandal, A.; Dhaka, A.; Koundal, D.; Bogatinoska, D.C.; Alyami, H. Enhanced Watershed Segmentation Algorithm-Based Modified ResNet50 Model for Brain Tumor Detection. BioMed Res. Int. 2022, 2022, 7348344.
    30. Arif, M.; Ajesh, F.; Shamsudheen, S.; Geman, O.; Izdrui, D.; Vicoveanu, D. Brain Tumor Detection and Classification by MRI Using Biologically Inspired Orthogonal Wavelet Transform and Deep Learning Techniques. J. Healthc. Eng. 2022, 2022, 2693621.
    31. Mamatha, S.; Krishnappa, H.; Shalini, N. Graph Theory Based Segmentation of Magnetic Resonance Images for Brain Tumor Detection. Pattern Recognit. Image Anal. 2022, 32, 153–161.
    32. Belfin, R.; Anitha, J.; Nainan, A.; Thomas, L. An Efficient Approach for Brain Tumor Detection Using Deep Learning Techniques. In Proceedings of the International Conference on Innovative Computing and Communications, Singapore, 12–13 July 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 297–312.
    33. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324.
    34. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105.
    35. Mikołajczyk, A.; Grochowski, M. Data augmentation for improving deep learning in image classification problem. In Proceedings of the 2018 International Interdisciplinary PhD Workshop (IIPhDW), Swinoujscie, Poland, 9–12 May 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 117–122.
    36. Engstrom, L.; Tran, B.; Tsipras, D.; Schmidt, L.; Madry, A. A Rotation and a Translation Suffice: Fooling CNNs with Simple Transformations. 2018. Available online: https://openreview.net/forum?id=BJfvknCqFQ (accessed on 20 February 2022).
    37. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27.
    38. Nayak, D.R.; Padhy, N.; Mallick, P.K.; Zymbler, M.; Kumar, S. Brain Tumor Classification Using Dense Efficient-Net. Axioms 2022, 11, 34.
    39. LeCun, Y.; Jackel, L.; Bottou, L.; Brunot, A.; Cortes, C.; Denker, J.; Drucker, H.; Guyon, I.; Muller, U.; Sackinger, E.; et al. Comparison of learning algorithms for handwritten digit recognition. In Proceedings of the International Conference on Artificial Neural Networks, Perth, Australia, 27 November–1 December 1995; Volume 60, pp. 53–60.
    40. Wang, G.; Gong, J. Facial expression recognition based on improved LeNet-5 CNN. In Proceedings of the 2019 Chinese Control And Decision Conference (CCDC), Nanchang, China, 3–5 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 5655–5660.
    41. Zhang, Z.H.; Yang, Z.; Sun, Y.; Wu, Y.F.; Xing, Y.D. Lenet-5 Convolution Neural Network with Mish Activation Function and Fixed Memory Step Gradient Descent Method. In Proceedings of the 2019 16th International Computer Conference on Wavelet Active Media Technology and Information Processing, Chengdu, China, 13–15 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 196–199.
    42. Rongshi, D.; Yongming, T. Accelerator implementation of Lenet-5 convolution neural network based on FPGA with HLS. In Proceedings of the 2019 3rd International Conference on Circuits, System and Simulation (ICCSS), Nanjing, China, 20–22 July 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 64–67.
    43. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9.
    44. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778.
    45. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556.
    46. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708.
    47. Shah, U.; Harpale, A. A Review of Deep Learning Models for Computer Vision. In Proceedings of the 2018 IEEE Punecon, Pune, India 2018 IEEE, Piscataway, NJ, USA, 30 November–2 December 2018; pp. 1–6.
    48. Peters, J.F. Foundations of Computer Vision: Computational Geometry, Visual Image Structures and Object Shape Detection; Springer: Berlin/Heidelberg, Germany, 2017; Volume 124.
    49. Li, Y.H.; Aslam, M.S.; Yang, K.L.; Kao, C.A.; Teng, S.Y. Classification of body constitution based on TCM philosophy and deep learning. Symmetry 2020, 12, 803.
    50. Chen, Q.; Xie, Q.; Yuan, Q.; Huang, H.; Li, Y. Research on a real-time monitoring method for the wear state of a tool based on a convolutional bidirectional LSTM model. Symmetry 2019, 11, 1233.
    51. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360.
    52. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861.
    More
    Information
    Contributors MDPI registered users' name will be linked to their SciProfiles pages. To register with us, please refer to https://encyclopedia.pub/register : , , , , , ,
    View Times: 62
    Revisions: 2 times (View History)
    Update Date: 28 Apr 2023
    Table of Contents
      1000/1000

      Confirm

      Are you sure to Delete?

      Video Upload Options

      Do you have a full video?
      Cite
      If you have any further questions, please contact Encyclopedia Editorial Office.
      Alsaif, H.; Guesmi, R.; Alshammari, B.M.; Hamrouni, T.; Guesmi, T.; Alzamil, A.; Belguesmi, L. Data Augmentation-Based Brain Tumor Detection. Encyclopedia. Available online: https://encyclopedia.pub/entry/43622 (accessed on 07 June 2023).
      Alsaif H, Guesmi R, Alshammari BM, Hamrouni T, Guesmi T, Alzamil A, et al. Data Augmentation-Based Brain Tumor Detection. Encyclopedia. Available at: https://encyclopedia.pub/entry/43622. Accessed June 07, 2023.
      Alsaif, Haitham, Ramzi Guesmi, Badr M. Alshammari, Tarek Hamrouni, Tawfik Guesmi, Ahmed Alzamil, Lamia Belguesmi. "Data Augmentation-Based Brain Tumor Detection" Encyclopedia, https://encyclopedia.pub/entry/43622 (accessed June 07, 2023).
      Alsaif, H., Guesmi, R., Alshammari, B.M., Hamrouni, T., Guesmi, T., Alzamil, A., & Belguesmi, L. (2023, April 28). Data Augmentation-Based Brain Tumor Detection. In Encyclopedia. https://encyclopedia.pub/entry/43622
      Alsaif, Haitham, et al. "Data Augmentation-Based Brain Tumor Detection." Encyclopedia. Web. 28 April, 2023.
      Top