Brain Tumor Detection Using Convolutional Neural Network Supervised by
Brain Tumor Detection Using Convolutional Neural Network Supervised by
net/publication/334947653
CITATIONS READS
0 912
5 authors, including:
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Md Abdullah Al Nasim on 03 August 2019.
A thesis
Submitted in partial fulfillment of the requirements for the Degree of
Bachelor of Science in Computer Science and Engineering
Submitted by
Supervised by
Mr. Faisal Muhammad Shah
Assistant Professor
Department of Computer Science and Engineering
Ahsanullah University of Science and Technology
Dhaka, Bangladesh
June 2019
CANDIDATES’ DECLARATION
We, hereby, declare that the thesis presented in this report is the outcome of the investi-
gation performed by us under the supervision of Mr. Faisal Muhammad Shah, Assistant
Professor, Department of Computer Science and Engineering, Ahsanullah University of Sci-
ence and Technology, Dhaka, Bangladesh. The work was spread over two final year courses,
CSE4100: Project and Thesis-I and CSE4250: Project and Thesis-II, in accordance with the
course curriculum of the Department for the Bachelor of Science in Computer Science and
Engineering program.
It is also declared that neither this thesis nor any part thereof has been submitted anywhere
else for the award of any degree, diploma or other qualifications.
Mohsena Ashraf
150104012
Tonmoy Hossain
150104032
MD Abdullah Al Nasim
150104085
i
CERTIFICATION
This thesis titled, “Brain Tumor Detection using Convolutional Neural Network”, submit-
ted by the group as mentioned below has been accepted as satisfactory in partial fulfillment
of the requirements for the degree B.Sc. in Computer Science and Engineering in June 2019.
Group Members:
ii
ACKNOWLEDGEMENT
We would like to express our deep gratitude to our honourable thesis supervisor Mr. Faisal
Muhammad Shah, Assistant Professor, Department of Computer Science and Engineering,
Ahsanullah University of Science and Technology, for his patient guidance, enthusiastic en-
couragement and constructive critiques of this thesis work. We would also like to thank
Professor Dr. Kazi A Kalpoma, the honourable Head of the Department of Computer Sci-
ence and Engineering of Ahsanullah University of Science and Technology, and all our re-
spected teachers, lab assistants, and friends for their moral support and helpful discussions.
Finally, we would like to thank our parents for their constant support and encouragement
throughout our study.
MD Abdullah Al Nasim
iii
ABSTRACT
Brain Tumor segmentation is one of the most crucial and arduous tasks in the field
of medical image processing as a human-assisted manual classification can result in in-
accurate prediction and diagnosis. Moreover, it becomes a tedious task when there is a
large amount of data present to be processed manually. Brain tumors have diversified
appearance and there is a similarity between tumor and normal tissues and thus the
extraction of tumor regions from images becomes complicated. In this thesis work, we
developed a model to extract brain tumor from 2D Magnetic Resonance brain Images
(MRI) by Fuzzy C-Means clustering algorithm which was followed by both traditional
classifiers and deep learning methods. The experimental study was carried on a real-
time dataset with diverse tumor sizes, locations, shapes, and different image intensities.
In traditional classifier part, we applied six traditional classifiers namely- Support Vector
Machine (SVM), K-Nearest Neighbor (KNN), Multi-layer Perceptron (MLP), Logistic Re-
gression, Naive Bayes and Random Forest. Among these classifiers, SVM provided the
best result. Afterwards, we moved on to Convolutional Neural Network (CNN) which
shows an improvement in performance over the traditional classifiers. We compared
the result of the traditional classifiers with the result of CNN. Furthermore, the perfor-
mance evaluation was done by changing the split ratio of CNN and traditional classifiers
multiple times. We also compared our result with the existing research works in terms
of segmentation and detection and achieved better results than many state-of-the-art
methods. For the traditional classifier part, we achieved an accuracy of 92.42% which
was obtained by Support Vector Machine (SVM) and CNN gave an accuracy of 97.87%.
iv
Contents
CANDIDATES’ DECLARATION i
CERTIFICATION ii
ACKNOWLEDGEMENT iii
ABSTRACT iv
List of Figures ix
List of Tables xi
1 Introduction 1
1.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Brain Tumor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.3 Classification of Brain Tumor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.3.1 Benign Tumor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.3.2 Malignant Tumor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.4 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.5 Objective . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.6 Thesis Contribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.7 Thesis Structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.8 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
v
3.3.2 Region Based Segmentation . . . . . . . . . . . . . . . . . . . . . . . . . 18
3.3.2.1 Region Growing . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
3.3.2.2 Clustering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
3.3.3 Edge Based Segmentation . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
3.3.4 Thresholding Based Segmentation . . . . . . . . . . . . . . . . . . . . . 21
3.3.5 Dataset wise Distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
3.3.6 A Statistical Survey . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
3.4 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
4 Background Study 31
4.1 Medical Image . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
4.2 Importance of Medical Image Analysis . . . . . . . . . . . . . . . . . . . . . . . 31
4.3 Machine Learning in Medical Image Analysis . . . . . . . . . . . . . . . . . . . 32
4.4 Image Processing for Brain Tumor Segmentation . . . . . . . . . . . . . . . . 32
4.4.1 Skull Stripping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
4.4.2 Pre-processing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
4.5 Traditional Machine Learning Classifiers . . . . . . . . . . . . . . . . . . . . . . 37
4.5.1 K-Nearest Neighbor . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
4.5.2 Logistic Regression . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
4.5.3 Multi-layer Perceptron . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
4.5.4 Naive Bayes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
4.5.5 Random Forest . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
4.5.6 Support Vector Machine (SVM) . . . . . . . . . . . . . . . . . . . . . . . 39
4.6 Deep Learning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
4.6.1 Neuron . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
4.6.2 Artificial Neural Network . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
4.6.3 Using Neural Networks for Images . . . . . . . . . . . . . . . . . . . . . 41
4.6.4 Types of Neural Networks . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
4.6.5 Convolutional Neural Network (CNN) . . . . . . . . . . . . . . . . . . . 43
4.7 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56
5 Proposed Methodology 57
5.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
5.2 Our Working Approach for Brain Tumor Segmentation . . . . . . . . . . . . . 57
5.2.1 Stage-1: Brain Tumor Detection using Traditional Classifiers . . . . 57
5.2.1.1 Step-1: Skull Stripping . . . . . . . . . . . . . . . . . . . . . . . 58
5.2.1.2 Step-2: Filtering and Enhancement . . . . . . . . . . . . . . . . 60
5.2.1.3 Step-3: Segmentation by Fuzzy C-means (FCM) Algorithm . 60
5.2.1.4 Step-4: Morphological Operations . . . . . . . . . . . . . . . . 60
5.2.1.5 Step-5: Tumor extraction & contouring . . . . . . . . . . . . . 61
vi
5.2.1.6 Step-6: Classification by traditional classifiers . . . . . . . . . 61
5.2.2 Classification Using Convolutional Neural Network . . . . . . . . . . 62
5.2.2.1 Convolutional Layer . . . . . . . . . . . . . . . . . . . . . . . . . 64
5.2.2.2 Max Pooling Layer . . . . . . . . . . . . . . . . . . . . . . . . . . 64
5.2.2.3 Flatten Layer . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
5.2.2.4 Fully Connected Layer . . . . . . . . . . . . . . . . . . . . . . . . 65
5.3 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
vii
6.8 Performance comparison Between the existing model the proposed CNN
model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
6.9 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
References 89
viii
List of Figures
ix
4.17 Curve of ReLU Function [91] . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
4.18 Curve of leaky ReLU Function [91] . . . . . . . . . . . . . . . . . . . . . . . . . 53
4.19 Network before and after Dropout [92] . . . . . . . . . . . . . . . . . . . . . . . 54
4.20 A Neural Network before and after Batch Normalization [92] . . . . . . . . . 55
x
List of Tables
xi
1
Chapter 1
Introduction
1.1 Overview
Medical imaging techniques are used to image the inner portions of a human body for med-
ical diagnosis. And medical image classification is one of the most challenging & affluent
topics in the field of Image Processing. Medical image classification problems, tumor de-
tection or detection of Cancer is the most prominent one. The statistics about the death
rate from brain tumor suggest that it is one of the most alarming and critical cancer types
in the Human body. As per the International Agency of Research on Cancer (IARC), more
than 1,000,000 people are diagnosed with brain tumor per year around the world, with
ever increasing fatality rate. It is the second most fatal cause of death related to Cancer in
children and adults younger than 34 years [1]. In recent times, the physicians are follow-
ing the advanced methods to identify the tumor which is more painful for the patients. To
analyze the abnormalities in different parts of the body, CT (Computed Tomography) scan
and MRI (Medical Reasoning Imaging) are two convenient methods. MRI-based medical
image analysis for brain tumor studies has been gaining attention in recent times due to
an increased need for efficient and objective evaluation of large amounts of medical data.
Analysis of this diverse range of image types requires sophisticated computerized quantifi-
cation and visualization tools. So, automatic brain tumor detection from MRI images will
play a crucial role in this case by alleviating the need of manual processing of huge amount
of data.
According to Ilhan et al. [2], a brain tumor occurs when abnormal cells form within the
brain. Many different types of brain tumors exist. Some brain tumors are noncancerous (be-
1.3. CLASSIFICATION OF BRAIN TUMOR 2
nign), whereas some brain tumors are cancerous (malignant) and some are pre-malignant.
Cancerous tumors can be divided into primary tumors that start within the brain, and sec-
ondary tumors that have spread from somewhere else, known as brain metastasis tumors
[2].
There are two types of brain tumor. One is Benign Tumor characterized as non-cancerous
and the other one is Malignant Tumor- also known as Cancerous Tumor.
Benign brain tumors are usually defined as a group of similar cells that do not follow normal
cell division and growth, thus developing into a mass of cells that microscopically do not
have the characteristic appearance of a cancer [3]. These are the properties of a benign
tumor:
• Grows slowly, do not invade surrounding tissues or spread to other organs, and often
have a border or edge that can be seen on CT scans.
• It can be life threatening because they can compress brain tissues and other structures
inside the skull, so the term ‘benign’ can be misleading.
Malignant brain tumors contain cancer cells and often do not have clear borders. They are
considered to be life threatening because they grow rapidly and invade surrounding brain
tissues [4]. These are the properties of a malignant tumor:
• Fast growing cancer that spreads to other areas of the brain and spine.
• Generally these are more serious and often more fatal threat to life.
1.4. MOTIVATION 3
Figure 1.1: Benign Tumor (left) and Malignant Tumor (Right) [5]
1.4 Motivation
Observing the recent statistics of death rate caused by brain tumors, we selected brain tumor
detection and classification which belongs to the field of medical image analysis. Tumor
detection in medical images are time consuming as it depends on human judgment. The
experts in this field, such as radiologists, specialized doctors examine CT scan, MRI, PET
scan images and give decisions upon which the treatment depends. This whole process is
time consuming. Automated medical image analysis can help to reduce the time and effort
taken here and the workload of a human as it will be done by machines.
Figure 1.2: New cases and survival rate caused by brain tumor [6]
Figure 1.2 shows that death caused by brain cancer is higher than other types of cancers.
Brain tumor detection in an early stage can help to reduce the death rate in this field. For
supporting faster communication, where patient care can be extended to remote areas using
information technology, automated image analysis will help to a great extent.
1.5. OBJECTIVE 4
The developed countries of the world have been introduced to the automation of medical
image analysis. But, in Bangladesh, it has not been adopted well yet. We want to build a
model which will be efficient and feasible in the perspective of Bangladesh.
If we go through the most recent decade’s statistics, it had been shown that there was an
estimated 14.1 million cancer cases around the world in 2012. Men comprised of 7.4 million
among them, while rest 6.7 million were consisted of female. This number is expected to
increase to 24 million by 2035. Among all the form of cancers, Lungs cancer was the most
common cancer worldwide contributing 13% of the total number of new cases diagnosed
in 2012 [7]. So analyzing these statistics, we wanted to contribute in the field of medical
image analysis.
1.5 Objective
The main objective of our thesis is to build a model that can predict whether the medical
images contain a tumor or not and find its properties. Primarily, dataset collection is the
main task to work on a medical image because brain tumor dataset is scarce as well as very
much complicated to acquire. Most of the researchers focused on definitive work like filter-
ing, segmentation, feature selection or skull removing. Here we tried to establish a model
which can accomplish all the fundamental and major necessary tasks to find a tumor and its
properties. We proposed an efficient and effective method which helps in the segmentation
and detection of the brain tumor without any human assistance, based on both traditional
classifiers and Convolutional Neural Network. Finally, we compared all the experimental
results to find out which model provides better performance in terms of accuracy, sensitivity
and other performance metrics.
• For segmentation of the tumor, we developed a skull stripping method using basic
image processing techniques which produces a much better result than the existing
skull removal techniques which is discussed in the experimental results section.
1.7. THESIS STRUCTURE 5
• We carried out two types of classifications to detect the tumor. Tumor classification us-
ing Traditional Machine Learning classifiers and Convolutional Neural Network were
carried out and comparison of performance measures was done between these two
models.
Our thesis book consists of seven chapters based on our research work. Following, we will
briefly discuss about the basis of the chapters.
• Chapter 1: Introduction This is the introductory chapter of the thesis book. In this
chapter, we will describe shortly the objective, our goal, and contribution of our thesis.
We will also characterize Brain Tumor and its classification shortly.
• Chapter 2: Related Works We will describe some associated work that had been done
before and describe their working approach, advantages and disadvantages.
• Chapter 4: Background Study This chapter includes all the prerequisite knowledge
related to our thesis topic. We will discuss the Image processing techniques, Tradi-
tional Machine Learning Classifiers, and Deep Learning.
• Chapter 6: Experimental Results & Evaluation In this chapter, we will explain the
performance measures, our proposed algorithm for performance evaluation and we
will discuss about the experimental results.
1.8. SUMMARY 6
• Chapter 7: Conclusion Finally, we will end our work with the conclusion chapter. In
this division, we will discuss the limitations of our work and future directions to work
on and improve.
1.8 Summary
In this chapter, a short description of Brain tumor and its sub-fields are described. We have
discussed different types of Brain tumor and its characteristics. We shortly explained the
motivation behind our work and the objective for doing it. At last, our thesis contribution
and the structure of our thesis was briefly described.
7
Chapter 2
2.1 Overview
In recent years, numerous and diverse types of work have been carried out in the field of
medical image processing. Researchers from the various ground such as- computer vision,
image processing, machine learning came into a place in the field of Medical Image Pro-
cessing. We have studied some of the existing papers to find the most useful and advanced
methods that were used in the existing articles in recent times. We worked on a total of 52
research articles. In this chapter, we will discuss thoroughly about these papers and their
working procedures which are related to our work.
Shehzad et al. [8] proposed an algorithm to detect brain tumor from the MR images and
calculate the area of the tumor. The designed algorithm claims to detect and extract the
tumor of any shape, intensity, size, and location.
Working Approach:
• Gaussian low pass filter is used to blur the image and then the blurred image is added
to the original image.
• Morphological gradient image and filtered image are added for image enhancement.
• The threshold value is calculated with the help of standard deviation and mean of the
filtered image.
• Image is binarized by comparing the threshold value with every pixel value of the
image.
• Erosion is done again for thinning the image and dilation is done again to get removed
(caused by erosion) part of the tumor back.
• Tumor is extracted by comparing the original image with the dilated image.
• Erosion is done to shrink any noise remaining in the resultant tumor extracted image.
Sankari et al. [9] along with the other researchers came up with a model for cancer diagnosis
for a brain tumor which is the toughest task. Most researches has been done in this field
using PCA, Route set theory and Wavelet method. The authors here used Convolutional
neural networks to solve the problem.
Working Approach:
• The bilateral filter is used to remove the noise from the MRI.
• Histogram Equalization is used for enhancing and feature extraction of the image.
Borase et al. [10] used computer-based procedures to detect tumor blocks and classify the
type of tumor by using Artificial Neural Network. They used MRI images for their training
and testing stage.
Working Approach:
• After segmentation, every set of connected pixels having the same gray-level values
are assigned the same unique region label.
• K-means clustering technique could be a better option for pre-processing of the images.
• Erosion is done again for thinning the image and dilation is done again to get removed
(caused by erosion) part of the tumor back.
• The tumor is extracted comparing the original image with the dilated image.
• Erosion is done to shrink any noise remained in the resultant tumor extracted image.
Working Approach:
• The image obtained from the opening operation is subtracted from the original im-
age, thus separating the skull from the original image. The image obtained from the
opening operation is subtracted from the skull image which removes the undesired
gray pixels from the skull image. Lastly, the cleaned skull image is subtracted from
the original image to get the resulting image which is excluding the skull.
• For segmentation, they proposed a threshold approach in which the threshold value
was calculated by the sum of unique pixel values excluding zeros (black pixels) divided
by the count of unique pixel values.
Joseph et al. [12] used K-means as well as SVM method for segmentation and to maintain
the pattern for future uses respectively. They defined a relation between SVM and skull
masking technique. They combine the K-means segmentation and SVM technique with skull
2.2. REVIEWS OF THE RELATED PAPERS 10
masking for getting a better result. Maintaining the tumor pattern is also a computationally
complex task which they did and also justified with the projected idea. They modified the
feature extraction technique and existing K-means technique in a way that the amount of
skull tissue is obscured and a proper tumor-detecting image is diversified. Hence we could
do more by distinguishing the type of tumor and the location as well as the stage which is
not yet identified properly.
Working Approach:
• Support Vector Machine (SVM) is used in an unsupervised manner which will use to
create and maintain the pattern for future use.
In hunnur et al.[13], the authors proposed a method to detect brain tumors mainly based
on Thresholding approach and morphological operations. It also calculates the area of the
tumor and displays the stage of the tumor.
Working Approach:
• MRI images of the brain are used as input and the images are converted into grayscale
images.
• High pass filter is used to remove the noise present in the converted images.
• Thresholding is used to extract the object from the background by selecting a threshold
value.
• Tumor region is detected and then the image is shrunk to remove the unwanted details
present in the images.
• The tumor area is calculated and at last, the stage of the tumor patient is displayed.
2.2. REVIEWS OF THE RELATED PAPERS 11
Working Approach:
• Obtained MRI images are displayed in two-dimensional matrices having pixels as its
elements.
• Gaussian low pass filter and averaging filters are used to remove salt and pepper noise
from the image.
• Gaussian high pass filter is used to enhance boundaries of the objects in the image.
• Threshold segmentation is used to convert the grayscale image into a binary image
format.
• Morphological operators are applied to the converted binary image to separate the
tumor part of the image.
Furthermore, we extended our study into some more recent articles and a thorough descrip-
tion is given by-
Devkota et al. [15] established the whole segmentation process based on Mathematical
Morphological Operations and spatial FCM algorithm which improves the computation time,
but the proposed solution has not been tested up to the evaluation stage. It detects cancer
with 92% accuracy and classifier has an accuracy of 86.6%. Yantao et al. [16] resembled
Histogram based segmentation technique. The brain tumor segmentation task as a three-
class (tumor including necrosis and tumor, edema and normal tissue) classification problem
regarding two modalities FLAIR and T1. The abnormal regions were detected by using a
region-based active contour model on FLAIR modality. The edema and tumor tissues were
distinguished in the abnormal regions based on the contrast enhancement T1 modality by
the k-means method and accomplished a Dice coefficient and sensitivity of 73.6% and 90.3%
respectively.
2.2. REVIEWS OF THE RELATED PAPERS 12
Based on edge detection approaches, Badran et al. [17] adopted the canny edge detection
model accumulated with Adaptive thresholding to extract the ROI. The dataset contained
102 images. Images were first preprocessed, then for the two sets of the neural network, for
the first set canny edge detection was applied, And for the second set, adaptive thresholding
was applied. The segmented image is then represented by a level number and characteristics
features are extracted by the Harris method. Then two neural network is applied, first for
the detection of healthy or tumor containing the brain and the second one is for detecting
tumor type. Depicting the outcomes and comparing these two models, the Canny edge
detection method showed better results in terms of accuracy. Pei et al. [18] proposed a
technique which utilizes tumor growth patterns as novel features to improve texture based
tumor segmentation in longitudinal MRI. Label maps are being used to obtain tumor growth
modeling and predict cell density after extracting textures such as fractal, mBm etc. and
intensity features. Performance of the model reflected as the Mean DSC with tumor cell
density- LOO: 0.819302 and 3-Folder: 0.82122.
Dina et al. [19] introduced a model based on the Probabilistic Neural Network model re-
lated to Learning Vector Quantization. The model was evaluated on 64 MRI images, among
which 18 MRI images were used as the test set, and the rest was used as a training set.
The Gaussian filter smoothed the images. 79% of the processing time was reduced by the
modified PNN method. A Probabilistic Neural Network based segmentation technique was
implemented by Othman et al. [20]. Principal Component Analysis (PCA) was used for
feature extraction and also to reduce the large dimensionality of the data [20]. The MRI
images are converted into matrices, and then Probabilistic Neural Network is used for clas-
sification. Finally, performance analysis is done. The training dataset contained 20 subjects,
and the test dataset included 15 subjects. Based on the spread value, accuracy ranged from
73% to 100%.
Concentrating on Region based Fuzzy Clustering and deformable model, Rajendran et al.
[21] accomplished 95.3% and 82.1% of ASM and Jaccard Index based on Enhanced Proba-
bilistic Fuzzy C-Means model with some morphological operations. Zahra et al. [22] applied
LinkNet network for tumor segmentation. Initially, they used a single Linknet network and
sent all training seven datasets to that network for segmentation. They did not consider the
view angle of the images and introduced a method for CNN to automatically segment the
most common types of a brain tumor which do not require pre-processing steps. Dice score
of 0.73 is achieved for a single network, and 0.79 is obtained for multiple systems.
2.3. SUMMARY 13
2.3 Summary
In this chapter, we reviewed a total number of 52 research articles and discussed about their
working procedures. This literature study shows that a lot of research works had been intro-
duced regarding brain tumor segmentation and detection. Some of the researchers applied
traditional classifiers, while the others implemented deep learning methods. Some works
achieved a significant result using traditional approaches while some works did not. But af-
ter studying these works, we can claim that deep learning performs better than traditional
classifiers because of their learning mechanism and use of memory in the network.
14
Chapter 3
3.1 Overview
From the inception of Image processing, Medical Image is undoubtedly one of the most de-
cisive sectors and brain tumor detection is one of the most crucial task in this field. Many
researchers have worked in this field. We have studied 52 research articles which have been
done for brain tumor detection and segmentation. In this chapter, we will present a statis-
tical analysis of these research articles based on years, citations, segmentation techniques
etc.
Working on a total of 52 research articles dated from 2007 to 2018 and based on the col-
lected information, we break down the total process of segmentation along with respective
figures. We try to select the articles based on various criteria such as- citation, year, dataset,
etc. but mostly focused on segmentation techniques. Apart from the single and mixed seg-
mentation technique, we further go through the articles which adopted the Neural Network.
3.2. YEAR AND CITATION WISE DISTRIBUTION 15
In figure 3.1 and 3.2, statistics of the articles according to years and based on segmentation
types is being represented with all the accessible information.
From figure 3.1, we can see that from year 2010 to 2013 and from year 2015 to 2017, max-
imum number of research works have been done. And from figure 3.2, we can depict that
most of the researchers adopted Neural network based detection (supervised) and clustering
based segmentation (unsupervised).
Further, in figure 3.3, categorizing the articles based on the number of citation is described,
where a total of 42 papers where the segmentation technique belongs to the primitive image
processing techniques.
In figure 3.4, the papers appertain to Neural Network based segmentation and the statistics
represent the information about these papers citation.
3.2. YEAR AND CITATION WISE DISTRIBUTION 16
From figure 3.3 and 3.4, we can see that the research articles which was done by Rajesh et
al. [43] and Shenbagarajan et al. [65] has the highest number of citations.
Figure 3.3: Citation-wise distribution of basic image processing techniques used in the arti-
cles
Segmentation subdivides an image into its constituent regions or objects based on some
concurrent characteristics where the objects which are depicted are strongly related to the
regions. The level of details to which the subdivision is carried depends on the problem
being solved, that is, segmentation should stop when the objects or regions of interest in
an application have been detected i.e. for the distinction of the tumor interests lied on
separating the abnormal tissues.
Distribution of Skull Stripping techniques based on uses is represented through Figure 3.5.
Summarizing all the information from figure 3.5, we can infer that skull stripping technique
is gradually getting undone according to years.
Articulating an ROI and extracting decisive features, Layer Based segmentation techniques
are the most indispensable one. In this segmentation process, three layers are generated
from an image, entitled as mask, graphics and text layer. JBIG (Joint Bi-level Image Experts
Group) algorithm is used to losslessly compress the mask layer, text layer is compressed
using token based order [24], and graphics layer are compressed using the JPEG coder.
A layered model is used for object detection and image segmentation that composites the
result of a bank of object detectors defining shape masks and explaining the appearance,
depth order, and that evaluates both class and instance segmentation [25].
3.3. SEGMENTATION TECHNIQUE WISE DISTRIBUTION 18
Deng et al. [26] proposed an adaptive region growing method based on the two preeminent
subjects which cover variances and gradients along and inside of the boundary curve in order
to overcome the difficulty of manual threshold selection. There are significant region-based
segmentation methods which we can use to accomplish our goal, which are given below-
Region growing is the simplest and straightforward region based segmentation that groups’
pixels or sub-regions into larger regions based on some pre-defined criteria. The common
procedure is to differentiate one pixel with its neighbors [27]. There are two types of region
growing method-
1. Seeded Region Growing Method: Along with the image, seeds are being taken as
input and marking each of the objects that are to be segmented. The regions are itera-
tively grown through comparison of all unallocated neighboring pixels to the regions
[28].
2. Unseeded Region Growing Method: It does not require the seed point and begins
with a single region. At each iteration, it works through considering the adjacent
pixels in the same way as seeded region growing.
3.3.2.2 Clustering
2. Fuzzy C-Means Clustering (FCM): One piece of data belongs to two or more clusters.
Developed by Dunn in 1973 [31] and improved by Bezdek in 1981 [32], is frequently
used in Image Segmentation. Use of FCM algorithm on intensity features where FCM
segments image into a pre-specified number of clusters (K), FCM gives a fuzzy mem-
bership (U) to describe the degree of similarity of one pixel to each cluster [33].
3. Spatial Fuzzy C-Means Clustering: This algorithm utilizes the local spatial infor-
mation which is convenient in reducing noise distortion and intensity inhomogeneity
in the segmentation [34]. Segmentation was done by spatial FCM. An extension of
Probabilistic Neural Network (PNN), called WPNN, which uses anisotropic Gaussians
rather than the isotropic Gaussians used by PNN was used for classification [35].
Figure-3.6 shows the depiction of the above information and figure-3.7 shows the in-
formation about the total number of papers which used clustering according to years.
From figure 3.6, we see that K-means Clustering and Fuzzy C Means Clustering were
mostly used. Furthermore, from figure 3.7, we can infer that, segmentation based on
clustering techniques was mostly carried out in the year 2017 and clustering based
segmentation techniques are still in practice.
Figure 3.6: Number wise distribution of clustering techniques used in the articles
3.3. SEGMENTATION TECHNIQUE WISE DISTRIBUTION 20
Figure 3.7: Year wise distribution of clustering techniques used according to years
Edge based segmentation means segmentation of an image by identifying the edges of the
Region of Interests. Edges can be connected and disconnected. According to the data or
methodology, one needs region boundaries which are closed and the desired edges are the
boundaries between such objects or spatial-taxon [36].
1. Edge Detection: Working with their developed thresholding and edge detection tech-
nique, Debnath et al. [37] identified the tumor successfully but the computation time,
as well as the classifier and efficiency, is below average in contrast with the others’
work.
2. Canny edge detection: By optimizing the canny edge detection model and GA for
canny edge detection. Malathi et al. [38] improved the efficiency used closed contour
segmentation. Badran et al. [17] used two sets of a neural network in their work. For
the first set, they used canny edge detection.
Figure 3.8 shows the information about the number of papers which used edge detection
techniques in their work. And from the figure, we see that most of the papers carried out
watershed segmentation technique.
3.3. SEGMENTATION TECHNIQUE WISE DISTRIBUTION 21
Figure 3.8: Number wise distribution of edge detection techniques used in articles
1. Binary Thresholding: Debnath et al. [37] presented their algorithm which includes
thresholding for tumor segmentation. Converting 24-Bit Color Images to 256 Gray
Color Images and Calculating histograms the resulting images were converted to a
binary thresholded image, histograms were calculated and at last edge detection al-
gorithm was used [11].
3. Otsu Thresholding: This algorithm presumes that the image encompasses two classes
of pixels following bi-modal histogram [41]. Mittal et al. [42] used Otsu Threshold-
ing segmentation along with watershed technique, asserting that Otsu’s thresholding
chooses the threshold for minimizing the intra-class variance of the thresholded black
and white pixels.
Figure 3.9 represents the number of papers where thresholding based approach was used for
brain tumor segmentation and detection, where most of the papers adopted binary thresh-
olding approach.
3.3. SEGMENTATION TECHNIQUE WISE DISTRIBUTION 22
Figure 3.9: Type wise distribution of thresholding techniques used in the articles
Several competition and resources from universities are open for all to work on the images
of the brain tumor. For working on the segmentation of brain tumors in MRI scans, BRATS
dataset is available for all which contains T1, T2 and FLAIR images.
Several sources of databases were used in the papers. Figure 3.10 represents the names of
the data-sets used by the articles and distribution of the dataset based on their usage. And
from this figure, it can be deduced that most of the researchers did not use any standard
dataset.
3.3. SEGMENTATION TECHNIQUE WISE DISTRIBUTION 23
We worked on a total of 52 papers, in which 10 of them used neural network based seg-
mentation and the rest of them worked on various traditional segmentation techniques.
Gleaning all the essential and statistical information, we compose a table which reflects the
bottom line of the articles about their performance and relative work.
No of
Authors &
Technique Images Citation Result Total
Year
used
Mean, Median, Std. Dev.
Edge (Debnath et And number of white
12 36 1
Detection al., 2011)[37] pixels measured
to detect the tumor
(Malathi,
Canny Edge Unspecified 0 Dice: 90.13% - 93.26%
2014)[38] 2
Detection
3.3. SEGMENTATION TECHNIQUE WISE DISTRIBUTION 26
No of
Authors &
Technique Images Citation Result Total
Year
used
Parametric deformable
Average similarity
active contour
(Rajendran et metrics: 95.3%,
model model 15 46 1
al., 2011)[21] average Jaccard
with gradient
index: 82.1%
vector field(GVF)
Gives substantial
results for
Content Based
(Jainy et al., homogeneous
Active Contour 428 96 1
2012)[67] tumors against
Model
different and
similar background
Region Based (Shenbagarajan, et
80 13 Accuracy: 94% 1
Active Contour al., 2016) [65]
3.3. SEGMENTATION TECHNIQUE WISE DISTRIBUTION 29
3.4 Summary
In this chapter, we presented a statistical analysis of various brain tumor segmentation tech-
niques such as- Layer based, Region based, Thresholding based, Edge detection based seg-
mentation tecniques which was presented with necessary tables, figures and graphical rep-
resentations along with their descriptions. Furthermore, we constituted a numerical rep-
resentation of the research papers along with their performance measures where neural
network based segmentation has been carried out.
31
Chapter 4
Background Study
Medical imaging means the visualization of body parts, tissues, or organs, for use in clinical
diagnosis, treatment and disease monitoring. Imaging techniques encompass the fields of
radiology, nuclear medicine and optical imaging and image-guided intervention.
Digital image processing is one of the modern and advance technology which process on
the photos or videos. Nowadays x-ray is the main important application of Digital Image
Processing. Before x-ray, it was very difficult to examine human bone of his body because
the doctor has to cut the body skin and flesh to know about the bone of the body whether
it is crack or damaged or not. In every perspective and every lane, image processing can be
applied whether it is for security or for personal use.
From the discovery of X-ray by Roentgen in 1895 to the present day imaging tools like
Magnetic Resonance Imaging (MRI) and Computed Tomography (CT), the technology has
progressed much. The advances in imaging technology will continue as time progresses.
However, today the focus of systems is shifting from medical imaging focus from the gener-
ation and acquisition of images to post-processing and management of image data. This is
stimulated by the need to make efficient use of the data that already exists. Recent progress
in imaging research has shown the potential the technology can have to improve and trans-
form many aspects of clinical medicine.
4.3. MACHINE LEARNING IN MEDICAL IMAGE ANALYSIS 32
Advances in both imaging and computers have synergistically led to a rapid rise in the
potential use of artificial intelligence in various radiological imaging tasks, such as risk as-
sessment, detection, diagnosis, prognosis, and therapy response, as well as in rapid disease
discovery. Computer-extracted (radio-mic) features can serve as input to machine learning
algorithms (i.e., computer algorithms that “learn” a specific task given specific input data).
With such machine learning methods, multiple radio-mic features are merged into a single
value, such as a tumor signature, which might be related to the likelihood of disease state
[74]. Various machine learning techniques have been applied across the decades, for ex-
ample, linear discriminant analysis, support vector machines, decision trees, and random
forests, and neural networks. Reviews of machine learning have been written over the past
many years including those that serve as tutorials to new investigators into the field [75]
[76].
One of the most challenging as well as demanding task is to segment the region of interest
from an object and segmenting the tumor from an MRI Brain image is an ambitious one.
Brain tumor segmentation from MRI is one of the most crucial tasks in medical image pro-
cessing as it generally involves a considerable amount of data. Moreover, the tumors can be
ill-defined with soft tissue boundaries. So it is a very extensive task to obtain the accurate
segmentation of tumors from the human brain.
Image processing helps to improve the quality of the MRI images and subsequently to per-
form features extraction and classification. For brain tumor segmentation, image processing
involves various steps such as skull stripping, pre-processing, tumor contouring etc.
Skull removing is one of the most important and essential steps for detecting the brain tumor
and extracting the features. It is the process of eliminating all the non-brain tissues in the
MRI images. By this method, it is possible to remove additional cerebral tissues such as fat,
skin and skull in the brain images [77]. Some of the popular skull removing techniques are-
image contour, skull stripping based on segmentation, morphological operation and based
on histogram analysis and using threshold value.
4.4. IMAGE PROCESSING FOR BRAIN TUMOR SEGMENTATION 33
4.4.2 Pre-processing
In medical image processing, medical images are corrupted by different type of noises. It
is very important to obtain precise images to facilitate accurate observations for the given
application [78]. Pre-processing steps include filtering, morphological operations etc.
• Filtering: Image filtering is useful for many applications, including smoothing, sharp-
ening, removing noise, and edge detection. A filter is defined by a kernel, which is a
small array applied to each pixel and its neighbors within an image. The process used
to apply filters to an image is known as convolution, and may be applied in either the
spatial or frequency domain. Spatial domain filtering can be classified into two types-
smoothing and sharpening filters, according to their outputs.
– Smoothening Filter: Smoothening spatial filters are used for noise reductions
and blurring operations. It includes box filter, Gaussian blur filter, median blur
filter, bilateral filter etc.
1. Box Filtering is basically an average-of-surrounding-pixel kind of image fil-
tering.
2. Gaussian blur filtering is a 2D convolution operator that is used to ‘blur’
images and remove detail and noise. In this sense it is similar to the mean
filter, but it uses a different kernel that represents the shape of a Gaussian
(‘bell-shaped’) hump. The degree of smoothing is determined by the stan-
dard deviation of the Gaussian.
3. Median blur filter is normally used to reduce noise in an image. Instead of
simply replacing the pixel value with the mean of neighboring pixel values,
it replaces it with the median of those values.
4. Bilateral filter is a non-linear, edge-preserving, and noise-reducing smooth-
ing filter for images. This preserves sharp edges.
– Sharpening Filter: Sharpening spatial filters seek to highlight fine details. It
removes blurring from images and highlights edges. Sharpening filters are based
on spatial differentiation. It includes laplacian filter, sobel filter, difference filter
etc.
1. Laplacian filters are derivative filters used to find areas of rapid change
(edges) in images.
2. Sobel filters are typically used for edge detection.
3. Difference filters enhance the details in the direction specific to the mask
selected.
4.4. IMAGE PROCESSING FOR BRAIN TUMOR SEGMENTATION 34
• Morphological Operations: Morphology means pixel shape based analysis. The ob-
jective of using morphological operations is to remove the imperfections in the struc-
ture of the image [79]. The most basic morphological operations are dilation and
erosion. Dilation adds pixels to the boundaries of objects in an image, while erosion
removes pixels on object boundaries. Morphological operations also include opening,
closing, hit and miss transform etc.
Morphological erosion removes islands and small objects so that only substantive ob-
jects remain.
• Segmentation: Segmentation means dividing the image into different regions and
separating objects from background. Accurate segmentation of objects of interest in
an image greatly facilitates further analysis of these objects. There are various types
of segmentation algorithms such as edge based, thresholding based, region based,
clustering based etc.
Figure 4.1 represents the segmentation of an object from its background.
which define the resulting groups. Examining the centroid feature weights can
be used to qualitatively interpret what kind of group each cluster represents.
Figure 4.3 is the visual representation of the segmentation which was done by
K-means clustering technique.
We have used six traditional machine learning classifiers which are K-Nearest Neighbor, Lo-
gistic Regression, Multi-layer Perceptron, Naive Bayes, Random Forest, and Support Vector
Machine to get the accuracy of the tumor detection of our proposed model.
KNN is one of the simplest of classification algorithms available for supervised learning.
The idea is to search for the closest match of the test data in feature space. Typically the
object is classified based on the labels of its k nearest neighbors by majority vote. If k=1,
the object is simply classified as the class of the object nearest to it. When there are only
two classes, k must be an odd integer. However, there can still be times when k is an odd
integer when performing multi-class classification. After we convert each image to a vector
of fixed-length with real numbers, we used the most common distance function for KNN
which is Euclidean distance between two points x and y and it can be written as:
m
X 1
Æ
d(x, y) = |x − y| = (x − y).(x − y) = ( (x i − yi )2 ) 2 (4.3)
i=1
Like all regression analyses, the logistic regression is a predictive analysis. Logistic regres-
sion is used to describe data and to explain the relationship between one dependent binary
variable and one or more nominal, ordinal, interval or ratio-level independent variables.
Logistic regression uses an equation as the representation, very much like linear regression.
Input values (x) are combined linearly using weights or coefficient values (referred to as
the Greek capital letter Beta) to predict an output value (y). A key difference from linear
regression is that the output value being modeled is a binary values (0 or 1) rather than a
numeric value. Below is the logistic regression equation [93]:
(e b0 +b1 x)
y= (4.4)
1 + e(b0 +b1 x)
4.5. TRADITIONAL MACHINE LEARNING CLASSIFIERS 38
where y is the predicted output, b0 is the bias or intercept term and b1 is the coefficient for
the single input value (x).
A multi layer perceptron (MLP) is a feed forward artificial neural network that generates
a set of outputs from a set of inputs. A MLP consists of at least three layers of nodes: an
input layer, a hidden layer and an output layer [94]. Except for the input nodes, each node
is a neuron that uses a nonlinear activation function. An MLP is characterized by several
layers of input nodes connected as a directed graph between the input and output layers.
MLP uses back-propagation for training the network. Multi-layer are most of the neural
networks expect deep learning. It uses one or two hidden layers.
In a machine learning classification problem, there are multiple features and classes, say,
C1 , C2 , ..., Ck . The main aim in the Naive Bayes algorithm is to calculate the conditional
probability of an object with a feature vector x 1 , x 2 , ..., x n belongs to a particular class Ci .
And the equation can be written as [95]:
j=n
Y P(Ci )
P(Ci |x 1 , x 2 , ..., x n ) = ( P(x j |Ci )) ∗ , f or 1 ¶ j ¶ n (4.6)
j=1
P(x 1 , x 2 , ..., x n )
Random forest is a type of supervised machine learning algorithm based on ensemble learn-
ing. Ensemble learning is a type of learning where we join different types of algorithms or
same algorithm multiple times to form a more powerful prediction model. The random
forest algorithm combines multiple algorithm of the same type i.e. multiple decision trees.
4.6. DEEP LEARNING 39
Random decision forests correct for decision trees’ habit of over-fitting to their training set
[96]. It operates by constructing a multitude of decision trees at training time and out-
putting the class that is the mode of the classes (classification) or mean prediction (regres-
sion) of the individual trees [96].
Support vector machines are supervised learning models with associated learning algo-
rithms that analyze data used for classification and regression analysis.
Given a set of training examples, each marked as belonging to one or the other of two
categories, an SVM training algorithm builds a model that assigns new examples to one
category or the other, making it a non-probabilistic binary linear classifier [97]. An SVM
model is a representation of the examples as points in space, mapped so that the examples of
the separate categories are divided by a clear gap that is as wide as possible. New examples
are then mapped into that same space and predicted to belong to a category based on which
side of the gap they fall.
Deep Learning is a subset of machine learning algorithms that is very good at recognizing
patterns but typically requires a large number of data. Deep learning excels in recognizing
objects in images as it is implemented using three or more layers of artificial neural networks
where each layer is responsible for extracting one or more features of the image.
4.6.1 Neuron
The basic building unit of neural networks are artificial neurons, which imitate human brain
neurons. These artificial neurons are powerful computational units that have weighted input
signals and produce an output signal using an activation function. These neurons are spread
across the several layers in a neural network.
• Weight: When a signal (value) arrives, a neuron gets multiplied by a weight value.
If a neuron has three inputs, it has three weight values which can be adjusted during
training time [87].
• Bias: It is an extra input to neurons and it is always 1, and has its own connection
weight. This makes sure that even when all the inputs are none (all 0’s) there is going
to be an activation in the neuron.
Deep learning consists of artificial neural networks that are modelled on similar networks
present in the human brain. These neural networks work in multiple layers so this kind of
machine learning is called deep learning.
Artificial neural networks are a way of calculating an output from an input (a classifica-
tion) using weighted connections (“synapses") that are calculated from repeated iterations
through training data. Each pass through the training data alters the weights such that the
neural network produces the output with greater “accuracy" (lower error rate). The com-
bination of working memory and speed is crucial when we’re doing hundreds of thousands
of matrix multiplications. Figure 4.6 represents a simple artificial neural network:
An artificial neural network generally has three layers. Layers are made up of some nodes
which are interconnected. The three layers of ANN are described in the following:
• Input Layer: This layer consists of neurons and they just receive the inputs and pass
4.6. DEEP LEARNING 41
it to the next layer. The number of layers in the input layer should be equal to the
attributes or features in the dataset.
• Hidden Layer: In between input and output layer there are hidden layers based on
the type of model. Hidden layers contain vast number of neurons. The neurons in the
hidden layer apply transformations to the inputs before passing them to the next layer.
As the network is trained the weights get updated, to be more predictive. The actual
processing of the data is done via a system of weighted connections in the hidden
layer. The hidden layers are linked to the output layer.
• Output Layer: The output layer is the predicted feature, or class in a classification
problem, it basically depends in the type of the built model. The output layer gives
the output based on the information passed from the hidden layer.
The internal structure of ANN can be changed by itself based on the information passing
through it. This is done by the adjustment of the weights. Every connection in the neural
network generally has a weight that controls the signal between the two neurons. If the
output is good, the adjustment is no longer needed, but if the output is poor, then the
system adapts by changing the weights to improve the output. The performance evaluation
of the output is done by the system by comparing the output with the original output given
before in the training mode [6].
Neural network can be used to recognize or detect object category but it will require more
works to uniquely identify an object. A classical neural network requires to input a set of
features extracted from each of the image. Deep neural network (DNN) works with image
pixels.
4.6. DEEP LEARNING 42
An image can be represented as a matrix, each element of the matrix containing color in-
formation for a pixel. The matrix is used as input data into the neural network. The small
dimensions of the images, to easily and quickly help learning, establish the size of the vec-
tor and the number of input vectors. The transfer function used is also called an activation
function. Processing of images with artificial neural network involves different processes,
such as:
Processing images with artificial neural networks successfully resolve the problems of clas-
sification, identification, authentication, diagnostics, optimization and approximation [87].
There are several kinds of artificial neural networks. These type of networks are imple-
mented based on the mathematical operations and a set of parameters required to determine
the output. Some of the popular neural networks are [86]:
• Feedforward Neural Network: This neural network is one of the simplest forms of
ANN, where the data or the input travels in one direction. The data passes through
the input nodes and exit on the output nodes. This neural network may or may not
have the hidden layers. In simple words, it has a front propagated wave and no back
propagation by using a classifying activation function usually.
• Radial Basis Function Neural Network: Radial basis functions consider the distance
of a point with respect to the center. RBF functions have two layers, first where the
features are combined with the Radial Basis Function in the inner layer and then the
4.6. DEEP LEARNING 43
output of these features are taken into consideration while computing the same output
in the next time-step which is basically a memory.
• Recurrent Neural Network: The Recurrent Neural Network works on the principle
of saving the output of a layer and feeding this back to the input to help in predicting
the outcome of the layer. This is also known as Long Short Term Memory.
Among these mostly used neural networks, Convolutional Neural Networks are applied in
techniques like signal processing and image classification techniques.
The basic idea of Convolutional Neural Network was introduced by Kunihiko Fukushima in
1980s [88]. Convolutional Neural Networks (ConvNets or CNNs) are a category of Neural
Networks that have proven very effective in areas such as image recognition and classifica-
tion. Computer vision techniques are dominated by convolutional neural networks because
of their accuracy in image classification. CNN is a class of deep, feed-forward artificial neu-
ral networks (where connections between nodes do not form a cycle) & use a variation of
multi-layer perceptrons designed to require minimal pre-processing.
ConvNet architectures make the explicit assumption that the inputs are images, which allows
us to encode certain properties into the architecture. These then make the forward function
more efficient to implement and vastly reduce the amount of parameters in the network.
ConvNets are made up of neurons that have learnable weights and biases. Each neuron
receives some inputs, performs a dot product and optionally follows it with a non-linearity.
4.6. DEEP LEARNING 44
The whole network still expresses a single differentiable score function: from the raw image
pixels on one end to class scores at the other. And they have a loss function on the last layer.
(i) Why CNN is different than simple Neural Network: Convolutional Neural Networks
have a different architecture than regular Neural Networks [86]. Regular Neural Networks
transform an input by putting it through a series of hidden layers. Every layer is made up
of a set of neurons, where each layer is fully connected to all neurons in the layer before
and where neurons in a single layer function completely independently and do not share
any connections between themselves. Finally, there is a last fully-connected layer, which is
the output layer that represent the predictions. Regular Neural Networks do not scale well
to full images.
Convolutional Neural Networks are a bit different. First of all, the layers are organized
in three dimensions: width, height and depth. Further, the neurons in one layer do not
connect to all the neurons in the next layer but only to a small region of it. Lastly, the final
output will be reduced to a single vector of probability scores, organized along the depth
dimension.
Figure 4.7: A simple neural network and A Convolutional Neural Network [86]
In the above, in figure 4.7, the left side represents a regular three Layer neural network. On
the other hand, the right side of the figure represents a CNN which arranges its neurons in
three dimensions (width, height, depth) [86].
Every layer of a CNN transforms the 3D input volume to a 3D output volume of neuron
activation’s. In this example, the red input layer holds the image, so its width and height
would be the dimensions of the image, and the depth would be three (Red, Green, Blue
channels).
one is reversed and shifted. This operation is a particular kind of integral transform [89]:
Z ∞
There are three elements that enter into the convolution operation [89]:
• Feature map: The feature map is also known as an activation map. It is called feature
map because it is also a mapping of where a certain kind of feature is found in the
image.
Figure 4.8 represents the convolution operation.
(iii) The Convolution Arithmetic: Here the facts are shown how properties of an output
image changes from input by some factors [90].
Let, Input Image Size = I, (means width I and Height I i.e. image size is I * I)
Filter Size = F, (Filter is F * F)
Number of Filters = K
Number of Strides = S
Amount of Zero Padding = P
Then the output image size will be:
I − F + 2P
O= +1 (4.8)
S
4.6. DEEP LEARNING 46
• Forward Pass for Convolutional Layer: In forward pass of convolutional layer, a filter
performs dot multiplication with all the parts of the input matrix one after another of
filter size, then sum all the elements of the single dot product and adds the bias value
with it and places the final value in the corresponding location of the output matrix
for that dot product. This process continues across full input image with all filters and
for each filter there comes an output.
The iteration of forward pass and back-propagation will continue until the network finds
the desired output.
(iv) Layer’s Used to a Build CNN Model: A simple CNN is a sequence of layers, and every
layer of a CNN transforms one volume of activation’s to another through a differentiable
function. Three main types of layers used to build CNN architectures are:
• Convolutional Layer: The Convolutional layer is the core block of the Convolutional
Neural Network. It has some special properties. It does most of the computational
heavy lifting.The CONV layer’s parameters consist of a set of learn-able filters. Every
filter is small spatially (along width and height), but extends through the full depth
of the input volume.
For example, a typical 3X3 filter on a first layer of a ConvNet might have size 5*5*3
(i.e. 5 pixels width and height, and 3 because images have depth 3, the color chan-
nels). During the forward pass, each filter is convolved across the width and height of
the input volume and dot products are computed between the entries of the filter and
the input at any position. As the filters are slided over the width and height of the in-
put volume, a 2-dimensional activation map will be produced that gives the responses
of that filter at every spatial position. Intuitively, the network will learn filters that ac-
tivate when they see some type of visual feature such as an edge of some orientation.
These activation maps are stacked along the depth dimension and the output volume
is produced.
Figure 4.9 shows an example of convolution in the convolutional layer of CNN. More-
over, Convolutional layer has some basic features [90], such as:
4.6. DEEP LEARNING 47
These features help to reduce the number of parameters in the whole system and
makes the computation more efficient.
Three hyper-parameters control the size of the output volume of the convolutional
layer. These parameters are:
– Depth: The depth of input volume in first layer is the number of color channels
of that input image. If the input image is a color image then the depth is 3 that
is the red channel, the green channel and the blue channel. If the image is black
and white or gray-scale, then the depth is 1.
The depth of the output volume is the number of filters that we use in the input.
Figure 4.8 represents the depth which changes from 3 to 32 when we use 32
filters and the image size gets smaller.
4.6. DEEP LEARNING 48
– Stride: Stride is used to slide along width and height of the input image. When
the stride is 1, then we move the filters one pixel at a time. When the stride is 2,
then the filters jump 2 pixels at a time as we slide them around.
Figure 4.11: Sliding filter along an input image when the stride is 1 [90]
Figure 4.11 shows that a 2*2 filter moves along the input size of 4*4 through
width and height when the stride is 1.
– Zero Padding: Sometimes in the input layer, we pad the input image with zero
that is called zero-padding. Zero padding allows us to control the size of the
input layer. If we don’t use zero-padding, sometimes some property from the
edges can be lost.
Figure 4.12 illustrates the scenerio of zero-padding of an input.
• Pooling Layer: Pooling layer is another building block of CNN. Usually pooling layer
is placed after the convolutional layer. Its function is to progressively reduce the spa-
tial size of the representation to reduce the amount of parameters and computation
in the network, and hence to also control over fitting. The Pooling Layer operates
independently on every depth slice of the input and resizes it spatially. Pooling does
not affect the depth dimension of the input volume.
4.6. DEEP LEARNING 49
The task of pooling is done by is done by summarizing the sub-regions of the input
using some methods like taking the average or the maximum or the minimum value
of the sub-regions only. These methods are called pooling functions.
The equation [90] shows that the output size from the pooling layer will be,
w − f
O= +1 (4.9)
s
4.6. DEEP LEARNING 50
Figure 4.13 shows the connection between two layers where the right one is the fully
connected layer.
(iv) Activation Function: Activation functions are used to introduce non-linearity to neural
networks. It squashes the values in a smaller range. For example, a sigmoid activation
function squashes values between a ranges 0 to 1.
Figure 4.14 demonstrates the importance of activation function. The figure on the left side is
4.6. DEEP LEARNING 51
the scenario of applying logistic regression without the application of activation function and
the figure on the right side is the same scenario with the application of activation function.
• Commonly Used Activation Functions: There are many activation functions used in
deep learning industry. Here we will discuss about some activation functions in brief
which are commonly in use.
– Sigmoid: The Sigmoid function bounds the input value in between 0 to 1 range.
For large positive number, it returns 1 and for large negative number, it returns
0. The mathematical representation of the sigmoid function is [91]:
1
σ(x) = (4.10)
1 + e−x
– ReLU (Rectified Linear Unit): ReLU refers to Rectified Linear Unit. It simply
thresholds the input value to zero. For positive, it returns the number and for
negative, it returns 0. In AlexNet architecture, after using ReLU as an activation
function, it was 6 times faster than using the tanh function [91]. The formula of
ReLu is as following [91]:
f (x) = max(0, x) (4.11)
Moreover, there are some other activation functions which are used in the field of deep
learning.
4.6. DEEP LEARNING 54
If a simple model is not performing well at predicting due to poor generalization and a
complex model may not perform well due to over-fitting. In this case, regularization helps
to choose the preferred complexity for the model.
The two most important regularization techniques are Dropout and Batch Normalization.
The above figure- 4.19 shows how dropout performs in the neural network. Cross
neuron represents inactive neurons in this network.
4.6. DEEP LEARNING 55
Figure 4.20: A Neural Network before and after Batch Normalization [92]
The above figure- 4.20 shows a standard neural network before and after having Batch
Normalization.
• Bias: Bias is an error from erroneous assumptions in the learning algorithm. High
bias can cause an algorithm to miss the relevant relations between features and target
outputs (under-fitting).
• Learning rate: The learning rate or step size in machine learning is a hyper-parameter
which determines to what extent newly acquired information overrides old informa-
tion.
4.7. SUMMARY 56
• Beta_1 and Beta_2: The algorithm calculates an exponential moving average of the
gradient and the squared gradient, and the parameters beta1 and beta2 control the
decay rates of these moving averages.
• Epsilon: It is a very small number to prevent any division by zero in the implementa-
tion.
• Dropout or Decay weight: A simple and powerful regularization technique for neural
networks and deep learning models is dropout.
• Amsgrad: AMSGrad (in Keras, it is controlled by setting amsgrad = True for Adam
optimizer), which uses the maximum of past squared gradients in order to allow the
rarely-occurring minibatches with large and informative gradients to have greater in-
fluence on the overall direction, otherwise diminished by exponential averaging in
plain Adam.
• Epoch: Epoch is defined as the number forward and backward pass of all training
examples.
(vii) Limitations of CNN: CNNs only capture local ‘spatial’ patterns in data. If the data
can’t be made to look like an image, ConvNets are less useful. Moreover, CNNs are very bad
at encoding different representations of pose and orientation within themselves. In CNNs,
the orientations and their surroundings are not taken into account.
4.7 Summary
This chapter includes a brief explanation of all the topics which are related to our thesis
work along with their working procedure, their advantages and disadvantages etc. Firstly,
we have tried to describe the topics of basic image processing and then we moved on to the
description of the basic CNN architecture along with its different hyper-parameters.
57
Chapter 5
Proposed Methodology
5.1 Overview
There are two proposed model by which we can detect the abnormal cells in brain MRI. We
have tried to detect the tumor using traditional machine learning algorithms and also using
a convolutional neural network. In classification using traditional machine learning step,
we have tried to train the proposed model using six machine learning algorithms: KNN,
Logistic Regression, Multi-layer perceptron, Naive Bayes, Random Forest, and SVM. At first,
we will discuss the proposed segmentation technique to distinguish the tumor. After that,
we will discuss the detection of the tumor using traditional machine learning algorithm.
After that, we will introduce the proposed model and thoroughly describe the all the layers
related to detect the brain tumor using CNN.
We have conducted our work for brain tumor segmentation and detection in two stages.
The stages are:
In our first prospective model, brain tumor segmentation and detection using machine learn-
ing algorithm have been done, and a comparison of the classifiers for our model is illustrated.
5.2. OUR WORKING APPROACH FOR BRAIN TUMOR SEGMENTATION 58
Our proposed system of Brain tumor detection using traditional machine learning algorithm
consists of seven stages:
The results of our work accomplished satisfactory results. The main stages of our proposed
model (Fig-5.1) will be illustrated in the following sections.
We will comprehensively describe the proposed model of tumor segmentation and for clas-
sification using traditional Machine learning classifier, we used six classifiers and evaluated
the performance.
In the later section, we will comprehensively describe the sections of the proposed method-
ology (fig-5.1)
Skull stripping is a very important step in medical image processing as the background of
the MRI image not containing any useful information, and it only increases the processing
5.2. OUR WORKING APPROACH FOR BRAIN TUMOR SEGMENTATION 59
time.
In our work, we come up with a hybrid model to remove the skull portion from the MRI
images in three steps. These three steps can be elaborately explained in fig. 5.2.
• For skull removal, at first we used Otsu’s Thresholding method which automatically
calculates the threshold value and segments the image into background and fore-
ground.
• After binarization of the MRI, erosion operation had been performed before applying
connected component analysis.
• At the last stage of our skull stripping step, we used connected component analysis to
extract only the brain region and as a consequence, we removed the skull from the
MRI.
• We found the largest component which is the skull and then we found the mask by
assigning 1 to inside and 0 to outside of the brain region.
• Multiplying the mask to T1, T2 and FLAIR images we got the skull removed MRI.
5.2. OUR WORKING APPROACH FOR BRAIN TUMOR SEGMENTATION 60
For better segmentation, we need to maximize the MRI image quality with minimized noise
as brain MRI images are more sensitive to noise than any other medical image.
Gaussian blur filter was used in our work for Gaussian noise reduction existing in Brain MRI
which prevailed the performance of the segmentation. Because we just need the abnormal
tissues of the Brain MRI, we can avoid the subtle detail. In a MRI, there may be very small
area of abnormal region or tumor tissues, but we need the maximal one. Also in MRI image
the amount of Gaussian noise can be comparatively higher than any other noises. That is
why Gaussian filter was used. Then we enhanced the image by using add-Weighted method
(it provides a blending effect to the images). First, we blur the image where we know that by
using smoothing filter to an image, we can suppress most of the high-frequency components.
Then, we subtract this smoothed image from the original image (the resulting difference is
known as a mask). Thus, the output image will have most of the high-frequency components
that are blocked by the smoothing filter. Adding this mask back to the original will enhance
the high-frequency components.
Fuzzy C-Means clustering algorithm is used for segmentation, which allows one piece of
data to belong to two or more clusters. We got the fuzzy clustered segmented image at this
stage, which ensured a better segmentation. Unlike k-means where data point must exclu-
sively belong to one cluster center here data point is assigned membership to each cluster
center as a result of which data point may belong to more than one cluster center.
We also employed the Normalized cut algorithm, Thresholding technique, K-Means Clus-
tering, and watershed segmentation. But considering the pros and cons described in back-
ground studies, we adopt FCM algorithm for segmentation.
To segment the tumor, we only need the brain part rather than the skull part. For this, we
applied morphological operations in our images.
At first, erosion was done to separate weakly connected regions of the MRI image. After
erosion, we will get multiple disconnected regions in our images. Dilation is applied after-
ward. In case of noise removal, erosion is followed by dilation using the same structuring
5.2. OUR WORKING APPROACH FOR BRAIN TUMOR SEGMENTATION 61
We will find the edge of the abnormal tissues by which we can mark the perimeter. The
maximum length between edges to another edge will be the perimeter of the Tumor tissues.
Tumor cluster extraction was done by an intensity-based approach which is thresholding.
The output of this image is the highlighted tumor area with a dark background.
After tumor Contouring, we can find the segmented tumor of the brain MRI. Further, we
will classify about the tumor and find some imperative features of the tumor which are
co-related with the brain MRI.
We used six traditional machine learning classifiers which are K-Nearest Neighbor, Logis-
tic Regression, Multi-layer Perceptron, Naive Bayes, Random Forest, and Support Vector
Machine to get the accuracy of tumor detection of our proposed model. Among these six
classifiers, SVM gave the best result. In this section, we will describe why SVM gives the
most improved result.
A main advantage of the KNN algorithm is that it performs well with multi-modal classes
because the basis of its decision is based on a small neighborhood of similar objects [100].
Therefore, even if the target class is multi-modal, the algorithm can still lead to good accu-
racy. However, a major disadvantage of the KNN algorithm is that it uses all the features
equally in computing for similarities which can lead to classification errors, especially when
there is only a small subset of features that are useful for classification [100].
Logistic regression is a better traditional learning algorithm because the output of logistic
regression is more informative than other classification algorithms. But Logistic regression
tends to underperform when there are multiple or non-linear decision boundaries and for
MRI images it will not work well. They are not flexible enough to naturally capture more
complex relationships. Also this classifier needs long training time sometimes.
Naive Bayes classifier does not work well with the brain MRI because there is a great chance
of losing the accuracy. In brain MRI, dependencies between the abnormal and normal tissue
play considerable importance.
Random forests algorithm has less variance than a single decision tree. It means that it works
correctly for a large range of data items than single decision trees. But the BRATS dataset
is not a large one to employ the random forest algorithm on it. The main disadvantage of
5.2. OUR WORKING APPROACH FOR BRAIN TUMOR SEGMENTATION 62
Random forests is their complexity. They are much harder and time-consuming to construct
than decision trees. It also requires more computational resources and is also less intuitive.
Support Vector Machine (SVM) gives more improve result in terms of classifying an image.
SVM works well with even unstructured and semi-structured data like text, images, and
trees. With an appropriate kernel function, we can detect the tumor. We consider all the
pixels of the MRI as a feature. Unlike in neural networks, SVM is not solved for local optima.
It scales relatively well to high dimensional data.
We can say that for classifying tumor MRI, the outcomes of classification of an image using
SVM gives better result with respect to other classifiers . That is why, We used Support Vector
Machine algorithm for tumor detection in our proposed model. We got the best result using
SVM over the other traditional classifiers.
Convolutional Neural Network is broadly used in the field of Medical image processing.
Over the years lots of researchers tried to build a model which can detect the tumor more
efficiently. It is a class of deep neural networks which is applied to interpreting visual im-
agery. A fully-connected neural network can detect the tumor, but because of parameter
sharing and sparsity of connection, we adopted the Convolutional Neural Network (CNN)
for our model.
A Five-Layer Convolutional Neural Network is introduced and implemented for tumor detec-
tion. The aggregated model consisting of seven stages including the hidden layers provides
us with the better performing result for the apprehension of the tumor. One convolutional
layer, One Pooling layer, One Flatten layer, and Two Fully connected layer work as the hid-
den layers along with the input and output layer.
We characterized the complete working flow into seven steps from which we can detect the
brain tumor using CNN. In fig-5.3 the seven-step working flow diagram is shown.
Fig-5.4 is the proposed methodology for tumor detection using 5-Layer Convolutional Neu-
ral Network. In fig-5.4, five hidden layers are shown along with its dimension. We will
discuss about the layer and its characteristics in the following section.
5.2. OUR WORKING APPROACH FOR BRAIN TUMOR SEGMENTATION 63
Figure 5.3: Working flow of the proposed five layer CNN Model.
Figure 5.4: Proposed Methodology for Brain tumor detection using 5-Layer Convolutional
Neural Network
5.2. OUR WORKING APPROACH FOR BRAIN TUMOR SEGMENTATION 64
A Convolutional layer is the core building block of a CNN model. Using convolutional layer
as the opening layer, an input shape of the MRI images is generated which is 64*64*3,
converting all the images into a same dimension. After accumulating all the images in the
same aspect, we created a convolutional kernel that is convoluted with the input layer -
administering with 32 convolutional filters of size 3*3 each with the support of 3 channels
tensors. Rectified Linear Unit (ReLU) is used as an activation function.
The input volume has size 64*64*3 which means 64 pixels width, 64 pixels height, depth of
3 and the filter size is 3*3. Then each neuron in the Convolutional Layer will have weights
to a 3*3*3 region in the input volume, for a total of 3*3*3 = 27 weights and one will be
added for bias parameter.
There are three hyper parameters which we will evaluate and those are- depth, stride and
zero-padding. For our model, the input volume size is 64*64*3, filter size is 3*3 so the
spatial extent or filter size is 3. We did not use padding in the border so that Zero Padding,
and stride is 1. Using stride as 1 grant us to do all the spatial down-sampling in the pooling
layers, with the CONV layers only transforming the input volume depth-wise. Now using
the convolutional layer equation - (4.9) we find the dimension of convolutional layer which
is 62, 62 and 32 respectively. After Convolutional layer, we introduced the max pooling
layer as the second layer.
The main focus of pooling layer is to progressively reduce the spatial size of the represen-
tation in order to reduce the number of parameters and computational task in the network.
It can control over-fitting because it can scale down the parameters. Using the max pooling
layer, it can operate independently on every depth slice of the input and resizes it spatially.
Working on the Brain MRI image can also cost the contamination of the over-fitting and
for this Max Pooling layer perfectly works for this perception. So, for our input image, we
used MaxPooling2D for the proposed model. This convolutional layer runs on 31*31*32
dimension. Input size is 62*62*32 and after finding out the dimension of max pooling layer
using the equation - (4.10), we got the output volume size of 31*31*32. Because of dividing
the input images in both spatial dimensions, the pool size is (2, 2) which means a tuple of
two integers by which to downscale by vertically and horizontally.
5.2. OUR WORKING APPROACH FOR BRAIN TUMOR SEGMENTATION 65
After the pooling layer, a pooled feature map is obtained. Flatten layer is one of the essential
layers after the pooling because we have to transform the whole matrix representing the
input images into a single column vector and its imperative for processing. It is then fed to
the Neural Network for the processing. The dimension of this layer is 31*31*32 = 30752.
Two fully connected layers were employed- Dense-1 and Dense-2 represented the dense
layer. The dense function is applied in Keras for the processing of the Neural Network, and
the obtained vector works as an input for this layer.
There are 128 nodes in the hidden layer. Because the number of dimension or nodes pro-
portional with the computing resources we need to fit our model we kept it as moderate as
possible and for this perspective 128 nodes gives the most substantial result. ReLU is used
as the activation function because of showing better convergence performance.
After the first dense layer, the second fully connected layer was used as the final layer of the
model. In this layer, we used the sigmoid function as activation function where the total
number of the node is one because we need to lower the uses of computing resources so that
a more significant amount assuages the execution time. There is a chance of hampering the
learning in deep networks for using of the sigmoid as the activation function. So we scale
down the sigmoid function, and the number of the nodes is much lesser and easy to handle
for this deep network.
In a nutshell, we have a five-layer CNN model by which we can detect a tumor from an MRI
image. The entire working flow of the five-layer CNN model is illustrated in figure-3. First,
we have to load the input dataset and all the images should be identical in size in the input
image. After the input layer, we introduced a Convolution layer with 32 convolutional fil-
ters along ReLU as an activation function. Max pooling layer is employer after the ConvNet
architecture. Pooled feature map is obtained from Max pooling layer which is used to get
the single column vector and the mechanism is known as Flattening. Finally, we develop
two fully connected layers which are named as a dense layer in our model with 128 nodes.
The sigmoid function is used as the activation function in the final dense layer. Using Adam
optimizer and binary cross-entropy as a loss function, we compiled the model and find the
accuracy of detecting the tumor.
The performance of the CNN model is done by finding the accuracy of our proposed model.
5.3. SUMMARY 66
We have developed an algorithm (Algorithm-1) to find the performance to detect the brain
tumor. The following algorithm was used to evaluate the performance of our CNN model:
5.3 Summary
In this chapter, the proposed methodology of segmentation and detection of brain tumor is
described. Segmentation of the abnormal tissues along with the detection using two distinct
process is thoroughly demonstrated with proper diagram and explanation.
67
Chapter 6
6.1 Overview
In this section, we will comprehensively describe the outcomes of our proposed methodol-
ogy. We carried out the tumor segmentation using Fuzzy C Means (FCM) algorithm and
classify our model as tumor or non-tumor using two distinct mechanisms which are detec-
tion using traditional machine learning and detection using a convolutional neural network.
Following we will do the performance evaluation process and compare the performance of
the two models. Furthermore, we also analyze our model with the existing model in terms
of segmentation and classification.
We used the Jupyter Notebook and various python packages such as Numpy, Pandas, OpenCV
etc for image precessing. For the traditional classifiers, we used the Scikit-Learn. We used
python version 3.6 with Anaconda. For training and testing our model through CNN, we
used Tensorflow and keras framework. We used the dedicated GPU which is provided by
Google Colab.
We used the benchmark dataset in the field of Brain Tumor Segmentation, and that is BRATS
dataset [16]. BRATS is a Multimodal Brain Tumor Segmentation dataset named after MIC-
CAI (The Medical Image Computing and Computer Assisted Intervention Society). The
dataset is a labeled dataset differentiating tumor and non-tumor. The dataset breaks down
6.4. PERFORMANCE MEASURES 68
into two segment- Training set and Testing set. All the images are from three categories- T1-
weighted, T1-weighted and FLAIR (Fluid Attenuated Inversion Recovery). Both sets consist
of two classes. One is for tumorous MRI (class-1) and the other one is for Non-tumorous
MRI (class-0). In the training set, there are 187 tumor MRI images and 30 non-tumor MRI
images. And there are 24 testing images for performance evaluation.
Furthermore, we also used a dataset developed by Jun Cheng [102] where a total of 3064
T1-weighted contrast-enhanced images from 233 patients with three kinds of brain tumor:
meningioma (708 slices), glioma (1426 slices), and pituitary tumor (930 slices). It is only
used for the evaluation of the CNN model.
We have to discuss about the performance matrices in order to know in what extent our
model works accurately. In this section we will discuss about the performance measure that
we consider to evaluate our model. We need to familiarize with some terms regarding the
performance measures.
The Confusion matrix is one of the most intuitive metrics used to find the correctness and
accuracy of the model. It is used for Classification problem where the output can be of
two or more types of classes. The Confusion Matrix consists of four parameters which are
described below-
• True Positive (TP): Number of tumor images that are correctly classified.
• True Negative (TN): Number of non-tumor images that are correctly classified.
• False Positive (FP): Number of non-tumor images that are misclassified as tumor.
• False Negative (FN): Number of tumor images that are misclassified as non-tumor.
6.4.2 Accuracy
It’s the most popular performance matrix which measures how often the classifier produce
the correct prediction. Mathematically Accuracy defined as the ratio of the number of correct
6.4. PERFORMANCE MEASURES 69
predicted images and the total number of images and symbolically represented as-
C or r ec t P r edic t ions TP + TN
Accur ac y = = (6.1)
Tot al number o f images TP + TN + FP + FN
6.4.3 Precision
It is the retrieved information that are relevant to the model. Precision is the ratio of the
number of tumor images that are correctly classified (TP) and the number of images clas-
sified or misclassified as tumor (TP + FP). The lower the FP the higher the Precision. The
model is more effective in case of higher precision rate.
TP
P r ecision = (6.2)
TP + FP
6.4.4 Recall
It is the fraction of the images which are successfully retrieved. Recall is the ratio of the
number of tumor images that are correctly classified and the number of images that are to
be predicted. Sensitivity, Hit Rate, True Positive Rate are the other names of Recall. The
lower the False negative the higher the recall because the number of tumor images that are
classified as non-tumor is low.
TP
Recall = (6.3)
TP + FN
6.4.5 F-Score
It is the harmonic mean of Precision and Recall and is a measure of test accuracy. F-score
reaches its best value at 1 (100% precision and recall) and worst value at 0. F-Score can be
defined as-
P r ecision ∗ Recall 2T P
F − Scor e = 2 ∗ = (6.4)
P r ecision + Recall 2T P + F P + F N
6.4.6 Specificity
Specificity is the True Negative Rate (TNR) of the model and the statistical measure of the
binary classification test. As we are dealing with a binary classification (tumor or non-
tumor) so we can use this as the performance evaluation. It is the ratio of the number
of non-tumor images that are correctly classified (TN) and the number of images that are
6.6. EXPERIMENTAL RESULTS 70
classified or misclassified as non-tumor (TN + FP). The lower the false positive (FP) the
higher the specificity or selectivity.
TN
Speci f ici t y = T N R = (6.5)
TN + FP
In this section, the hyper-parameter values in initialization and training stage that were used
to implement our CNN model are mentioned. These information are depicted in the chart
below:
In this section, we will discuss the experimental result of the proposed methodology. We
divide this section into three subsection which we will describe thoroughly.
In this section, we will illustrate all the output images after each step of segmentation based
on the proposed methodology.
6.6. EXPERIMENTAL RESULTS 71
In Brain MRI image, Skull plays a less important role when it comes to segment the tumor. So
we can say the skull is unnecessary or redundant part of the MRI. In six steps we remove the
skull from the MRI. These six steps are: Gray-scale conversion, thresholding, Morphological
operations, enumeration of the connected component, finding the mask and lastly multiply
the mask to the MRI image. Fig-6.1 is the depiction of the skull removed image from input
Brain MRI image.
In this section, we enhanced the MRI image. Using the Gaussian blur filter, we remove
the noise and enhancement is done using the Addweighted method. After that, we used
our segmentation algorithm. We have shown the steps of this step in Figure-6.2 where the
output of step-1 (skull removed MRI) is going as input in this section and output is the
enhanced MRI. We also have shown the filtered image which is done using Gaussian blur
filter, in between the skull removed MRI and Enhanced MRI.
(a) Skull Removed Image - 1 (b) Filtered Image - 1 (c) Enhanced Image - 1
(d) Skull Removed Image - 2 (e) Filtered Image - 2 (f) Enhanced Image - 2
(g) Skull Removed Image - 3 (h) Filtered Image - 3 (i) Enhanced Image - 3
After Image Enhancement, we segment the abnormal tissues which are the brain tumor from
the MRI. Using Fuzzy C Means segmentation algorithm we distinguish between the normal
cells and abnormal cells. We used FCM, K-Means clustering, Normalized cut, Watershed
segmentation, and thresholding method. But FCM gives us the improved result as it ensures
that one piece of data to belong to two or more clusters which are important for Brain MRI.
In Figure-6.3 the segmented brain tumor is shown.
Figure 6.3: Segmentation using FCM of the enhance brain MRI image
6.6. EXPERIMENTAL RESULTS 74
Finally, the last step is tumor contouring. In this step, we showed the contoured brain
tumor MRI image. It is circled in green color indicates the location of the brain tumor.
Tumor cluster extraction is done by an intensity-based approach named as thresholding.
Figure-6.4 illustrated the contoured brain tumor MRI images.
We employed Six traditional Machine Learning Algorithms to classify our model. The Six
traditional Machine Learning Algorithms are K-Nearest Neighbor, Logistic Regression, Multi-
layer Perceptron, Naive Bayes, Random Forest, and Support Vector Machine.
Based on splitting ratio we have shown the performance of the proposed model into two
experiments which are described in the following section.
6.6.2.1 Experiment - I
In the first experiment, we split the dataset into 70 by 30 ratio which means 70% of the
data goes for training and 30% of the data goes for testing. In Table - 6.2, we represented
the performance of our proposed model based on this ratio. We can see that, SVM gives the
best result in terms of Accuracy, Recall, Dice score and Jaccard Index.
Jaccard
Classifiers Accuracy Recall Specificity Precision Dice Score
Index
K-Nearest Neighbor 0.8939 0.949 0.4280 0.9330 0.9410 0.889
Logistic Regression 0.8788 0.949 0.286 0.918 0.933 0.875
Multi-layer Perception 0.8939 1.000 0.167 0.894 0.944 0.894
Naive Bayes 0.7879 0.797 0.714 0.959 0.870 0.770
Random Forest 0.8939 0.983 0.167 0.903 0.943 0.892
Support Vector Machine 0.9242 0.983 0.428 0.935 0.959 0.921
6.6.2.2 Experiment - II
In the second experiment, we split the dataset into 80 by 20 ratio which means 80% of the
data goes for training and 20% of the data goes for testing. In Table - 6.3, we represented
the performance of our proposed model based on this ratio. Accuracy of 88.63% is achieved
using SVM as the classifier.
Jaccard
Classifiers Accuracy Recall Specificity Precision Dice Score
Index
K-Nearest Neighbor 0.8409 0.40 0.8810 0.48 0.465 0.412
Logistic Regression 0.8863 0.50 0.9050 0.20 0.285 0.167
Multi-layer Perception 0.8863 0.41 0.8864 0.48 0.442 0.465
6.6. EXPERIMENTAL RESULTS 76
We trained our model based on two datasets. We have operated a different number of
layers with a different value of hyper-parameters and we got the best result using the five-
layer CNN model with respected splitting ratio and other criteria. Based on the number of
layers, we break down the performance of the proposed CNN model into several sections.
We will show the experimental results along with other characteristics and evaluation in
this following sub-section. In Experiment-I and Experiment-II, we have shown the accuracy
while tuning the learning rate and epoch using the five-layer model based on 70:30 and
80:20 splitting ratio respectively. Later, we have shown the complete outcomes of using
five, six and seven layers CNN model and compare the results.
6.6.3.1 Experiment-I
In this section, the proposed five-layer CNN model is trained by splitting the dataset into 70
by 30 ratio. Table - 6.4 represents the relation between learning rate, epochs, time to train
and accuracy of the model based on this splitting ratio. When the learning rate is 0.001,
epoch is 50, training time is 500 sec then we get the best accuracy: 96.55% in this splitting
ratio.
Table 6.4: Training Time and Accuracy of the proposed CNN model (splitting ratio 70:30)
6.6.3.2 Experiment - II
In this section, the proposed five-layer CNN model is trained by splitting the dataset into
80 by 20 ratio. In Table-6.5, a comparison of training time and accuracy is shown which is
based on learning rate and epochs. When the learning rate is 0.001, epoch is 10 then we
get the best accuracy: 96.55% and the training time is 175 sec in this splitting ratio.
Table 6.5: Training Time and Accuracy of the proposed CNN model (splitting ratio 80:20)
Furthermore, for the justification of the proposed five-layer CNN model, we have shown
several experimental results based on different hyper-parameters and splitting ratio. In
table-6.6, we showed the performance of five-layer CNN model based on 80:20 and 70:30
splitting ratio. From the table-6.6 we can see that when the splitting ratio is 80:20, the batch
size is 64 and epoch is 10 the model provides the best result of 97.87% as accuracy. After
this point, the model overfits the data. So, for five layer CNN model, we get the accuracy of
97.87%.
6.6. EXPERIMENTAL RESULTS 78
Convolution Layer Max Pooling Split Ratio Batch Size Epoch Accuracy (%)
8 92.72
9 85.82
32
10 86.85
11 87.88
80:20
8 93.67
9 94.98
64
10 97.87
11 94.89
62*62*32 31*31*32
8 81.35
9 83.71
32
10 87.87
11 89.13
70:30
8 88.07
9 88.76
64
10 91.23
11 94.90
In this section, we add one more convolutional layer of size 62*62*32. We calculated the
size of the convolutional layer using equation-4.8. Changing the dimension could cause
the model with poor accuracy. Using two convolutional layers, one max pooling layer, one
flatten layer and two fully connected layers, we get maximum accuracy of 94.21%. In table-
6.6, we got the maximum accuracy when the splitting ratio is 80:20, the batch size is 64 and
epoch is 11. After this point, the model accuracy started decreasing gradually. So adding
more convolutional layer will not increase the accuracy.
Convolution Layer Max Pooling Split Ratio Batch Size Epoch Accuracy (%)
8 89.29
9 92.83
32
10 93.76
62*62*32
11 93.62
62*62*32 31*31*32 80:20
8 94.01
(2 layer)
9 94.08
64
10 94.39
6.6. EXPERIMENTAL RESULTS 79
In the fifth phase of our experiment, we employed seven-layer (two convolutional layers,
two max pooling layers, one flatten layer and two fully connected layers) of CNN model
to measure the performance for the detection of brain tumor. As we seen in the earlier
experiment that, adding more convolutional layer does not improve the performance of the
model, we increase the Max Pooling layer. From table-6.8, we get the maximum accuracy
of 95.67% when the splitting ratio is 80:20, batch size is 64 and epoch is 9.
Convolution Layer Max Pooling Split Ratio Batch Size Epoch Accuracy (%)
8 88.17
9 90.16
32
10 89.02
8 95.33
80:20
9 95.67
64
10 95.21
11 94.99
62*62*32 31*31*32
8 79.73
31*31*128 15*15*128
9 84.17
(3 layer) (2 layer) 32
10 84.38
11 84.31
70:30
8 87.07
9 87.12
64
10 87.19
11 87.18
6.6. EXPERIMENTAL RESULTS 80
Two graphical representation is depicted below in the two figures based on the splitting
ratio 70:30 of the proposed CNN model using the BRATS dataset. In fig-6.5 and fig-6.6, the
splitting ratio is 70:30. In fig-6.6 which is the loss curve where the training and validation
curve intersected in the earlier cycle of the epoch. So, the validation rate is not proportional
with the model accuracy from the start which is not good for the model.
Figure 6.5: Model Accuracy curve based on 70:30 split ratio (BRATS Dataset)
Figure 6.6: Model Loss Curve based on 70:30 split ratio (BRATS Dataset)
6.6. EXPERIMENTAL RESULTS 81
In fig-6.7, a representation of accuracy curve of the model is depicted. We can see from the
loss curve (fig-6.8) that at epoch six the training and validation curve met. Hence we can
say that for splitting ratio 80:20 the model gives the best result.
Figure 6.7: Model Accuracy Curve based on 80:20 split ratio (BRATS Dataset)
Figure 6.8: Model Loss Curve based on 80:20 split ratio (BRATS Dataset)
6.6. EXPERIMENTAL RESULTS 82
Fig-6.9 is the depiction of the of the Accuracy of the six algorithms based on 70:30 distri-
bution ratio. Here, SVM gives the best result.
Figure 6.9: Accuracy of the Proposed Model using ML classifiers based on 70:30 split ratio
In fig-6.10, a representation of the of the Accuracy of the six algorithms based on 80:20
distribution ratio. Here, SVM gives the most improved result.
Figure 6.10: Accuracy of the Proposed Model using ML classifiers based on 80:20 split ratio
6.6. EXPERIMENTAL RESULTS 83
Fig 6.11 illustrated a bar chart of learning rate vs training time. We selected three values of
learning rate such as 0.001, 0.005 and 0.01 based on this splitting ratio. The training time
is increasing with the learning rate and at 0.001 learning rate and 10 epochs we get the
minimal amount of training time.
Figure 6.11: Learning Rate Vs Training Time of proposed CNN model (based on 80:20
splitting)
A bar chart of learning rate vs Accuracy is illustrated in fig. 6.12. We selected four values of
epoch such as 10, 20, 50 and 100 based on this splitting ratio. The accuracy of the model
is decreasing as the learning rate increases gradually. We get the best accuracy of 97.87%
at 0.001 learning rate.
Figure 6.12: Learning Rate Vs Accuracy of proposed CNN model (based on 80:20 split ratio)
6.6. EXPERIMENTAL RESULTS 84
Figure 6.13 illustrated a bar chart of learning rate vs training time. We selected three values
of learning rate such as 0.001, 0.005 and 0.01 based on this splitting ratio. The training
time is increasing with the learning rate and at 0.001 learning rate and 10 epochs we get
the minimal amount of training time: 180 seconds.
Figure 6.13: Learning Rate Vs Training Time of proposed CNN model (based on 70:30 split
ratio)
In fig. 6.14, a bar chart of learning rate vs Accuracy is shown. We selected four values of
epoch such as 10, 20, 50 and 100 based on this splitting ratio. We get the best accuracy of
92.88% at 0.001 learning rate.
Figure 6.14: Learning Rate Vs Accuracy of proposed CNN model (based on 70:30 split ratio)
6.7. PERFORMANCE COMPARISON 85
In this section, we compared the performance of the existing models to the proposed models.
Though we used two dataset to measure the performance of the proposed model, we will
take into consideration of the results of the BRATS dataset.
We compared the proposed model of classification using Traditional Machine Learning al-
gorithm and the CNN model depending on different splitting ratio. In traditional machine
learning model, we evaluate the model based on two splitting techniques. Among them
70:30 splitting ratio gives us the best result of 92.42% for traditional machine learning.
For the proposed CNN model, three types of CNN model is used and we got 97.87% as the
best accuracy for CNN models for five-layer CNN model. Table-6.9 showed the comparison
of the CNN model based on the experiment - III, IV, V.
No. of Convolution
Max Pooling Split Ratio Batch Size Epoch Accuracy (%)
Layers Layer
Five 62*62*32 31*31*32 80:20 64 10 97.87
62*62*32
Six 31*31*32 80:20 64 10 94.39
62*62*32
62*62*32 31*31*32
Seven 80:20 64 9 95.67
31*31*128 15*15*128
Table-6.10 illustrated the comparison of these two models. Comparing the two best results
from proposed traditional machine learning and CNN model, CNN outperforms ML in terms
of Accuracy. From this observation we can assert that the five-layer CNN model provides us
the most better result among the other distribution and the learning rate is 0.001, epoch
10 and the training time is 175 secs.
Table 6.10: Performance Comparison Between the proposed ML and CNN model
In this section, we will do a comparison among the existing model and our proposed method-
ology of five layer CNN model. Table-6.11 illustrated the comparison between the existing
models and our models. We took seven related research articles based on the same dataset
from different methods and compare their outcomes with the proposed CNN model.
Table 6.11: Performance comparison with the existing models and the proposed CNN model
6.9 Summary
We have shown the performance matrices with a short description and the value of the
hyper-parameters for which we got the best result for the proposed CNN model. After that,
the output of each step in the segmentation of the tumor is depicted. Later, we have shown
all the experimental and empirical results with proper justification. Finally, we have done a
comparison between our proposed models and also with the existing models.
87
Chapter 7
7.1 Conclusion
Performance analysis of automated brain tumor detection from MR imaging and CT scan
using basic image processing techniques based on various hard and soft computing has
been performed in our work. Moreover, we applied six traditional classifiers to detect brain
tumor in the images. Then we applied CNN for brain tumor detection to include deep
learning method in our work. We compared the result of the traditional one having the best
accuracy (SVM) with the result of CNN. Furthermore, our work presents a generic method
of tumor detection and extraction of its various features.
In the context of the full dataset, it is necessary to parallelize and utilize high-performance
computing platform for maximum efficiency. We tried our best to detect the tumors accu-
rately but, nevertheless we faced some problems in our work where tumor could not be
detected or falsely detected. So, we will try to work on those images and on the complete
dataset. Hence, we will try to apply other deep learning methods in the future so that we
can get a more precise and better result.
7.2 Limitations
There are some limitations of our thesis work that we have listed in this section which we
are leaving to improve in our future works.
There are more opportunities for improvement or research on our work in the future.
• Firstly, the number of images can be increased. The bigger the number of the images
is, the better the model is trained.
• Thirdly, more traditional classifiers can be applied to get more increased accuracy.
• Fourthly, we will try to classify the tumor if its benign or malignant after the detection
of the tumor.
• Last but not the least, more variations of deep learning methods can be tested in
future.
89
References
[1] McKinney PA Brain tumours: incidence, survival, and aetiology Journal of Neurology,
Neurosurgery & Psychiatry 2004;75:ii12-ii17.
[2] “General Information About Adult Brain Tumors”. NCI. Last Accessed Date: May 10,
2019.
[3] Nunn, Laura Silverstein; Silverstein, Alvin; Silverstein, Virginia B. (2006). “Cancer.
Brookfield, Conn: Twenty-First Century Books”. pp. 11-12. ISBN 0-7613-2833-5.
[4] David Lowell Strayer; Raphael Rubin; Rubin, Emanuel (2008). “Rubin’s pathology: clin-
icopathologic foundations of medicine”. Philadelphia: Wolters Kluwer/Lippincott Williams
& Wilkins. pp. 138-139. ISBN 0-7817-9516-8.
[5] https://www.researchgate.net/figure/Example-of-type-III-tumor-involvement-A-49-year-
old-man-with-lung-cancer-A-A-selective-fig3-38042515, Last Accessed Date: May 10, 2019.
[8] Khurram Shehzad, Imran Siddique, Obed Ullah Memon, “Efficient Brain Tumor Detec-
tion Using Image Processing Techniques”, International Journal of Scientific & Engineering
Research, December 2016.
[10] Prof. Vrushali Borase., Prof. Gayatri Naik. and Prof. Vaishali Londhe, “Brain MR
Image Segmentation for Tumor Detection using Artificial Neural”, International Journal Of
Engineering And Computer Science, 2017.
[11] Umit Ilhan, Ahmet Ilhan, “Brain tumor segmentation based on a new threshold ap-
proach,” Procedia Computer Science, ISSN: 1877-0509, Vol: 120, Page: 580-587, Publica-
REFERENCES 90
[12] Rohini Paul Joseph, C. Senthil Singh and M.Manikandan, “Brain MRI Image Segmen-
tation and Detection in Image Processing”, International Journal of Research in Engineering
and Technology, 2014.
[13] Miss. Shrutika Santosh Hunnur, Prof. Akshata Raut, Prof. Swati Kulkarni, “Implemen-
tation of Image Processing for Detection of Brain Tumors”, IEEE, 2017.
[14] Anam Mustaqeem, Ali Javed, Tehseen Fatima, “An Efficient Brain Tumor Detection
Algorithm Using Watershed & Thresholding Based Segmentation”, I.J. Image, Graphics and
Signal Processing, 2012,10,34-39.
[15] B. Devkota, Abeer Alsadoon, P.W.C. Prasad, A. K. Singh, A. Elchouemi, “Image Segmen-
tation for Early Stage Brain Tumor Detection using Mathematical Morphological Reconstruc-
tion,” 6th International Conference on Smart Computing and Communications, ICSCC 2017,
7-8 December 2017, Kurukshetra, India.
[16] Song, Yantao & Ji, Zexuan & Sun, Quansen & Yuhui, Zheng. (2016). “A Novel Brain
Tumor Segmentation from Multi-Modality MRI via A Level-Set-Based Model”. Journal of
Signal Processing Systems. 87. 10.1007/s11265-016-1188-4.
[17] Ehab F. Badran, Esraa Galal Mahmoud, Nadder Hamdy, “An Algorithm for Detecting
Brain Tumors in MRI Images”, 7th International Conference on Cloud Computing, Data Science
& Engineering - Confluence, 2017.
[18] Pei L, Reza SMS, Li W, Davatzikos C, Iftekharuddin KM. “Improved brain tumor seg-
mentation by utilizing tumor growth model in longitudinal brain MRI”. Proc SPIE Int Soc
Opt Eng. 2017.
[19] Dina Aboul Dahab, Samy S. A. Ghoniemy, Gamal M. Selim, “Automated Brain Tumor
Detection and Identification using Image Processing and Probabilistic Neural Network Tech-
niques”, IJIPVC, Vol. 1, No. 2, pp. 1-8, 2012.
[20] Mohd Fauzi Othman, Mohd Ariffanan and Mohd Basri, “Probabilistic Neural Network
for Brain Tumor Classification”, 2nd International Conference on Intelligent Systems, Mod-
elling and Simulation, 2011.
[21] A. Rajendran, R. Dhanasekaran, “Fuzzy Clustering and Deformable Model for Tumor
Segmentation on MRI Brain Image: A Combined Approach,” International Conference on
Communication Technology and System Design 2011.
[22] Sobhaninia, Zahra & Rezaei, Safiyeh & Noroozi, Alireza & Ahmadi, Mehdi & Zarrabi,
Hamidreza & Karimi, Nader & Emami, Ali & Samavi, Shadrokh. (2018). “Brain Tumor
Segmentation Using Deep Learning by Type Specific Sorting of Images”.
REFERENCES 91
[23] Kaatsch P (June 2010). “Epidemiology of childhood cancer”. Cancer Treatment Re-
views. 36 (4): 277-85.
[24] Maheswari, D. and Radha V. (2010) “Enhanced Layer Based Compound Image, Com-
pression Proceedings of the First Amrita ACM-W celebration of Women in Computing (A2CWIC
2010)”, Pp. 209- 216,ISBN:978-1-4503-0194-7.
[25] Yi Yang, Sam Hallman, Deva Ramanan, Charless C. Fowlkes. (2009-2010). “Layered
object Models for Image Segmentation”.
[26] Wankai Deng , Wei Xiao, He Deng, Jianguo Liu, “MRI Brain Tumor Segmentation With
Region Growing Method Based On The Gradients And Variances Along And Inside Of The
Boundary Curve”, 3rd International Conference on Biomedical Engineering and Informatics,
2010.
[27] Anam Mustaqeem, Ali Javed, Tehseen Fatima, “An Efficient Brain Tumor Detection
Algorithm Using Watershed & Thresholding Based Segmentation”, I.J. Image, Graphics and
Signal Processing, 2012, 10, 34-39.
[28] Jianping Fan, D. K. Y. Yau, A. K. Elmagarmid and W. G. Aref, “Automatic image segmen-
tation by integrating color-edge extraction and seeded region growing,” in IEEE Transactions
on Image Processing, vol. 10, no. 10, pp. 1454-1466, Oct. 2001.
[29] MacKay, David (2003). “Chapter 20. An Example Inference Task: Clustering” (PDF).
Information Theory, Inference and Learning Algorithms. Cambridge University Press. pp.
284&ndash, 292. ISBN 0-521-64298-1. MR 2012999.
[30] S. R. Telrandhe, A. Pimpalkar and A. Kendhe, “Detection of brain tumor from MRI
images by using segmentation & SVM,” 2016 World Conference on Futuristic Trends in
Research and Innovation for Social Welfare (Startup Conclave), Coimbatore, 2016, pp. 1-6.
[31] J. C. Dunn (1973): “A Fuzzy Relative of the ISODATA Process and Its Use in Detecting
Compact Well-Separated Clusters”, Journal of Cybernetics 3: 32-57
[32] J. C. Bezdek (1981): “Pattern Recognition with Fuzzy Objective Function Algoritms”,
Plenum Press, New York
[33] Wankai Deng , Wei Xiao, He Deng, Jianguo Liu, “MRI Brain Tumor Segmentation With
Region Growing Method Based On The Gradients And Variances Along And Inside Of The
Boundary Curve”, 3rd International Conference on Biomedical Engineering and Informatics ,
2010.
[34] S. K. Adhikari, J. K. Sing, D. K. Basu and M. Nasipuri, “A spatial fuzzy C-means algo-
rithm with application to MRI image segmentation,” 2015 Eighth International Conference
on Advances in Pattern Recognition (ICAPR) , Kolkata, 2015, pp. 1-6.
REFERENCES 92
[35] M. U. Akram and A. Usman, “Computer aided system for brain tumor detection and
segmentation,” International Conference on Computer Networks and Information Technology
, Abbottabad, 2011, pp. 299-302.
[36] Shree, N. Varuna and T. N. R. Kumar. “Identification and classification of brain tumor
MRI images with feature extraction using DWT and probabilistic neural network.” Brain
Informatics (2017).
[37] Debnath Bhattacharyya and Tai-hoon Kim, “Brain Tumor Detection Using MRI Image
Analysis,” UCMA 2011, Part II, CCIS 151, pp. 307-314, 2011., Springer-Verlag Berlin Hei-
delberg 2011.
[38] Malathi, R. and Dr. Nadirabanu Kamal, A.R., “Evaluating the Hausdorff Distance for
Contour Segmentation of Brain Images,” Australian Journal of Basic and Applied Sciences,
8(16) October 2014, Pages: 100-111.
[40] Gonzalez, Rafael C., and Richard E. Woods. Digital Image Processing. Addison-Wesley,
1993.
[41] M. Sezgin & B. Sankur (2004). “Survey over image thresholding techniques and quan-
titative performance evaluation". Journal of Electronic Imaging. 13 (1): 146-165
[42] Kirti Mittal, Abhishek Shekhar, Prashant Singh, Manish Kumar, “Brain Tumour Ex-
traction using Otsu Based Threshold Segmentation”, International Journal of Advanced Re-
search in Computer Science and Software Engineering, Volume 7, Issue 4, April 2017.
[43] K. Rajesh Babu, P.V. Naganjaneyulu and K. Satya Prasad, “Performance Analysis for Ef-
ficient Brain Tumor Segmentation by using Clustering Algorithm,” Indian Journal of Science
and Technology, Vol 10(11), March 2017.
[44] Ming-Ni Wu, Chia-Chen Lin, Chin-Chen Chang, “Brain Tumor Detection Using Color-
Based K-Means Clustering Segmentation”, Third International Conference on Intelligent In-
formation Hiding and Multimedia Signal Processing , 2007.
[45] J.selvakumar, A.Lakshmi, T.Arivoli, “Brain Tumor Segmentation and Its Area Calcu-
lation in Brain MR Images using K-Mean Clustering and Fuzzy C-Mean Algorithm”, IEEE-
International Conference On Advances In Engineering, Science And Management (ICAESM
-2012) March 30, 31, 2012.
[46] Jianwei Liu, Lei Guo, “A New Brain MRI Image Segmentation Strategy Based on K-
means Clustering and SVM”, 7th International Conference on Intelligent Human-Machine
Systems and Cybernetics , 2015.
REFERENCES 93
[47] J. Vijay and J. Subhashini, “An efficient brain tumor detection methodology using K-
means clustering algorithm,” 2013 International Conference on Communication and Signal
Processing, Melmaruvathur, 2013, pp. 653-657.
[48] C. H. Rao, P. V. Naganjaneyulu and K. S. Prasad, “Brain Tumor Detection and Segmen-
tation Using Conditional Random Field,” 2017 IEEE 7th International Advance Computing
Conference (IACC) , Hyderabad, 2017, pp. 807-810.
[49] Parveen, Amritpal singh, “Detection of Brain Tumor in MRI Images, using Combina-
tion of Fuzzy C-Means and SVM”, 2nd International Conference on Signal Processing and
Integrated Networks (SPIN) , 2015.
[50] T.Logeswari and M.Karnan “An Improved Implementation of Brain Tumor Detection
Using Segmentation Based on Hierarchical Self Organization Map” International Journal of
Computer Theory and Engineering, Vol. 2, No. 4, August, 2010 1793-201.
[51] N. N. Gopal and M. Karnan, “Diagnose brain tumor through MRI using image process-
ing clustering algorithms such as Fuzzy C Means along with intelligent optimization tech-
niques,” 2010 IEEE International Conference on Computational Intelligence and Computing
Research , Coimbatore, 2010, pp. 1-4.
[52] N. Gordillo, E. Montseny and P. Sobrevilla, “A new fuzzy approach to brain tumor
segmentation,” International Conference on Fuzzy Systems, Barcelona, 2010, pp. 1-8.
[53] Chaiyanan Sompong, Sartra Wongthanavasu, “Brain Tumor Segmentation Using Cel-
lular Automata-based Fuzzy C-Means”, 13th International Joint Conference on Computer
Science and Software Engineering (JCSSE) , 2013.
[54] Lawrene O. Hall, Amine M. Bensaid, Laurene P. Clarke, Robert P. Velthuizen, Martin
S. Silbiger, and James C. Bezdek, “A Comparison of Neural Network and Fuzzy Clustering
Techniques in Segmenting Magnetic Resonance Images of the Brain” IEEE TRANSACTIONS
ON NEURAL NETWORKS, VOL 3, NO 5, SEPTEMBER 1992.
[55] Bing Nan Li, Chee Kong Chui, Stephen Chang, S.H. Ong, “Integrating spatial fuzzy
clustering with level set methods for automated medical image segmentation” Computers
in Biology and Medicine 41(2011) 1-10.
[56] Pranita Balaji Kanade, Prof. P.P. Gumaste, “Brain Tumor detection Using MRI images”,
International Journal of Innovative Research in Electrical, Electronics, Instrumentation and
Control Engineering Vol. 3, Issue 2, February 2015.
[57] Eman Abdel-Maksoud, Mohammed Elmogy, Rashid Al-Awadi, “Brain tumor segmenta-
tion based on a hybrid clustering technique”, Elsevier. Volume 16, Issue 1, March 2015.
[58] K. S. Angel Viji and J. Jayakumari, “Automatic detection of brain tumor based on mag-
netic resonance image using CAD System with watershed segmentation,” 2011 International
REFERENCES 94
[59] Dawood Dilber, Jasleen, “Brain Tumor Detection using Watershed Algorithm”, Interna-
tional Journal of Innovative Research in Science, Engineering and Technology, Vol. 5, Issue
3, March 2016.
[61] Bauer S., Nolte LP., Reyes M. (2011) “Fully Automatic Segmentation of Brain Tumor
Images Using Support Vector Machine Classification in Combination with Hierarchical Con-
ditional Random Field Regularization”. In: Fichtinger G., Martel A., Peters T. (eds) Medical
Image Computing and Computer-Assisted Intervention - MICCAI 2011. MICCAI 2011.
[62] S. Pereira, A. Pinto, V. Alves and C. A. Silva, “Brain Tumor Segmentation Using Con-
volutional Neural Networks in MRI Images,” in IEEE Transactions on Medical Imaging, vol.
35, no. 5, pp. 1240-1251, May 2016.
[63] Mohammad Havaei, Axel Davy, David Warde-Farley, Antoine Biard, Aaron Courville,
Yoshua Bengio, Chris Pal, Pierre-Marc Jodoin, Hugo Larochelle, “Brain tumor segmentation
with Deep Neural Networks”, Medical Image Analysis, Volume 35, 2017, Pages 18-31, ISSN
1361-8415
[64] J. J. Corso, E. Sharon, S. Dube, S. El-Saden, U. Sinha and A. Yuille, “Efficient Mul-
tilevel Brain Tumor Segmentation With Integrated Bayesian Model Classification,” in IEEE
Transactions on Medical Imaging, vol. 27, no. 5, pp. 629-640, May 2008.
[65] Shenbagarajan, A., Ramalingam, V., Balasubramanian, C., & Palanivel, S. (2016). “Tu-
mor Diagnosis in MRI Brain Image using ACM Segmentation and ANN-LM Classification
Techniques.” Indian Journal Of Science And Technology, 9(1).
[67] Jainy Sachdeva, Vinod Kumar, Indra Gupta, Niranjan Khandelwal, Chirag Kamal Ahuja,
“A novel content-based active contour model for brain tumor segmentation”, Magnetic Res-
onance Imaging, Volume 30, Issue 5, 2012, Pages 694-715, ISSN 0730-725X.
[68] Mariam Saii, Zaid Kraitem, “Automatic Brain tumor detection in MRI using image pro-
cessing techniques”, Biomedical Statistics and Informatics, Vol. 2, No. 2, pp. 73-76, 2017.
[69] Zexuan Ji, Yong Xia, , Quansen Sun, Qiang Chen, Deshen Xia, David Dagan Feng,
REFERENCES 95
“Fuzzy Local Gaussian Mixture Model for Brain MR Image Segmentation”, IEEE Transactions
on Information Technology in Biomedicine, Vol. 16, No. 3, May 2012.
[70] M. Huang, W. Yang, Y. Wu, J. Jiang, W. Chen and Q. Feng, “Brain Tumor Segmenta-
tion Based on Local Independent Projection-Based Classification,” in IEEE Transactions on
Biomedical Engineering, vol. 61, no. 10, pp. 2633-2645, Oct. 2014.
[71] Nilesh Bhaskarrao Bahadure, Arun Kumar Ray, Har Pal Thethi, “Image Analysis for MRI
Based Brain Tumor Detection and Feature Extraction Using Biologically Inspired BWT and
SVM”, International Journal of Biomedical Imaging Volume 2017.
[72] A. Demirhan, M. TÃűrÃij and Äř. GÃijler, “Segmentation of Tumor and Edema Along
With Healthy Tissues of Brain Using Wavelets and Neural Networks,” IEEE Journal of Biomed-
ical and Health Informatics, vol. 19, no. 4, pp. 1451-1458, July 2015.
[73] Binaghi, Elisabetta & Omodei, Massimo & Pedoia, Valentina & Balbi, Sergio & Lattanzi,
Desiree & Monti, Emanuele. (2014). “Automatic Segmentation of MR Brain Tumor Images
using Support Vector Machine in Combination with Graph Cut”. NCTA 2014 - Proceedings
of the International Conference on Neural Computation Theory and Applications. 152-157.
10.5220/0005068501520157.
[74] A. Rajendran, R. Dhanasekaran, “Fuzzy Clustering and Deformable Model for Tumor
Segmentation on MRI Brain Image: A Combined Approach,” International Conference on
Communication Technology and System Design, 2011.
[75] Avanzo M, Stancanello J, El Naqa I. “Beyond imaging: the promise of radiomics.” Phys
Med 2017;38:122-39.
[76] Erickson BJ, Korfiatis P, Akkus Z, Kline TL. “Machine learning for medical imaging.”
Radiographics 2017;37:505-15.
[79] Ravi Srisha, Am Khan, “Morphological Operations for Image Processing : Understand-
ing and its Applications”, National Conference on VLSI, Signal processing & Communications
, December 2013.
[80] https://stackoverflow.com/questions/44179793/segmenting-an-object-from-background-
using-matlab-based-on-feature-points/44180904#44180904, Last Accessed Date: April 26,
2019.
[82] Rohini Paul Joseph, C. Senthil Singh and M.Manikandan, “Brain MRI Image Segmen-
tation and Detection in Image Processing”, International Journal of Research in Engineering
and Technology, 2014.
[83] https://www.mathworks.com/matlabcentral/fileexchange/25532-fuzzy-c-means-segmentation,
Last Accessed Date: April 30, 2019.
[86] https://medium.com/machine-learning-for-humans/neural-networks-deep-learning-cdad8aeae49b,
Last Accessed Date: May 10, 2019.
[87] https://hackernoon.com/everything-you-need-to-know-about-neural-networks-8988c3ee4491,
Last Accessed Date: May 10, 2019.
[88] Mehrotra, Kishan & Mohan, Chilukuri & Ranka Preface, “Elements of Artificial Neural
Nets”, 1997.
[89] https://www.superdatascience.com/blogs/convolutional-neural-networks-cnn-step-1-
convolution-operation, Last Accessed Date: June 10, 2019.
[90] LeCun, Yann & Bengio, Y & Hinton, Geoffrey. “Deep Learning”, Nature, 521. 436-44.
10.1038/nature14539, 2015.
[92] https://medium.com/@amarbudhiraja/https-medium-com-amarbudhiraja-learning-less-
to-learn-better-dropout-in-deep-machine-learning-74334da4bfc5, Last Accessed Date: June
10, 2019.
[94] https://medium.com/@AI_with_Kain/understanding-of-multilayer-perceptron-mlp-8f179c4a135f,
Last Accessed Date: June 5, 2019.
[95] Sona Taheri, Musa Mammadov, “Learning the Naive Bayes Classifier with Optimization
Models”, Int. J. Appl. Math. Computer Sci., 2013, Vol. 23, No. 4, 787-795.
[96] Hastie, Trevor; Tibshirani, Robert; Friedman, Jerome (2008). The Elements of Statis-
REFERENCES 97
[97] Ho TK (1998). “The Random Subspace Method for Constructing Decision Forests”,
IEEE Transactions on Pattern Analysis and Machine Intelligence. 20 (8): 832-844.
[99] https://towardsdatascience.com/support-vector-machine-introduction-to-machine-learning-
algorithms-934a444fca47, Last Accessed Date: June 15, 2019.
[100] Kim, J. I. N. H. O., B. S. Kim, and Silvio Savarese. “Comparing image classification
methods: K-nearest-neighbor and support-vector-machines.” Proceedings of the 6th WSEAS
international conference on Computer Engineering and Applications, and Proceedings of the
2012 American conference on Applied Mathematics. Vol. 1001. 2012.
[101] Sujan, Md. Hayder Khan et al. “A Segmentation based Automated System for Brain
Tumor Detection.” (2016).
98