Review 3
Review 3
Review-III
School: SCOPE
Guide: Prof. Rajkumar R
Team Members:
1.Aishna Gupta - 19BCB0101
2.Sanjeev - 19BCE0181
3.Sudhanshu Gupta - 19BCE0188
CONTENTS:
1. Abstract
2. Literature survey
3. Aim of project
4. Objective of project
5. Motivation behind
6. System Architecture diagram
7. Module Description
8. Methodology
9. Conclusion and future scope
10. References
Neuroimaging-based approaches have been extensively applied
to study brain disease in recent years and have deepened our
understanding of both cognitively healthy and disordered brain
structure and function. The individual prediction and
characterization of patients with psychiatric problems have
shown promising results regarding recent breakthroughs in
machine-learning approaches. To evaluate patients with
The aim of the project is to develop an accurate and reliable deep learning-based
approach for the individual prediction and characterization of patients with psychiatric
disorders, specifically schizophrenia and bipolar disorder, using neuroimaging data. The
use of machine learning approaches can improve the diagnosis and treatment of
psychiatric disorders, as it allows for the identification of unique patterns and biomarkers
in brain structure and function that are difficult to detect through traditional clinical
methods.
Objective
[2020]
Skills Identified Gaps identified
● The ability of deep learning algorithms, ● While our project focuses on predicting the
specifically convolutional neural networks occurrence of schizophrenia and bipolar disorder
(CNNs), to accurately detect schizophrenia from in aged individuals using MRI images, the
structural MRI data. literature review only focuses on detecting
● The potential of deep learning algorithms to schizophrenia from structural MRI data
identify relevant structural features of ● Our project utilizes two deep learning models,
schizophrenia, such as the right temporal and VGG16 and Convolutional Neural Networks
parietal areas of the brain. (CNN), while the literature review only focuses
on the use of CNNs for detecting schizophrenia
from structural MRI data.
Literature Survey
● Knowledge of machine learning algorithms and ● Limited generalizability of findings due to small
their application in medical research, sample sizes and lack of diversity in study
particularly in predicting or diagnosing mental populations.
health conditions. ● Need for more rigorous evaluation and
● Ability to collect and analyze longitudinal data replication of findings to ensure that machine
from large, nationally representative surveys. learning models can be integrated into clinical
practice and improve patient outcomes.
● Lack of attention to ethical considerations, such
as issues related to data privacy and potential
bias in machine learning models.
Literature Survey
● This include knowledge of schizophrenia as a ● limited use of only one type of MRI modality,
brain disorder, familiarity with MRI and fMRI and the lack of external validation of the results.
imaging techniques ● Future studies should address these limitations
● understanding of machine learning algorithms, to ensure the reproducibility and generalizability
particularly convolutional neural networks, and of the findings.
expertise in data analysis and interpretation.
● The study also demonstrates skills in
experimental design and use of statistical
analysis.
Literature Survey
● Data analysis and preprocessing ● The study had a relatively small sample size,
● Machine learning algorithm selection and which may limit the generalizability of the
implementation findings to larger populations.
● Model evaluation using various metrics ● The data used in the study was self-reported,
● Interpretation of results and identification of which may introduce biases and inaccuracies in
potential implications for mental health the responses.
interventions and services
S
Y
S
T
E
M
A
R
C
H
I
T
E
C
T
U
R
E
System Model Description
➔ Data Collection
The dataset will be collected from various platforms like Center for Biomedical
Research Excellence (COBRE), Northwestern University Schizophrenia Data and
Software Tool (NUSDAST), domain experts, literature survey and other accessible
platforms.
➔ Data Augmentation
We enrich the training data to achieve a balanced distribution among the classes
because the image classes are severely unbalanced. To build a new augmented
dataset, we rotate and mirror the photos.
➔ Splitting of data
Data is standardized for the model's training and testing once it has been cleaned.
After the data is split, we train the algorithm on the training data set while putting
the test data aside. Based on the logic, methods, and values of the feature in the
training data, this training process will generate the training model. The primary
goal of feature extraction is to scale all values uniformly.
● Training set: A data scientist utilizes a training set to train a model and
determine its ideal parameters- parameters it must learn from the data.
● Test set: For an evaluation of the trained model's generalizability, a test set is
required. The latter refers to a model's capacity to find patterns in fresh,
unexplored data after being trained on training data. The inability for
generalization we described earlier, known as model overfitting, must be
avoided by using distinct subsets for training and testing.
➔ Classification
Deep Learning Technique is used to predict mental illness using
classification and ensemble techniques on brain MRI datasets. The
objective is to analyze the performance of these methods and find
accuracy, and identify the responsible/important features which play
a major role in prediction. So for comparing the accuracy and other
features we would be using two algorithms:
1. CNN(Convolutional Neural Network)
2. VGG16 model
Comparison
Once we have successfully implemented both the algorithms we will compare
the results using the accuracy and confusion matrix.
CNN VGG16
Within Deep Learning, a Convolutional VGG16 is a convolution neural
Neural Network or CNN is a type of net (CNN ) architecture which
artificial neural network, which is was used to win
widely used for image/object ILSVR(Imagenet) competition in
recognition and classification. Deep 2014. It is considered to be one of
Learning thus recognizes objects in an the excellent vision model
image by using a CNN. architecture
CNN
CNN(Convolution Neural Network)
The implementation of the CNN algorithm is done in following steps:
1. Classifier Training: A classifier is a function that takes features as input and generates a
class label prediction. Already Neuroimaging studies have applied various classifiers for
mental illness prediction, but there are dimensional issues which should be taken into
account before applying such classification algorithms.
2. Convolutional layer: This layer extracts the features from the image which is applied as
input. The neurons convolve the input image and produce a feature map in the output image
and this output image from this layer is fed as an input to the next convolutional layer.
3. Pooling layer: This layer is used to decrease the dimensions of the feature map still
maintaining all the important features. This layer is usually placed between two
convolutional layers.
4. Relu layer: A rectified linear unit (ReLU) is an activation function which interprets the positive
part of its argument.
5. Fully Connected layer: FLC means that each filter in the previous layer is connected to each
filter in the next layer. This is used to classify the input image based on the training dataset into
various classes.
It has 4 phases:
● Model construction: Model construction depends on machine learning algorithms. In this
projects case, it was Convolution Neural Networks.
● Model training: Here, the model is trained using training data and expected output for this
data. Once the model has been trained it is possible to carry out model testing.
● Model testing: During this phase a second set of data is loaded. This data set has never been
seen by the model and therefore it’s true accuracy will be verified.
● Model evaluation : After the model training is complete, the saved model can be used in the
real world. The name of this phase is model evaluation.
VGG16
Algorithm
The ImageNet dataset contains images of fixed size of 224*224 and have RGB channels. This model process the
input image and outputs the a vector of 1000 values. This vector represents the classification probability for the
corresponding class. To make sure these probabilities are added, we use softmax function.
Softmax Function uses the 5 most probable candidates into the vector.and our ground truth vector which then uses
the complex mathematical function to derive the results.
Architecture: The input to the network is an image of dimensions (48, 48, 3). The first two layers have 64 channels
of 3*3 filter size and same padding, followed by two layers of 256 filter size and padding size (3, 3). After that,
there are 2 sets of 3 convolution layers and a max pool layer. The stack of convolution and max-pooling layers
produces a (7, 7, 512) feature map, flatten it to a (1, 25088) vector, and pass it to a softmax layer to normalize the
classification vector. All the hidden layers use ReLU as their activation function, which is more computationally
efficient and decreases the likelihood of vanishing gradient problems.
Conclusion