Project Work Papers
Project Work Papers
Use Case
The paper addresses the challenge of classifying white blood cells (WBCs) under varying
lighting conditions. It aims to improve the generalisation ability of convolutional neural
networks (CNNs) for WBC classification, ensuring accurate results despite changes in
illumination intensity. This is particularly important in medical diagnostics, where variations in
lighting can affect the accuracy of automated cell classification systems.
The solution proposed involves the introduction of a Learnable Illumination Intensity Invariant
Layer (LII-Layer) within the CNN architecture. This LII-Layer is designed to create a
color-invariant representation of the input images. The key idea is to adjust the scale of
convolution dynamically during training, inspired by classical color invariant transformations.
This layer learns the optimal scale for edge detection, making the CNN more robust to
changes in illumination.
Steps:
● Frameworks/Libraries: PyTorch
● Model Architecture: CNNs, specifically ResNet34 and ResNet50 with the integrated
LII-Layer
● Training: Utilized a cosine learning rate decay strategy and trained the models over
100 epochs
● Data Preprocessing: Images were resized to 224x224 pixels, with random
horizontal and vertical flips applied during preprocessing
Implementation
1. Data Preparation:
○ Images were resized to 224x224 pixels.
○ Random horizontal and vertical flips were applied to augment the dataset.
2. Model Training:
○ CNN models (ResNet34 and ResNet50) were trained with and without the
LII-Layer.
○Training was conducted using a cosine learning rate decay strategy over 100
epochs.
3. Evaluation:
○ The trained models were tested on datasets with different lighting conditions.
○ The performance metrics were analyzed to assess the generalization
capability of the models.
4. Analysis:
○ Performance and training time of models with the LII-Layer were compared
against models without it.
○ The impact of the LII-Layer on the generalization ability of the models was
evaluated.
● Challenges:
○ The integration of the LII-Layer increased computational complexity and
training time.
● Improvements:
○ Optimization: Optimize the LII-Layer to reduce additional training time and
computational load.
○ Alternative Methods: Explore other methods for achieving illumination
invariance to compare effectiveness.
○ Augmentation: Integrate additional data augmentation techniques during
training to further enhance robustness.
○ Hybrid Approaches: Combine the LII-Layer with other invariant methods to
improve overall performance.
Use Case
This paper aims to accurately classify white blood cells (WBCs) for identifying acute myeloid
leukemia (AML) from blood smear microscopic images. The goal is to extract relevant
features from the images to improve WBC classification accuracy, aiding in the diagnosis of
AML. Accurate classification of WBCs is crucial for early detection and treatment planning
for leukemia patients.
The proposed approach combines a novel CMYK-moment localization technique with deep
learning. The methodology involves:
1. CMYK-Moment Localization: Transforming the RGB images into the CMYK color
space and using image moments to extract regions of interest (ROIs).
2. Deep Learning: Feeding these localized features into a CNN for feature extraction.
3. Feature Fusion: Combining pointwise and localized features within the CNN to
enable detailed and generalized feature extraction without losing pixel information.
Steps:
Implementation
1. Data Preparation:
○ Convert images to the CMYK color space.
○ Extract regions of interest using image moments.
2. Feature Extraction:
○ Use a hybrid model that combines CMYK-moment localization with a CNN for
extracting features.
○ Implement feature fusion within the CNN architecture.
3. Model Training:
○ Train the CNN on the extracted features.
○ Utilize labeled datasets for supervised learning.
4. Evaluation:
○ Test the trained model on the AML_Cytomorphology_LMU dataset and the
secondary dataset.
○ Evaluate the classification accuracy and generalization performance.
● Challenges:
○ Potential overfitting due to high model complexity.
○ High computational requirements for feature extraction and processing.
● Improvements:
○ Data Augmentation: Apply advanced data augmentation techniques to
enhance model generalization.
○ Model Refinement: Integrate additional layers or modules within the CNN to
refine feature extraction.
○ Dataset Expansion: Use larger and more diverse datasets to improve model
robustness and reduce overfitting.
○ Hybrid Approaches: Combine CMYK-moment localization with other feature
extraction techniques to improve overall performance.
White Blood Cell (WBC) Segmentation and Classification
Use Case Description
The proposed method utilizes a combination of deep learning models and algorithms to
achieve accurate WBC segmentation and classification. The framework includes the use of
PSPNet for segmentation, followed by MobilenetV3 and Artificial Gravitational Cuckoo
Search (AGCS) for feature extraction and selection. Finally, the ShufflenetV2 model is
employed to classify the WBCs into five categories.
Implementation Details
1. Data Preprocessing:
● Resizing input images to 256x256 pixels.
● Applying Contrast Limited Adaptive Histogram Equalization (CLAHE)
to enhance contrast.
● Using image augmentation techniques to reduce overfitting and data
imbalance.
2. Segmentation:
● Utilizing PSPNet with modifications to segment WBC nuclei from the
images.
● Incorporating Pyramid Pooling Modules (PPMs) and attention gates to
improve segmentation accuracy.
3. Feature Extraction and Selection:
● Extracting significant features using MobilenetV3.
● Selecting prominent features using the AGCS algorithm.
4. Classification:
Implementing ShufflenetV2 to classify the WBCs into monocytes,
basophils, eosinophils, neutrophils, and lymphocytes.
Achieving high accuracy on the BCCD and Raabin-WBC datasets.
Challenges and Further Improvements
Challenges:
○ Variability in WBC shapes, sizes, and colors makes classification
challenging.
Further Improvements:
○ Incorporating more robust feature extraction techniques to handle diverse
cell appearances.
○ Optimizing the network for faster inference times without sacrificing
accuracy.
○ Expanding the dataset with more labeled images to improve model
robustness and generalization capabilities .
A Review on Traditional Machine Learning and Deep Learning Models for WBCs
Classification in Blood Smear Images
The research paper "A Review on Traditional Machine Learning and Deep Learning Models
for WBCs Classification in Blood Smear Images" discusses the use case of leukocyte
classification in blood smear images using machine learning techniques. The approach to the
solution involves the application of various traditional machine learning and deep learning
models to accurately classify different types of white blood cells.
The tech stack used in the implementation of these models may include programming
languages such as Python, libraries like TensorFlow, Keras, Scikit-learn, and image
processing tools like OpenCV. The implementation process involves preprocessing the blood
smear images, extracting relevant features, training the machine learning models, and
evaluating their performance using metrics like accuracy, precision, recall, and F1-score.
Some further improvements that could be made based on the challenges faced in the
publication date include:
Model Interpretability: Enhancing the interpretability of the models to provide insights into
the decision-making process for better trust and acceptance in real-world applications.
Integration with Clinical Systems: Integrating the developed models with existing clinical
systems to facilitate seamless adoption by healthcare professionals and improve the
efficiency of diagnosis.
By addressing these improvements and challenges, the research can further enhance the
accuracy, reliability, and applicability of machine learning models for WBCs classification in
blood smear images, ultimately contributing to advancements in medical image analysis and
healthcare diagnostics.
The paper focuses on the detection of White Blood Cells (WBC) using integrated design
based on dual thresholding and features optimization. The primary goal is to identify different
types of white blood cells accurately, which is critical for diagnosing various medical
conditions, including leukemia.
Dual Thresholding: This technique is used to segment the white blood cells from the
microscopic images.
Deep Learning: Convolutional Neural Networks (CNN) are used to extract deep features,
which are then optimized using the Non-dominated Sorting Genetic Algorithm (NSGA).
Classification: The optimized feature vector is passed to different classifiers to achieve high
accuracy.
● MobileNetV2
● ShuffleNet
● AlexNet
● ResNet-18
Implementation:
Classification: Different classifiers are used to classify the optimized feature vectors.
Further Improvements:
Enhanced Feature Extraction: Incorporate additional features or use more advanced CNN
architectures to improve accuracy.
Hybrid Models: Combine different models to leverage their strengths and improve
performance.
Real-time Implementation: Develop real-time systems for WBC detection to assist in clinical
settings.
Automation: Automate the entire pipeline for faster and more efficient processing.
Expand Dataset: Use larger and more diverse datasets to train the models for better
generalization.
Challenge: Existing feature extraction methods might miss subtle details crucial for accurate
classification.
Improvement: Integrate multi-modal data (e.g., clinical data, patient history) alongside image
data to improve the model's context-awareness and accuracy.
4. Explainable AI (XAI):
Challenge: Deep learning models often act as black boxes, making it hard to interpret their
decisions.
5. Real-Time Processing:
Improvement: Optimize the model for real-time processing using techniques like model
pruning, quantization, and deploying on edge devices with high computational power (e.g.,
NVIDIA Jetson).
6. Robustness to Variability:
Challenge: Variability in image quality and staining techniques can affect the model's
performance.
Improvement: Train the model on a more diverse dataset that includes images with different
staining techniques and quality levels to improve robustness.
Challenge: Existing optimization algorithms may not fully explore the feature space.
8. Continuous Learning:
Challenge: Static models may become outdated as new data and techniques emerge.
Challenge: Lack of extensive clinical validation can limit the model's adoption.
Improvement: Establish a feedback loop with clinicians to continuously validate and refine
the model, ensuring it meets clinical standards and requirements.
Improvement: Develop user-friendly interfaces and visualization tools that allow clinicians to
interact with the model easily and interpret its results effectively.
Data Privacy and Security: Implement robust data encryption and access control mechanisms
to ensure patient data privacy and compliance with regulations like GDPR and HIPAA.
BRAIN TUMOUR
TransDoubleU-Net: Dual Scale Swin Transformer With Dual Level Decoder for 3D
Multimodal Brain Tumor Segmentation
Use Case:
Approach:
Tech Stack:
Implementation:
● Training: Used Adam optimizer with an initial learning rate of 0.0003 for 50
epochs.
● Data Augmentation: Random flip, rotation, scaling, and intensity shift. ●
Evaluation: Dice score (DSC) and Hausdorff distance (HD) metrics.
Results:
Use Case:
Approach:
Tech Stack:
● Programming Language: Python
● Framework: PyTorch
● Datasets: BraTS2019 and BraTS2020
Implementation:
● Training: Involves optimizing the model parameters using advanced deep learning
techniques.
● Data Augmentation: Applies various transformations to increase the diversity of
training data.
● Evaluation: Uses standard segmentation metrics like DSC and HD.
Results:
● Challenges: High computational cost and the need for large-scale annotated
datasets.
● Improvements: Reducing model complexity, enhancing training efficiency, and
exploring semi-supervised learning techniques to utilize unlabeled data.
1. Attention Mechanisms:
○ Enhance the self-attention mechanism to better capture long-range
dependencies.
○ Experiment with different attention mechanisms to improve segmentation
accuracy.
2. Data Augmentation:
○ Implement more sophisticated data augmentation techniques to improve
model robustness.
3. Computational Efficiency:
○ Optimize the model architecture to reduce computational costs and improve
training speed.
○ Explore the use of efficient transformers and lightweight models.
4. Training Techniques:
○ Investigate semi-supervised and unsupervised learning methods to
leverage unlabeled data.
○ Utilize advanced optimization algorithms to improve convergence and
performance.
A Hybrid Deep Learning Technique for Personality Trait Classification From Text
Use Case
The paper describes a system for personality trait classification based on textual data. This
system is intended to analyze written text to determine personality traits according to the Big
Five Model. The primary application is in domains where understanding written
communication can provide insights into individuals' personalities, such as social media
analysis, human resources, and customer service.
Implementation Details
● Data Collection: Text samples are collected from various sources.
● Preprocessing: Includes tokenization, stop-word removal, and stemming.
● Feature Extraction: Utilizes word embeddings and other NLP features.
● Hybrid Model: Combines CNNs and RNNs to capture both local and sequential
patterns in the text.
Technology Stack
● Libraries: NLTK, SpaCy for text preprocessing; TensorFlow, Keras for model
training
● Machine Learning: CNNs, RNNs for feature extraction and classification
● Word Embeddings: Pre-trained embeddings like GloVe or Word2Vec
Challenges:
Improvements:
● Incorporate transformer-based models (like BERT) for better contextual
understanding.
● Use advanced text augmentation techniques to improve model robustness.
● Develop interpretability tools to better understand how the model makes its
predictions.
Use Case
The paper "A Method for MBTI Classification Based on Impact of Class Components"
presents a method for classifying Myers-Briggs Type Indicator (MBTI) types from textual
data obtained from a social network. The primary use case is to improve the accuracy and
reliability of MBTI classification by addressing the imbalance and overlap in MBTI class
components.
Approach to Solution
The approach involves the use of Convolutional Neural Networks (CNN) and Long
Short-Term Memory (LSTM) networks with a custom loss function called CECI (Class
Encoding Component Impact). This loss function adjusts the weight of each MBTI class
component during training, aiming to mitigate the impact of class imbalance and improve
classification performance.
Tech Stack
1. Data Preprocessing: Textual data from the social network is preprocessed, including
tokenization, padding, and conversion to numerical format suitable for neural network
input.
2. Model Architecture:
○ LSTM Model: Includes an embedding layer, LSTM layers, and dense layers
for classification.
○ CNN Model: Comprises an embedding layer, convolutional layers, pooling
layers, and dense layers for classification.
3. Custom Loss Function (CECI): The CECI loss function is implemented to weigh
the impact of different MBTI class components during training.
4. Training and Evaluation: Both models are trained using the preprocessed data, and
their performance is evaluated using metrics such as F1-score, accuracy, precision,
recall, and confusion matrix.
Challenges:
Further Improvements:
These suggestions aim to address the limitations identified in the research and provide new
directions for improving the MBTI classification model(A_Method_for_MBTI_Class…).
A Quantitative Assessment of the Impact of Homogeneity in Personality Traits
on Software Quality and Team Productivity
The research paper discusses a case study involving two software projects: MyShop
project for students and Online Job Portal project for professionals. Here is a summary
of the use case, approach to the solution, tech stack used, implementation details,
potential improvements, and challenges faced:
Use Case:
● Both projects were developed using PHP programming language for backend
and HTML/CSS/JS for front end development.
Implementation Details:
● Training manuals and tools were provided to the teams for different phases of the
projects.
● Students were given two weeks for analysis and design, four weeks for
implementation, and two working days for testing.
● Professionals were given specific deadlines for each phase of the project.
● Tools like 'Visio' for creating models and 'TimeKeeper' for time tracking were
utilized.
Potential Improvements:
Challenges Faced:
● Ensuring consistent team engagement and participation throughout the project
phases.
● Managing time effectively to meet project deadlines.
● Addressing potential conflicts arising from diverse personality traits within the
teams.
Addressing challenges:
CERVICAL CANCER
Approach to Solution: The authors propose a hybrid approach that combines feature
selection using the Least Absolute Shrinkage and Selection Operator (LASSO) with
classification using an Ensemble Learning Support Vector Machine (EL-SVM). The
methodology includes:
1. LASSO for Feature Selection: LASSO is employed to select the most relevant
features from the biopsy images, reducing dimensionality and improving model
interpretability.
2. EL-SVM for Classification: An ensemble of SVM classifiers is used to enhance
the robustness and accuracy of the classification. This ensemble approach leverages
multiple SVMs trained on different subsets of the data to produce a consensus
classification.
● Image preprocessing and feature extraction using image processing libraries such
as OpenCV.
● Feature selection with LASSO using statistical computing environments like R or
Python's scikit-learn.
● Classification with EL-SVM, implemented using machine learning frameworks like
scikit-learn or MATLAB.
Implementation:
CRIME
Use Case
The research paper describes a use case focused on crime event extraction for low-resource
languages. The primary objective is to develop a system that can efficiently extract
crime-related information from texts in languages that lack substantial linguistic resources.
This system aims to assist law enforcement agencies and other relevant organizations in
analyzing and understanding crime patterns from various textual data sources.
The approach taken in the paper involves a parallel corpus-based method for crime event
extraction. This method leverages parallel corpora, which are collections of texts and their
translations in multiple languages. By using a high-resource language (e.g., English) as a
bridge, the researchers can transfer knowledge to the low-resource language. The main steps
in the approach include:
Tech Stack
● Natural Language Processing (NLP) Tools: Tools and libraries for text processing
and annotation, such as NLTK, SpaCy, or custom annotation tools.
● Machine Learning Frameworks: Frameworks like TensorFlow or PyTorch for
training the extraction models.
● Parallel Corpora: Datasets containing parallel texts in high-resource and
low-resource languages.
● Annotation Tools: Software for manual annotation of the corpus, possibly including
tools like BRAT or Prodigy.
● Evaluation Metrics: Standard metrics for evaluating the performance of NLP
models, such as precision, recall, and F1-score.
Implementation
Further Improvements
To address challenges and improve the approach, the following enhancements could be
considered:
1. Expansion of Parallel Corpora: Increasing the size and diversity of the parallel
corpora to cover more crime-related contexts and variations in language usage.
2. Advanced Annotation Techniques: Implementing semi-automatic or crowd-sourced
annotation methods to reduce the manual effort and improve annotation consistency.
3. Multilingual Models: Leveraging multilingual models and transfer learning
techniques to better transfer knowledge from high-resource to low-resource
languages.
4. Integration of Additional Data Sources: Incorporating other data sources such as
news articles, police reports, and social media to enrich the corpus and improve model
robustness.
5. Continuous Learning: Implementing continuous learning frameworks where the
models can be updated and refined with new data and annotations over time.
6. Addressing Bias and Fairness: Ensuring that the models are fair and unbiased,
particularly when dealing with sensitive topics like crime, by incorporating
fairness-aware algorithms and evaluation metrics.