Cancer Detection and Segmentation Project PPT Compressed
Cancer Detection and Segmentation Project PPT Compressed
Research Question
• Liver cancer is the leading cause of cancer deaths worldwide.
Manual examination of histopathological slides to detect cancer
is time consuming and even experienced pathologists differ
opinions 27% of the time.
• Automation of detection of cancer would
• help reduce time and work burden for pathologists,
• provide a consistent and maybe even more accurate
diagnosis.
• In recent years, scientists have proposed using advances in
Artificial Intelligence (AI) such as deep learning algorithms to
segment and detect cancerous regions in histopathological
Whole Slide Images (WSI) [4, 6, 9]
• Most methods use advanced Semantic Segmentation Neural Nets
(SSN) such as FCN, Unet, LinkNet, FPN etc. to segment tumor
regions. However, there has been no systematic evaluation of
their performance.
• Each model’s performance also depends on several parameters
whose impact has been inadequately studied. An example of a Whole Slide Image (WSI) of a
cancerous liver tissue.
• In this work, we aim to provide a thorough evaluation of major Where is the tumor?
semantic segmentation models for detection and diagnosis of
liver cancer in histopathological WSI. We want to analyze the
dependence of the cancer segmentation accuracy on :
• The choice of a particular SSN model,
• The size of local image patches,
• The input image resolution,
• The tumor type (viable or whole).
3
Approach
Generation of Patches
• A WSI is usually stored as an image pyramid allowing us to view it in different resolution
levels: 0, 1, 2 with level 0 as the highest resolution.
• While WSI have huge sizes (80k x 80k at level 0, 20k x 20k at level 1, 5k x 5k at level 2),
typical SSNs operate on images of much smaller sizes between 50-150 pixels and at a
single resolution. Therefore, WSI is divided into small local patches of the same size. The
patches may overlap.
• The choice of the patch size and the input resolution level may impact the accuracy of
the tumor segmentation algorithm.
Level 2
....
....
Level 1
Patch
Level 0
WSI
• Unet [8]
Predicted Masks
• LinkNet [1]
A dataset Patch
Backpropagation
of WSIs Sampler • DeepLabV3 [2]
• Feature Pyramid Network (FPN) [5]
Image
Updated
• To train the networks, we used Pytorch and the
SSN Weights open-source Semantic Segmentation Models
package [10]. Each model was trained for 1000
backpropagation iterations.
The algorithm for SSN training
Generation of Global Mask and Determination 6
of Cancer Status
• SSN produces for each pixel in each patch a number
from 0 to 1 that represents the probability of the
pixel being cancerous.
• The final segmentation mask is computed by taking
the average probability of each pixel over local
patches and rounding to 0 or 1.
• The final segmentation mask helps determine the
tumor burden, i.e. the amount of cancer present:
,%), $- ./,'0) !"#$%
!"#$% '"%()* =
,%), $- 1ℎ$0) !"#$%
• Note: We train two separate SSNs: one detects
viable tumors and one detects whole tumors.
An example of viable and whole tumor
• A viable tumor is the region containing the cell nest
and active cancerous parts of a tumor while a whole
tumor contains cell nests as well as dead tissue
surrounding the viable tumor and tumor capsules.
7
,%), $- $.)%0,4
3 4%)(/5!)(, 7%$"*(!%"!ℎ =
,%), $- "*/$*
Qualitative Results
These results were obtained with the PAN network at level 2 and patch size 512.
More example results are provided in the attached research paper
9
Quantitative Results
Average Jaccard Index between predicted tumor mask and groundtruth for viable
tumors for different combinations of SSN model, patch size, and resolution levels
Average Jaccard Index between predicted tumor mask and groundtruth for whole Tumor Burden Estimation accuracy
tumors for different combinations of SSN model, patch size, and resolution levels
10
Interpretation of Results
• Most of the tested SSNs were able to detect and segment viable tumor regions with a
relatively high accuracy. In the majority of the results, the predicted viable tumor masks
were mostly similar to the groundtruth except for scattered small size details.
• Our best average Jaccard Index value is 0.73 obtained with FPN at patch size 512 and
level 2. It is lower than the top scores reported at the PAIP challenge which are between
0.75-0.79. The method of [6], the only current published work PAIP 2019, combines the
result from multiple models through an ensemble approach. On the other hand, the
objective of our work is to investigate the behavior of individual models.
• As expected, the detection results of whole tumors are generally less accurate than that
of viable tumors due to the difficulty in detecting peripheral tissues which sometimes
may be visually similar to healthy tissue.
• DeepLabV3, FPN, and PAN outperformed the other three SSN models. While FPN
performed the best for segmentation of viable tumors, DeepLabV3 and PAN performed
well for both viable and whole tumors, and therefore got the best tumor burden
estimation accuracy. PAN appears to exhibit stable good performance across levels and
patch sizes.
• The patch size 512 exhibits the best performance, suggesting that a large patch better
captures global information which helps distinguish a tumor from other parts of the
liver.
• To our surprise, the pyramid levels 1 and 0 tend to have worse results than level 2.
Again, this suggests that the global information captured by level 2 could play a more
important role in detecting tumors than local details provided by levels 1 and 0. Another
possible reason could be that models at higher resolutions need to be trained for a
longer amount of time. We will further investigate this issue in future work.
11
Conclusion
• We have implemented 6 major semantic segmentation models, trained and
quantitatively evaluated their performance on the PAIP dataset for various
parameter settings.
• Results show that SSN are able to detect and segment tumors in WSI with
relatively high accuracy. The accuracy of tumor burden estimates almost
reaches 90%.
• PAN and DeepLabV3 came out as the best choices among the tested
models
• A large patch size and low input resolution tend to lead to better
performance, although more experiments are needed to confirm this
observation.
• Whole tumors are more difficult to segment than viable tumors.
• Our work provide a better understanding of the behavior of deep learning
algorithms for tumor segmentation. It helps researchers quickly set up and fine
tune the algorithms for their work.
Acknowledgements:
I would like to thank Professor Yuankai Huo from the department of Electrical Engineering
and Computer Science at the Vanderbilt University for giving his valuable time and
feedback.
12
References
1. A. Chaurasia and E. Culurciello. “LinkNet: Exploiting encoder representations for efficientsemantic
segmentation”. In: (2017), pp. 1–4.
2. L. Chen et al. “DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous
Convolution, and Fully Connected CRFs”. In: IEEE Transactions on Pattern Analysis and Machine
Intelligence 40.4 (2018), pp. 834–848.
3. Chung NC et al. Jaccard/Tanimoto similarity test and estimation methods for biological presence-
absence data. BMC Bioinformatics. 2019 Dec 24;20(Suppl 15):644
4. Baris Gecer et al. “Detection and classification of cancer in whole slide breast histopathology
images using deep convolutional networks”. In: Pattern Recognition 84 (2018), pp. 345–356.
5. Jonathan Hui. Understanding Feature Pyramid Networks for object detection (FPN). 2020.
6. Mahendra Khened et al. A Generalized Deep Learning Framework for Whole-Slide Image
Segmentation and Analysis. 2020. arXiv:2001.00258 [eess.IV].
7. Hanchao Li et al. “Pyramid Attention Network for Semantic Segmentation”. In:
CoRRabs/1805.10180(2018). arXiv:1805.10180.
8. Olaf Ronneberger et al. “U-net: Convolutional networks for biomedical image segmentation”. In:
International Conference on Medical image computing and computer-assisted intervention.
Springer. 2015, pp. 234–241.
9. Mary Shapcott et al. “Deep Learning With Sampling in Colon Cancer Histology”. In: Frontiers in
Bioengineering and Biotechnology 7 (2019), p. 52.
10. Pavel Yakubovskiy. segmentation-models-pytorch. 2020.
url:https://pypi.org/project/segmentation-models-pytorch/.
11. Hengshuang Zhao et al. “Pyramid scene parsing network”. In: Proceedings of the IEEE conference
on computer vision and pattern recognition. 2017, pp. 2881–2890.