Paper 14
Paper 14
Paper 14
Segmentation
Awadelrahman M. A. Ahmed
University of Oslo, Norway; [email protected]
3.2 Results
To quantify the overlap percentage between the ground truth mask
and our generated masks, we report both Jaccard index the Dice
similarity coefficient (DSC). We also report the per pixel recall,
precision, accuracy and F2 (giving more weight to recall). Results
on the test set are in Table 2. However it is difficult to comment on
those results objectively because of the uniqueness of the dataset
[4], i.e. no previous publications to compare with, but in general
the higher recall than precision reflects the model accountability
for false-negatives which is desirable for this application. The test
throughput is 16 frames/sec on a 2×14-core Intel/128Gib machine.
We report the model outputs for some samples in Figure 3. Even
Figure 1: Model block diagram
though we do not have the access to the ground truth of the test
set, we may observe that the model incorrectly identified the polyp
Generator Network location of the bottom two samples; whereas it did far better in
Conv1 𝑓 *1, LeakyReLU locating the polyps area of other samples (top and middle rows).
Conv2 𝑓 *2, LeakyReLU We do not have a clear explanation for that, but we speculate that
Conv3 𝑓 *4, LeakyReLU the small receptive field of the convolution layers could be a reason.
Conv4, 5, 6, 7, 8 𝑓 *8, LeakyReLU In other words, the convolution layers pay attention to the close by
DeConv1, 2, 3 ,4 𝑓 *8, ReLU, Dropout(0.5) area to each pixel and if this area is rich enough with features (e.g.
DeConv5 𝑓 *4, ReLU, Dropout(0.5) has sharp edges) the polyp can be more distinguishable. This why
DeConv6 𝑓 *2, ReLU, Dropout(0.5) incorporating attention mechanism [7] might help the model to
DeConv7 𝑓 *1, ReLU, Dropout(0.5) attend to fine details and larger ranges; hence we suggest studying
DeConv8 1 , Tanh it as an extended model.
Discriminator Network Jaccard DSC Recall Precision Accuracy F2
Conv1 𝑓 *1, LeakyReLU 0.4382 0.562 0.697 0.556 0.881 0.611
Conv2 𝑓 *2, LeakyReLU
Table 2: Test results as reported by the challenge organizers
Conv3 𝑓 *4, LeakyReLU
Conv4 1 , Sigmoid
𝑓 : number of filters=64
Table 1: Model details
REFERENCES
[1] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley,
Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial
nets. In Advances in neural information processing systems. 2672–2680.
[2] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A Efros. 2017. Image-to-
image translation with conditional adversarial networks. In Proceedings of the
IEEE conference on computer vision and pattern recognition. 1125–1134.
[3] Debesh Jha, Steven A. Hicks, Krister Emanuelsen, Håvard Johansen, Dag Johansen,
Thomas de Lange, Michael A. Riegler, and Pål Halvorsen. 14-15 December 2020.
Medico Multimedia Task at MediaEval 2020:Automatic Polyp Segmentation. In
Proc. of the MediaEval 2020 Workshop, Online.
[4] Debesh Jha, Pia H Smedsrud, Michael A Riegler, Pål Halvorsen, Thomas de Lange,
Dag Johansen, and Håvard D Johansen. 2020. Kvasir-SEG: A segmented polyp
dataset. In Proc. of International Conference on Multimedia Modeling. 451–462.
[5] Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang,
Zachary DeVito, Zeming Lin, Alban Desmaison, Luca Antiga, and Adam Lerer.
2017. Automatic differentiation in PyTorch. (2017).
[6] Guido Van Rossum and Fred L. Drake. 2009. Python 3 Reference Manual. CreateS-
pace, Scotts Valley, CA.
[7] Han Zhang, Ian Goodfellow, Dimitris Metaxas, and Augustus Odena. 2019. Self-
attention generative adversarial networks. In International conference on machine
learning. PMLR, 7354–7363.