Cannizzaro, Renato
A multi-centre polyp detection and segmentation dataset for generalisability assessment
Ali, Sharib, Jha, Debesh, Ghatwary, Noha, Realdon, Stefano, Cannizzaro, Renato, Salem, Osama E., Lamarque, Dominique, Daul, Christian, Riegler, Michael A., Anonsen, Kim V., Petlund, Andreas, Halvorsen, Pål, Rittscher, Jens, de Lange, Thomas, East, James E.
Polyps in the colon are widely known cancer precursors identified by colonoscopy. Whilst most polyps are benign, the polyp's number, size and surface structure are linked to the risk of colon cancer. Several methods have been developed to automate polyp detection and segmentation. However, the main issue is that they are not tested rigorously on a large multicentre purpose-built dataset, one reason being the lack of a comprehensive public dataset. As a result, the developed methods may not generalise to different population datasets. To this extent, we have curated a dataset from six unique centres incorporating more than 300 patients. The dataset includes both single frame and sequence data with 3762 annotated polyp labels with precise delineation of polyp boundaries verified by six senior gastroenterologists. To our knowledge, this is the most comprehensive detection and pixel-level segmentation dataset (referred to as \textit{PolypGen}) curated by a team of computational scientists and expert gastroenterologists. The paper provides insight into data construction and annotation strategies, quality assurance, and technical validation. Our dataset can be downloaded from \url{ https://doi.org/10.7303/syn26376615}.
A translational pathway of deep learning methods in GastroIntestinal Endoscopy
Ali, Sharib, Dmitrieva, Mariia, Ghatwary, Noha, Bano, Sophia, Polat, Gorkem, Temizel, Alptekin, Krenzer, Adrian, Hekalo, Amar, Guo, Yun Bo, Matuszewski, Bogdan, Gridach, Mourad, Voiculescu, Irina, Yoganand, Vishnusai, Chavan, Arnav, Raj, Aryan, Nguyen, Nhan T., Tran, Dat Q., Huynh, Le Duy, Boutry, Nicolas, Rezvy, Shahadate, Chen, Haijian, Choi, Yoon Ho, Subramanian, Anand, Balasubramanian, Velmurugan, Gao, Xiaohong W., Hu, Hongyu, Liao, Yusheng, Stoyanov, Danail, Daul, Christian, Realdon, Stefano, Cannizzaro, Renato, Lamarque, Dominique, Tran-Nguyen, Terry, Bailey, Adam, Braden, Barbara, East, James, Rittscher, Jens
The Endoscopy Computer Vision Challenge (EndoCV) is a crowd-sourcing initiative to address eminent problems in developing reliable computer aided detection and diagnosis endoscopy systems and suggest a pathway for clinical translation of technologies. Whilst endoscopy is a widely used diagnostic and treatment tool for hollow-organs, there are several core challenges often faced by endoscopists, mainly: 1) presence of multi-class artefacts that hinder their visual interpretation, and 2) difficulty in identifying subtle precancerous precursors and cancer abnormalities. Artefacts often affect the robustness of deep learning methods applied to the gastrointestinal tract organs as they can be confused with tissue of interest. EndoCV2020 challenges are designed to address research questions in these remits. In this paper, we present a summary of methods developed by the top 17 teams and provide an objective comparison of state-of-the-art methods and methods designed by the participants for two sub-challenges: i) artefact detection and segmentation (EAD2020), and ii) disease detection and segmentation (EDD2020). Multi-center, multi-organ, multi-class, and multi-modal clinical endoscopy datasets were compiled for both EAD2020 and EDD2020 sub-challenges. An out-of-sample generalisation ability of detection algorithms was also evaluated. Whilst most teams focused on accuracy improvements, only a few methods hold credibility for clinical usability. The best performing teams provided solutions to tackle class imbalance, and variabilities in size, origin, modality and occurrences by exploring data augmentation, data fusion, and optimal class thresholding techniques.