Al-Maskari, Rami
SELMA3D challenge: Self-supervised learning for 3D light-sheet microscopy image segmentation
Chen, Ying, Al-Maskari, Rami, Horvath, Izabela, Ali, Mayar, Hoher, Luciano, Yang, Kaiyuan, Lin, Zengming, Zhai, Zhiwei, Shen, Mengzhe, Xun, Dejin, Wang, Yi, Xu, Tony, Goubran, Maged, Wu, Yunheng, Mori, Kensaku, Paetzold, Johannes C., Erturk, Ali
Recent innovations in light sheet microscopy, paired with developments in tissue clearing techniques, enable the 3D imaging of large mammalian tissues with cellular resolution. Combined with the progress in large-scale data analysis, driven by deep learning, these innovations empower researchers to rapidly investigate the morphological and functional properties of diverse biological samples. Segmentation, a crucial preliminary step in the analysis process, can be automated using domain-specific deep learning models with expert-level performance. However, these models exhibit high sensitivity to domain shifts, leading to a significant drop in accuracy when applied to data outside their training distribution. To address this limitation, and inspired by the recent success of self-supervised learning in training generalizable models, we organized the SELMA3D Challenge during the MICCAI 2024 conference. SELMA3D provides a vast collection of light-sheet images from cleared mice and human brains, comprising 35 large 3D images-each with over 1000^3 voxels-and 315 annotated small patches for finetuning, preliminary testing and final testing. The dataset encompasses diverse biological structures, including vessel-like and spot-like structures. Five teams participated in all phases of the challenge, and their proposed methods are reviewed in this paper. Quantitative and qualitative results from most participating teams demonstrate that self-supervised learning on large datasets improves segmentation model performance and generalization. We will continue to support and extend SELMA3D as an inaugural MICCAI challenge focused on self-supervised learning for 3D microscopy image segmentation.
TopCoW: Benchmarking Topology-Aware Anatomical Segmentation of the Circle of Willis (CoW) for CTA and MRA
Yang, Kaiyuan, Musio, Fabio, Ma, Yihui, Juchler, Norman, Paetzold, Johannes C., Al-Maskari, Rami, Höher, Luciano, Li, Hongwei Bran, Hamamci, Ibrahim Ethem, Sekuboyina, Anjany, Shit, Suprosanna, Huang, Houjing, Waldmannstetter, Diana, Kofler, Florian, Navarro, Fernando, Menten, Martin, Ezhov, Ivan, Rueckert, Daniel, Vos, Iris, Ruigrok, Ynte, Velthuis, Birgitta, Kuijf, Hugo, Hämmerli, Julien, Wurster, Catherine, Bijlenga, Philippe, Westphal, Laura, Bisschop, Jeroen, Colombo, Elisa, Baazaoui, Hakim, Makmur, Andrew, Hallinan, James, Wiestler, Bene, Kirschke, Jan S., Wiest, Roland, Montagnon, Emmanuel, Letourneau-Guillon, Laurent, Galdran, Adrian, Galati, Francesco, Falcetta, Daniele, Zuluaga, Maria A., Lin, Chaolong, Zhao, Haoran, Zhang, Zehan, Ra, Sinyoung, Hwang, Jongyun, Park, Hyunjin, Chen, Junqiang, Wodzinski, Marek, Müller, Henning, Shi, Pengcheng, Liu, Wei, Ma, Ting, Yalçin, Cansu, Hamadache, Rachika E., Salvi, Joaquim, Llado, Xavier, Estrada, Uma Maria Lal-Trehan, Abramova, Valeriia, Giancardo, Luca, Oliver, Arnau, Liu, Jialu, Huang, Haibin, Cui, Yue, Lin, Zehang, Liu, Yusheng, Zhu, Shunzhi, Patel, Tatsat R., Tutino, Vincent M., Orouskhani, Maysam, Wang, Huayu, Mossa-Basha, Mahmud, Zhu, Chengcheng, Rokuss, Maximilian R., Kirchhoff, Yannick, Disch, Nico, Holzschuh, Julius, Isensee, Fabian, Maier-Hein, Klaus, Sato, Yuki, Hirsch, Sven, Wegener, Susanne, Menze, Bjoern
The Circle of Willis (CoW) is an important network of arteries connecting major circulations of the brain. Its vascular architecture is believed to affect the risk, severity, and clinical outcome of serious neuro-vascular diseases. However, characterizing the highly variable CoW anatomy is still a manual and time-consuming expert task. The CoW is usually imaged by two angiographic imaging modalities, magnetic resonance angiography (MRA) and computed tomography angiography (CTA), but there exist limited public datasets with annotations on CoW anatomy, especially for CTA. Therefore we organized the TopCoW Challenge in 2023 with the release of an annotated CoW dataset. The TopCoW dataset was the first public dataset with voxel-level annotations for thirteen possible CoW vessel components, enabled by virtual-reality (VR) technology. It was also the first large dataset with paired MRA and CTA from the same patients. TopCoW challenge formalized the CoW characterization problem as a multiclass anatomical segmentation task with an emphasis on topological metrics. We invited submissions worldwide for the CoW segmentation task, which attracted over 140 registered participants from four continents. The top performing teams managed to segment many CoW components to Dice scores around 90%, but with lower scores for communicating arteries and rare variants. There were also topological mistakes for predictions with high Dice scores. Additional topological analysis revealed further areas for improvement in detecting certain CoW components and matching CoW variant topology accurately. TopCoW represented a first attempt at benchmarking the CoW anatomical segmentation task for MRA and CTA, both morphologically and topologically.
Panoptica -- instance-wise evaluation of 3D semantic and instance segmentation maps
Kofler, Florian, Möller, Hendrik, Buchner, Josef A., de la Rosa, Ezequiel, Ezhov, Ivan, Rosier, Marcel, Mekki, Isra, Shit, Suprosanna, Negwer, Moritz, Al-Maskari, Rami, Ertürk, Ali, Vinayahalingam, Shankeeth, Isensee, Fabian, Pati, Sarthak, Rueckert, Daniel, Kirschke, Jan S., Ehrlich, Stefan K., Reinke, Annika, Menze, Bjoern, Wiestler, Benedikt, Piraud, Marie
This paper introduces panoptica, a versatile and performance-optimized package designed for computing instance-wise segmentation quality metrics from 2D and 3D segmentation maps. panoptica addresses the limitations of existing metrics and provides a modular framework that complements the original intersection over union-based panoptic quality with other metrics, such as the distance metric Average Symmetric Surface Distance. The package is open-source, implemented in Python, and accompanied by comprehensive documentation and tutorials. panoptica employs a three-step metrics computation process to cover diverse use cases. The efficacy of panoptica is demonstrated on various real-world biomedical datasets, where an instance-wise evaluation is instrumental for an accurate representation of the underlying clinical task. Overall, we envision panoptica as a valuable tool facilitating in-depth evaluation of segmentation methods.
blob loss: instance imbalance aware loss functions for semantic segmentation
Kofler, Florian, Shit, Suprosanna, Ezhov, Ivan, Fidon, Lucas, Horvath, Izabela, Al-Maskari, Rami, Li, Hongwei, Bhatia, Harsharan, Loehr, Timo, Piraud, Marie, Erturk, Ali, Kirschke, Jan, Peeken, Jan C., Vercauteren, Tom, Zimmer, Claus, Wiestler, Benedikt, Menze, Bjoern
Deep convolutional neural networks (CNN) have proven to be remarkably effective in semantic segmentation tasks. Most popular loss functions were introduced targeting improved volumetric scores, such as the Dice coefficient (DSC). By design, DSC can tackle class imbalance, however, it does not recognize instance imbalance within a class. As a result, a large foreground instance can dominate minor instances and still produce a satisfactory DSC. Nevertheless, detecting tiny instances is crucial for many applications, such as disease monitoring. For example, it is imperative to locate and surveil small-scale lesions in the follow-up of multiple sclerosis patients. We propose a novel family of loss functions, \emph{blob loss}, primarily aimed at maximizing instance-level detection metrics, such as F1 score and sensitivity. \emph{Blob loss} is designed for semantic segmentation problems where detecting multiple instances matters. We extensively evaluate a DSC-based \emph{blob loss} in five complex 3D semantic segmentation tasks featuring pronounced instance heterogeneity in terms of texture and morphology. Compared to soft Dice loss, we achieve 5% improvement for MS lesions, 3% improvement for liver tumor, and an average 2% improvement for microscopy segmentation tasks considering F1 score.
Approaching Peak Ground Truth
Kofler, Florian, Wahle, Johannes, Ezhov, Ivan, Wagner, Sophia, Al-Maskari, Rami, Gryska, Emilia, Todorov, Mihail, Bukas, Christina, Meissen, Felix, Peng, Tingying, Ertürk, Ali, Rueckert, Daniel, Heckemann, Rolf, Kirschke, Jan, Zimmer, Claus, Wiestler, Benedikt, Menze, Bjoern, Piraud, Marie
Machine learning models are typically evaluated by computing similarity with reference annotations and trained by maximizing similarity with such. Especially in the biomedical domain, annotations are subjective and suffer from low inter- and intra-rater reliability. Since annotations only reflect one interpretation of the real world, this can lead to sub-optimal predictions even though the model achieves high similarity scores. Here, the theoretical concept of PGT is introduced. PGT marks the point beyond which an increase in similarity with the \emph{reference annotation} stops translating to better RWMP. Additionally, a quantitative technique to approximate PGT by computing inter- and intra-rater reliability is proposed. Finally, four categories of PGT-aware strategies to evaluate and improve model performance are reviewed.