Goto

Collaborating Authors

 Gyawali, Prashnna


NCDD: Nearest Centroid Distance Deficit for Out-Of-Distribution Detection in Gastrointestinal Vision

arXiv.org Artificial Intelligence

The integration of deep learning tools in gastrointestinal vision holds the potential for significant advancements in diagnosis, treatment, and overall patient care. A major challenge, however, is these tools' tendency to make overconfident predictions, even when encountering unseen or newly emerging disease patterns, undermining their reliability. We address this critical issue of reliability by framing it as an out-of-distribution (OOD) detection problem, where previously unseen and emerging diseases are identified as OOD examples. However, gastrointestinal images pose a unique challenge due to the overlapping feature representations between in- Distribution (ID) and OOD examples. Existing approaches often overlook this characteristic, as they are primarily developed for natural image datasets, where feature distinctions are more apparent. Despite the overlap, we hypothesize that the features of an in-distribution example will cluster closer to the centroids of their ground truth class, resulting in a shorter distance to the nearest centroid. In contrast, OOD examples maintain an equal distance from all class centroids. Based on this observation, we propose a novel nearest-centroid distance deficit (NCCD) score in the feature space for gastrointestinal OOD detection. Evaluations across multiple deep learning architectures and two publicly available benchmarks, Kvasir2 and Gastrovision, demonstrate the effectiveness of our approach compared to several state-of-the-art methods. The code and implementation details are publicly available at: https://github.com/bhattarailab/NCDD


TE-SSL: Time and Event-aware Self Supervised Learning for Alzheimer's Disease Progression Analysis

arXiv.org Artificial Intelligence

Alzheimer's Disease (AD) represents one of the most pressing challenges in the field of neurodegenerative disorders, with its progression analysis being crucial for understanding disease dynamics and developing targeted interventions. Recent advancements in deep learning and various representation learning strategies, including self-supervised learning (SSL), have shown significant promise in enhancing medical image analysis, providing innovative ways to extract meaningful patterns from complex data. Notably, the computer vision literature has demonstrated that incorporating supervisory signals into SSL can further augment model performance by guiding the learning process with additional relevant information. However, the application of such supervisory signals in the context of disease progression analysis remains largely unexplored. This gap is particularly pronounced given the inherent challenges of incorporating both event and time-to-event information into the learning paradigm. Addressing this, we propose a novel framework, Time and Event-aware SSL (TE-SSL), which integrates time-to-event and event and data as supervisory signals to refine the learning process. Our comparative analysis with existing SSL-based methods in the downstream task of survival analysis shows superior performance across standard metrics.


Multimodal Federated Learning in Healthcare: a review

arXiv.org Artificial Intelligence

Bigger picture Building efficient multimodal AI algorithms from biomedical data, including medical imaging, electronic health records, and personalized sensors would have immense utility for a broad range of health applications. An increasingly common approach towards building such algorithms is a centralized framework, where all the training databases are held together for training models. However, especially in healthcare, due to variety of reasons, including privacy, security, and ethical reasons, gathering datasets from different hospitals, and health centers is not feasible. Federated learning has emerged as an elegant alternative to centralized machine learning where models are trained separately and collaboratively without a centralized database. However, must of the federated learning models have shown their efficacy in the unimodal system, and there is an increasing need for advanced FL algorithms to allow the training of larger and capable model that can absorb heterogeneous private data across multiple modalities. Accordingly, we explore the recent advancements and opportunities for such multimodal federated learning in healthcare; we then discuss the key challenges and promising strategies for overcoming these. Abstract: Recent advancements in multimodal machine learning have empowered the development of accurate and robust AI systems in the medical domain, especially within centralized database systems. Simultaneously, Federated Learning (FL) has progressed, providing a decentralized mechanism where data need not be consolidated, thereby enhancing the privacy and security of sensitive healthcare data. The integration of these two concepts supports the ongoing progress of multimodal learning in healthcare while ensuring the security and privacy of patient records within local data-holding agencies. This paper offers a concise overview of the significance of FL in healthcare and outlines the current state-of-the-art approaches to Multimodal Federated Learning (MMFL) within the healthcare domain. It comprehensively examines the existing challenges in the field, shedding light on the limitations of present models. Finally, the paper outlines potential directions for future advancements in the field, aiming to bridge the gap between cutting-edge AI technology and the imperative need for patient data privacy in healthcare applications. Introduction: Artificial intelligence (AI) tools have been transforming several domains (for example, language translation, speech recognition), and in recent years, it has been showing promise in healthcare applications. Most of such demonstrations have been narrowly focused on tasks using a single modality.