Goto

Collaborating Authors

 Wang, Chaoli


AI-Powered Automated Model Construction for Patient-Specific CFD Simulations of Aortic Flows

arXiv.org Artificial Intelligence

Effectively understanding and managing CVD requires advanced diagnostic tools capable of accurately characterizing complex hemodynamics within the cardiovascular system. While medical imaging modalities such as computed tomography (CT) and magnetic resonance imaging (MRI) provide high-resolution anatomical detail, they lack the capability to directly capture hemodynamics information (e.g., blood flow patterns, pressure, and wall shear stress fields) critical for understanding vascular function and pathology. To bridge this gap, image-based computational fluid dynamics (CFD) has emerged as a powerful computational paradigm that derives hemodynamic information from anatomical images via conservation laws. Although widely utilized in cardiovascular research, the clinical application of image-based CFD for diagnosis and surgical planning remains limited, largely due to the challenges associated with efficient and accurate model construction [2-4]. Constructing patient-specific vascular models for image-based CFD involves multiple steps, including image segmentation, geometry modeling, and mesh generation for the computational domain, all of which are critical to ensuring the fidelity of the final simulation results. However, the standard workflow heavily relies on manual methods, making it highly labor-intensive and time-consuming.


Meta-INR: Efficient Encoding of Volumetric Data via Meta-Learning Implicit Neural Representation

arXiv.org Artificial Intelligence

Implicit neural representation (INR) has emerged as a promising solution for encoding volumetric data, offering continuous representations and seamless compatibility with the volume rendering pipeline. However, optimizing an INR network from randomly initialized parameters for each new volume is computationally inefficient, especially for large-scale time-varying or ensemble volumetric datasets where volumes share similar structural patterns but require independent training. To close this gap, we propose Meta-INR, a pretraining strategy adapted from meta-learning algorithms to learn initial INR parameters from partial observation of a volumetric dataset. Compared to training an INR from scratch, the learned initial parameters provide a strong prior that enhances INR generalizability, allowing significantly faster convergence with just a few gradient updates when adapting to a new volume and better interpretability when analyzing the parameters of the adapted INRs. We demonstrate that Meta-INR can effectively extract high-quality generalizable features that help encode unseen similar volume data across diverse datasets. Furthermore, we highlight its utility in tasks such as simulation parameter analysis and representative timestep selection. The code is available at https://github.com/spacefarers/MetaINR.


Self Pre-training with Topology- and Spatiality-aware Masked Autoencoders for 3D Medical Image Segmentation

arXiv.org Artificial Intelligence

Masked Autoencoders (MAEs) have been shown to be effective in pre-training Vision Transformers (ViTs) for natural and medical image analysis problems. By reconstructing missing pixel/voxel information in visible patches, a ViT encoder can aggregate contextual information for downstream tasks. But, existing MAE pre-training methods, which were specifically developed with the ViT architecture, lack the ability to capture geometric shape and spatial information, which is critical for medical image segmentation tasks. In this paper, we propose a novel extension of known MAEs for self pre-training (i.e., models pre-trained on the same target dataset) for 3D medical image segmentation. (1) We propose a new topological loss to preserve geometric shape information by computing topological signatures of both the input and reconstructed volumes, learning geometric shape information. (2) We introduce a pre-text task that predicts the positions of the centers and eight corners of 3D crops, enabling the MAE to aggregate spatial information. (3) We extend the MAE pre-training strategy to a hybrid state-of-the-art (SOTA) medical image segmentation architecture and co-pretrain it alongside the ViT. (4) We develop a fine-tuned model for downstream segmentation tasks by complementing the pre-trained ViT encoder with our pre-trained SOTA model. Extensive experiments on five public 3D segmentation datasets show the effectiveness of our new approach.


Boosting Medical Image Classification with Segmentation Foundation Model

arXiv.org Artificial Intelligence

The Segment Anything Model (SAM) exhibits impressive capabilities in zero-shot segmentation for natural images. Recently, SAM has gained a great deal of attention for its applications in medical image segmentation. However, to our best knowledge, no studies have shown how to harness the power of SAM for medical image classification. To fill this gap and make SAM a true ``foundation model'' for medical image analysis, it is highly desirable to customize SAM specifically for medical image classification. In this paper, we introduce SAMAug-C, an innovative augmentation method based on SAM for augmenting classification datasets by generating variants of the original images. The augmented datasets can be used to train a deep learning classification model, thereby boosting the classification performance. Furthermore, we propose a novel framework that simultaneously processes raw and SAMAug-C augmented image input, capitalizing on the complementary information that is offered by both. Experiments on three public datasets validate the effectiveness of our new approach.


ECNR: Efficient Compressive Neural Representation of Time-Varying Volumetric Datasets

arXiv.org Artificial Intelligence

ECNR advocates MINER, a multiscale approach [31] proposed for implicit neural representation (INR) of image and point cloud Due to its conceptual simplicity and generality, compressive neural data. Similar to MINER, ECNR adaptively decomposes the spatiotemporal representation has emerged as a promising alternative to traditional volume into blocks via the Laplacian pyramid, starting compression methods for managing massive volumetric datasets. As such, a block is only partitioned further The current practice of neural compression utilizes a single large if its residual remains significant, demanding the capture of finer multilayer perceptron (MLP) to encode the global volume, incurring space-time details for accurate signal reconstruction. To fit the local slow training and inference. This paper presents an efficient compressive spatiotemporal blocks at each scale, we utilize multiple small neural representation (ECNR) solution for time-varying data MLPs, permitting fast encoding and decoding, reduced memory compression, utilizing the Laplacian pyramid for adaptive signal consumption, and enhanced reconstruction quality. Following a multiscale structure, we leverage multiple small Different from MINER, ECNR handles 4D (3D+time) volumetric MLPs at each scale for fitting local content or residual blocks. By datasets, while MINER only processes 2D static images or 3D mesh. Working in concert with the multiscale we group similar blocks into clusters, and each cluster consists of structure, we tailor a deep compression strategy to compact the nearly the same number of blocks. We then assign each cluster resulting model. We show the effectiveness of ECNR with multiple to an MLP and effectively train them in parallel. Furthermore, datasets and compare it with state-of-the-art compression methods we leverage a deep compression strategy (including block-guided (mainly SZ3, TTHRESH, and neurcomp).