Video-SwinUNet: Spatio-temporal Deep Learning Framework for VFSS Instance Segmentation
Zeng, Chengxi, Yang, Xinyu, Smithard, David, Mirmehdi, Majid, Gambaruto, Alberto M, Burghardt, Tilo
–arXiv.org Artificial Intelligence
This paper presents a deep learning framework for medical video segmentation. Convolution neural network (CNN) and transformer-based methods have achieved great milestones in medical image segmentation tasks due to their incredible semantic feature encoding and global information comprehension abilities. However, most existing approaches ignore a salient aspect of medical video data - the temporal dimension. Our proposed framework explicitly extracts features from neighbouring frames across the temporal dimension and incorporates them with a temporal feature blender, which then tokenises the high-level spatio-temporal feature to form a strong global feature encoded via a Swin Transformer. The final segmentation results are produced via a UNet-like encoder-decoder architecture. Our model outperforms other approaches by a significant margin and improves the segmentation benchmarks on the VFSS2022 dataset, achieving a dice coefficient of 0.8986 and 0.8186 for the two datasets tested. Our studies also show the efficacy of the temporal feature blending scheme and cross-dataset transferability of learned capabilities. Code and models are fully available at https://github.com/SimonZeng7108/Video-SwinUNet.
arXiv.org Artificial Intelligence
Jul-4-2023
- Country:
- Europe
- Spain > Andalusia
- Granada Province > Granada (0.04)
- Switzerland > Basel-City
- Basel (0.04)
- United Kingdom > England
- Bristol (0.04)
- Spain > Andalusia
- Europe
- Genre:
- Research Report > New Finding (0.54)
- Industry:
- Health & Medicine
- Diagnostic Medicine > Imaging (0.36)
- Therapeutic Area > Neurology (0.46)
- Health & Medicine
- Technology: