Goto

Collaborating Authors

 deeplabv3




Automatic Pith Detection in Tree Cross-Section Images Using Deep Learning

Liao, Tzu-I, Fakhry, Mahmoud, Varghese, Jibin Yesudas

arXiv.org Artificial Intelligence

Pith detection in tree cross-sections is essential for forestry and wood quality analysis but remains a manual, error-prone task. This study evaluates deep learning models -- YOLOv9, U-Net, Swin Transformer, DeepLabV3, and Mask R-CNN -- to automate the process efficiently. A dataset of 582 labeled images was dynamically augmented to improve generalization. Swin Transformer achieved the highest accuracy (0.94), excelling in fine segmentation. YOLOv9 performed well for bounding box detection but struggled with boundary precision. U-Net was effective for structured patterns, while DeepLabV3 captured multi-scale features with slight boundary imprecision. Mask R-CNN initially underperformed due to overlapping detections, but applying Non-Maximum Suppression (NMS) improved its IoU from 0.45 to 0.80. Generalizability was next tested using an oak dataset of 11 images from Oregon State University's Tree Ring Lab. Additionally, for exploratory analysis purposes, an additional dataset of 64 labeled tree cross-sections was used to train the worst-performing model to see if this would improve its performance generalizing to the unseen oak dataset. Key challenges included tensor mismatches and boundary inconsistencies, addressed through hyperparameter tuning and augmentation. Our results highlight deep learning's potential for tree cross-section pith detection, with model choice depending on dataset characteristics and application needs.



Looking Beyond Single Images for Contrastive Semantic Segmentation Learning - Supplementary Material - 1 Additional results 1.1 Controlled experiment on auxiliary label generation

Neural Information Processing Systems

Table 1 reports the results of a controlled experiment evaluating different components in our framework for auxiliary label generation. Positive correspondences are generated by matching pixels across different augmentations of the same image. With respect to the clustering algorithm, K-means performs better than DBSCAN (#4 vs. #5), which is We show qualitative results, comparing different feature extractors in Figure 1. DBSCAN is limited by the memory and computational complexity. Corresponding qualitative results are shown in Figure 3. Tables 3-5 show We observe the best performance when 5% outliers are removed.


AI Driven Water Segmentation with deep learning models for Enhanced Flood Monitoring

Mou, Sanjida Afrin, Chowdhury, Tasfia Noor, Mannan, Adib Ibn, Mim, Sadia Nourin, Tarannum, Lubana, Noman, Tasrin, Ahamed, Jamal Uddin

arXiv.org Artificial Intelligence

Flooding is a major natural hazard causing significant fatalities and economic losses annually, with increasing frequency due to climate change. Rapid and accurate flood detection and monitoring are crucial for mitigating these impacts. This study compares the performance of three deep learning models UNet, ResNet, and DeepLabv3 for pixelwise water segmentation to aid in flood detection, utilizing images from drones, in field observations, and social media. This study involves creating a new dataset that augments wellknown benchmark datasets with flood-specific images, enhancing the robustness of the models. The UNet, ResNet, and DeepLab v3 architectures are tested to determine their effectiveness in various environmental conditions and geographical locations, and the strengths and limitations of each model are also discussed here, providing insights into their applicability in different scenarios by predicting image segmentation masks. This fully automated approach allows these models to isolate flooded areas in images, significantly reducing processing time compared to traditional semi-automated methods. The outcome of this study is to predict segmented masks for each image effected by a flood disaster and the validation accuracy of these models. This methodology facilitates timely and continuous flood monitoring, providing vital data for emergency response teams to reduce loss of life and economic damages. It offers a significant reduction in the time required to generate flood maps, cutting down the manual processing time. Additionally, we present avenues for future research, including the integration of multimodal data sources and the development of robust deep learning architectures tailored specifically for flood detection tasks. Overall, our work contributes to the advancement of flood management strategies through innovative use of deep learning technologies.


Semantic Segmentation of Spontaneous Intracerebral Hemorrhage, Intraventricular Hemorrhage, and Associated Edema on CT Images Using Deep Learning

#artificialintelligence

"Just Accepted" papers have undergone full peer review and have been accepted for publication in Radiology: Artificial Intelligence. This article will undergo copyediting, layout, and proof review before it is published in its final version. Please note that during production of the final copyedited article, errors may be discovered which could affect the content. This study evaluated deep learning algorithms for semantic segmentation and quantification of intracerebral hemorrhage (ICH), perihematomal edema (PHE), and intraventricular hemorrhage (IVH) on noncontrast CT (NCCT) scans of patients with spontaneous ICH. Models were assessed on 1,732 annotated baseline NCCT scans obtained from the TICH-2 international multicenter trial (ISRCTN93732214), and different loss functions using three-dimensional nnU-Net were examined to address class imbalance (30% of participants with IVH in dataset).


MassMIND: Massachusetts Maritime INfrared Dataset

Nirgudkar, Shailesh, DeFilippo, Michael, Sacarny, Michael, Benjamin, Michael, Robinette, Paul

arXiv.org Artificial Intelligence

Recent advances in deep learning technology have triggered radical progress in the autonomy of ground vehicles. Marine coastal Autonomous Surface Vehicles (ASVs) that are regularly used for surveillance, monitoring and other routine tasks can benefit from this autonomy. Long haul deep sea transportation activities are additional opportunities. These two use cases present very different terrains -- the first being coastal waters -- with many obstacles, structures and human presence while the latter is mostly devoid of such obstacles. Variations in environmental conditions are common to both terrains. Robust labeled datasets mapping such terrains are crucial in improving the situational awareness that can drive autonomy. However, there are only limited such maritime datasets available and these primarily consist of optical images. Although, Long Wave Infrared (LWIR) is a strong complement to the optical spectrum that helps in extreme light conditions, a labeled public dataset with LWIR images does not currently exist. In this paper, we fill this gap by presenting a labeled dataset of over 2,900 LWIR segmented images captured in coastal maritime environment under diverse conditions. The images are labeled using instance segmentation and classified in seven categories -- sky, water, obstacle, living obstacle, bridge, self and background. We also evaluate this dataset across three deep learning architectures (UNet, PSPNet, DeepLabv3) and provide detailed analysis of its efficacy. While the dataset focuses on the coastal terrain it can equally help deep sea use cases. Such terrain would have less traffic, and the classifier trained on cluttered environment would be able to handle sparse scenes effectively. We share this dataset with the research community with the hope that it spurs new scene understanding capabilities in the maritime environment.


MTP: Multi-Task Pruning for Efficient Semantic Segmentation Networks

Chen, Xinghao, Zhang, Yiman, Wang, Yunhe

arXiv.org Artificial Intelligence

This paper focuses on channel pruning for semantic segmentation networks. Previous methods to compress and accelerate deep neural networks in the classification task cannot be straightforwardly applied to the semantic segmentation network that involves an implicit multi-task learning problem via pre-training. To identify the redundancy in segmentation networks, we present a multi-task channel pruning approach. The importance of each convolution filter \wrt the channel of an arbitrary layer will be simultaneously determined by the classification and segmentation tasks. In addition, we develop an alternative scheme for optimizing importance scores of filters in the entire network. Experimental results on several benchmarks illustrate the superiority of the proposed algorithm over the state-of-the-art pruning methods. Notably, we can obtain an about $2\times$ FLOPs reduction on DeepLabv3 with only an about $1\%$ mIoU drop on the PASCAL VOC 2012 dataset and an about $1.3\%$ mIoU drop on Cityscapes dataset, respectively.


The DeepLab Family

#artificialintelligence

Image segmentation tasks have seen lots of developments in recent years, and have become one of the most researched topics in Computer Vision⁶. One of the standards for segmentation is represented by the Deep Labelling for Image Segmentation architecture, also known as DeepLab. The approach was developed by Chen et al.¹ ² ³ ⁴ and different versions employing different mechanisms were proposed over time. In this article, a brief overview of the different DeepLab algorithms and their basic functioning will be given. The first appearance of the DeepLab architecture is found in [1].