yolo model
ShadowWolf -- Automatic Labelling, Evaluation and Model Training Optimised for Camera Trap Wildlife Images
The continuous growth of the global human population is leading to the expansion of human habitats, resulting in decreasing wildlife spaces and increasing human-wildlife interactions. These interactions can range from minor disturbances, such as raccoons in urban waste bins, to more severe consequences, including species extinction. As a result, the monitoring of wildlife is gaining significance in various contexts. Artificial intelligence (AI) offers a solution by automating the recognition of animals in images and videos, thereby reducing the manual effort required for wildlife monitoring. Traditional AI training involves three main stages: image collection, labelling, and model training. However, the variability, for example, in the landscape (e.g., mountains, open fields, forests), weather (e.g., rain, fog, sunshine), lighting (e.g., day, night), and camera-animal distances presents significant challenges to model robustness and adaptability in real-world scenarios. In this work, we propose a unified framework, called ShadowWolf, designed to address these challenges by integrating and optimizing the stages of AI model training and evaluation. The proposed framework enables dynamic model retraining to adjust to changes in environmental conditions and application requirements, thereby reducing labelling efforts and allowing for on-site model adaptation. This adaptive and unified approach enhances the accuracy and efficiency of wildlife monitoring systems, promoting more effective and scalable conservation efforts.
- Europe > Germany > Bremen > Bremen (0.28)
- North America (0.14)
- Europe > Spain (0.04)
- (2 more...)
- Workflow (1.00)
- Research Report (0.82)
- Overview (0.67)
- Information Technology (0.67)
- Energy (0.46)
- Information Technology > Sensing and Signal Processing > Image Processing (1.00)
- Information Technology > Communications > Social Media (1.00)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.47)
Appendix
In object detection and many other computer vision benchmarks, the image resolutions as well as the aspect ratios are usually not fixed as the image classification task. For the first layer, the PE is interpolated following ViT. In a word, Type-I uses more PEs and Type-II uses larger PE. In our paper, small-and base-sized models use this setting. The detailed configurations are given in Tab. 1. PE-cls to PE-det Rand.
YOLO-Based Defect Detection for Metal Sheets
Chou, Po-Heng, Wang, Chun-Chi, Mao, Wei-Lung
In this paper, we propose a YOLO-based deep learning (DL) model for automatic defect detection to solve the time-consuming and labor-intensive tasks in industrial manufacturing. In our experiments, the images of metal sheets are used as the dataset for training the YOLO model to detect the defects on the surfaces and in the holes of metal sheets. However, the lack of metal sheet images significantly degrades the performance of detection accuracy. To address this issue, the ConSinGAN is used to generate a considerable amount of data. Four versions of the YOLO model (i.e., YOLOv3, v4, v7, and v9) are combined with the ConSinGAN for data augmentation. The proposed YOLOv9 model with ConSinGAN outperforms the other YOLO models with an accuracy of 91.3%, and a detection time of 146 ms. The proposed YOLOv9 model is integrated into manufacturing hardware and a supervisory control and data acquisition (SCADA) system to establish a practical automated optical inspection (AOI) system. Additionally, the proposed automated defect detection is easily applied to other components in industrial manufacturing.
A Comparative Study of YOLOv8 to YOLOv11 Performance in Underwater Vision Tasks
Hung, Gordon, Rodriguez, Ivan Felipe
Autonomous underwater vehicles (AUVs) increasingly rely on on-board computer-vision systems for tasks such as habitat mapping, ecological monitoring, and infrastructure inspection. However, underwater imagery is hindered by light attenuation, turbidity, and severe class imbalance, while the computational resources available on AUVs are limited. One-stage detectors from the YOLO family are attractive because they fuse localization and classification in a single, low-latency network; however, their terrestrial benchmarks (COCO, PASCAL-VOC, Open Images) leave open the question of how successive YOLO releases perform in the marine domain. We curate two openly available datasets that span contrasting operating conditions: a Coral Disease set (4,480 images, 18 classes) and a Fish Species set (7,500 images, 20 classes). For each dataset, we create four training regimes (25 %, 50 %, 75 %, 100 % of the images) while keeping balanced validation and test partitions fixed. We train YOLOv8-s, YOLOv9-s, YOLOv10-s, and YOLOv11-s with identical hyperparameters (100 epochs, 640 px input, batch = 16, T4 GPU) and evaluate precision, recall, mAP50, mAP50-95, per-image inference time, and frames-per-second (FPS). Post-hoc Grad-CAM visualizations probe feature utilization and localization faithfulness. Across both datasets, accuracy saturates after YOLOv9, suggesting architectural innovations primarily target efficiency rather than accuracy. Inference speed, however, improves markedly. Our results (i) provide the first controlled comparison of recent YOLO variants on underwater imagery, (ii) show that lightweight YOLOv10 offers the best speed-accuracy trade-off for embedded AUV deployment, and (iii) deliver an open, reproducible benchmark and codebase to accelerate future marine-vision research.
- South America > Colombia > Bogotá D.C. > Bogotá (0.04)
- Asia > Taiwan (0.04)
- North America > United States > Rhode Island > Providence County > Providence (0.04)
- (3 more...)
HOTSPOT-YOLO: A Lightweight Deep Learning Attention-Driven Model for Detecting Thermal Anomalies in Drone-Based Solar Photovoltaic Inspections
Thermal anomaly detection in solar photovoltaic (PV) systems is essential for ensuring operational efficiency and reducing maintenance costs. In this study, we developed and named HOTSPOT - YOLO, a lightweight artificial intelligence (AI) model that integrat es an efficient convolutional neural network backbone and attention mechanisms to improve object detection. This model is specifically designed for drone - based thermal inspections of PV systems, addressing the unique challenges of detecting small and subtl e thermal anomalies, such as hotspots and defective modules, while maintaining real - time performance. Experimental results demonstrate a mean a verage p recision of 90.8%, reflecting a significant improvement over baseline object detection models. With a reduced computational load and robustness under diverse environmental conditions, HOTSPOT - YOLO offers a scalable and reliable solution for large - scale PV inspections. This work highlights the integration of advanced AI techniques with practical engineering ap plications, revolutionizing automated fault detection in renewable energy systems.
- Europe > Denmark (0.14)
- Asia > Nepal (0.04)
- Asia > Middle East > Republic of Türkiye (0.04)
- Asia > India (0.04)
- Overview (0.68)
- Research Report > New Finding (0.54)
Appendix
In object detection and many other computer vision benchmarks, the image resolutions as well as the aspect ratios are usually not fixed as the image classification task. For the first layer, the PE is interpolated following ViT. In a word, Type-I uses more PEs and Type-II uses larger PE. In our paper, small-and base-sized models use this setting. The detailed configurations are given in Tab. 1. PE-cls to PE-det Rand.
Privacy-Preserving Multi-Stage Fall Detection Framework with Semi-supervised Federated Learning and Robotic Vision Confirmation
Azghadi, Seyed Alireza Rahimi, Nguyen, Truong-Thanh-Hung, Fournier, Helene, Wachowicz, Monica, Richard, Rene, Palma, Francis, Cao, Hung
The aging population is growing rapidly, and so is the danger of falls in older adults. A major cause of injury is falling, and detection in time can greatly save medical expenses and recovery time. However, to provide timely intervention and avoid unnecessary alarms, detection systems must be effective and reliable while addressing privacy concerns regarding the user. In this work, we propose a framework for detecting falls using several complementary systems: a semi-supervised federated learning-based fall detection system (SF2D), an indoor localization and navigation system, and a vision-based human fall recognition system. A wearable device and an edge device identify a fall scenario in the first system. On top of that, the second system uses an indoor localization technique first to localize the fall location and then navigate a robot to inspect the scenario. A vision-based detection system running on an edge device with a mounted camera on a robot is used to recognize fallen people. Each of the systems of this proposed framework achieves different accuracy rates. Specifically, the SF2D has a 0.81% failure rate equivalent to 99.19% accuracy, while the vision-based fallen people detection achieves 96.3% accuracy. However, when we combine the accuracy of these two systems with the accuracy of the navigation system (95% success rate), our proposed framework creates a highly reliable performance for fall detection, with an overall accuracy of 99.99%. Not only is the proposed framework safe for older adults, but it is also a privacy-preserving solution for detecting falls.
- South America > Uruguay > Maldonado > Maldonado (0.04)
- North America > Canada > Newfoundland and Labrador > Labrador (0.04)
- North America > Canada > New Brunswick > Fredericton (0.04)
- (7 more...)
- Research Report > New Finding (1.00)
- Overview (1.00)
- Information Technology > Security & Privacy (1.00)
- Health & Medicine > Health Care Technology (1.00)
- Health & Medicine > Health Care Providers & Services (1.00)
- Health & Medicine > Consumer Health (1.00)
Rapid morphology characterization of two-dimensional TMDs and lateral heterostructures based on deep learning
He, Junqi, Zhang, Yujie, Wang, Jialu, Wang, Tao, Zhang, Pan, Cai, Chengjie, Yang, Jinxing, Lin, Xiao, Yang, Xiaohui
Leveraging advancements in artificial intelligence, we introduce a deep learning-based method for efficiently characterizing heterostructures and 2D materials, specifically MoS 2-MoSe 2 lateral heterostructures and MoS 2 flakes with varying shapes and thicknesses. By utilizing YOLO models, we achieve an accuracy rate of over 94.67% in identifying these materials. Additionally, we explore the application of transfer learning across different materials, which further enhances model performance. This model exhibits robust generalization and anti-interference ability, ensuring reliable results in diverse scenarios. To facilitate practical use, we have developed an application that enables real-time analysis directly from optical microscope images, making the process significantly faster and more cost-effective than traditional methods. This deep learning-driven approach represents a promising tool for the rapid and accurate characterization of 2D materials, opening new avenues for research and development in material science. Keywords 2D material, TMDs, lateral heterostructure, deep learning, instance segmentation, morphology characterization Introduction Two-dimensional (2D) materials have attracted significant attention due to their excellent mechanical, electrical, thermal, and optical properties, making them ideal candidates for next-generation technologies.
- Asia > China > Zhejiang Province > Hangzhou (0.06)
- North America > United States > Texas > Shelby County > Center (0.04)
- North America > United States > Texas > Harris County > Houston (0.04)
Drone Detection and Tracking with YOLO and a Rule-based Method
Bhattacharya, Purbaditya, Nowak, Patrick
Drones or unmanned aerial vehicles are traditionally used for military missions, warfare, and espionage. However, the usage of drones has significantly increased due to multiple industrial applications involving security and inspection, transportation, research purposes, and recreational drone flying. Such an increased volume of drone activity in public spaces requires regulatory actions for purposes of privacy protection and safety. Hence, detection of illegal drone activities such as boundary encroachment becomes a necessity. Such detection tasks are usually automated and performed by deep learning models which are trained on annotated image datasets. This paper builds on a previous work and extends an already published open source dataset. A description and analysis of the entire dataset is provided. The dataset is used to train the YOLOv7 deep learning model and some of its minor variants and the results are provided. Since the detection models are based on a single image input, a simple cross-correlation based tracker is used to reduce detection drops and improve tracking performance in videos. Finally, the entire drone detection system is summarized.
- Information Technology > Robotics & Automation (0.34)
- Law > Criminal Law (0.34)
Optimizing Helmet Detection with Hybrid YOLO Pipelines: A Detailed Analysis
M, Vaikunth, D, Dejey, C, Vishaal, S, Balamurali
Helmet detection is crucial for advancing protection levels in public road traffic dynamics. This problem statement translates to an object detection task. Therefore, this paper compares recent You Only Look Once (YOLO) models in the context of helmet detection in terms of reliability and computational load. Specifically, YOLOv8, YOLOv9, and the newly released YOLOv11 have been used. Besides, a modified architectural pipeline that remarkably improves the overall performance has been proposed in this manuscript. This hybridized YOLO model (h-YOLO) has been pitted against the independent models for analysis that proves h-YOLO is preferable for helmet detection over plain YOLO models. The models were tested using a range of standard object detection benchmarks such as recall, precision, and mAP (Mean Average Precision). In addition, training and testing times were recorded to provide the overall scope of the models in a real-time detection scenario.
- Asia > India > Tamil Nadu > Chennai (0.05)
- Oceania > Australia > Victoria > Melbourne (0.04)
- North America > Montserrat (0.04)
- (2 more...)
- Research Report (0.64)
- Personal (0.46)
- Overview (0.46)
- Information Technology (0.93)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (0.68)
- Transportation > Ground > Road (0.48)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.94)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis (0.90)