yolov11
Comparative Analysis of Object Detection Algorithms for Surface Defect Detection
This article compares the performance of six prominent object detection algorithms YOLOv11, RetinaNet, Fast R-CNN, YOLOv8, RT - DETR, and DETR on the NEU - DET surface defect detection dataset comprising images representing various metal surface defects, a crucial application in industrial quality control. Each model's performance was assessed regar ding detection accuracy, speed, and robustness across different defect types such as scratches, inclusions, and rolled-in scales. YOLOv11, a state-of-the-art real-time object detection algorithm, demonstrated superior performance compared to the other methods, achieving a remarkable 70% higher accuracy on average. This improvement can be attributed to YOLOv11's enhanced feature extraction capabilities and ability to process the entire image in a single forward pass, making it faster and more efficient in detecting smaller surface defects. Additionally, YOLOv11's architecture optimizations, such as improved anchor box generation and deeper convolutional layers, contributed to more precise localization of defects.
- South America > Ecuador > Pichincha Province > Quito (0.04)
- South America > Ecuador > Azuay Province > Cuenca (0.04)
- North America > United States > Tennessee > Davidson County > Nashville (0.04)
- (5 more...)
Image Segmentation and Classification of E-waste for Training Robots for Waste Segregation
Abstract--Industry partners provided a problem statement that involves classifying electronic waste using machine learning models, which will be utilized by pick-and-place robots for waste segregation. This was achieved by taking common electronic waste items, such as a mouse and a charger, unsol-dering them, and taking pictures to create a custom dataset. The state-of-the-art YOLOv11 model was trained and run to achieve 70 mAP in real-time. The Mask R-CNN model was also trained and achieved 41 mAP . The model can be integrated with pick-and-place robots to perform segregation of e-waste. Electronic waste (e-waste) is one of the fastest-growing solid waste streams globally [2].
- North America > United States > California > Santa Clara County > Palo Alto (0.05)
- Asia > India (0.05)
PickleBall: Secure Deserialization of Pickle-based Machine Learning Models (Extended Report)
Kellas, Andreas D., Christou, Neophytos, Jiang, Wenxin, Li, Penghui, Simon, Laurent, David, Yaniv, Kemerlis, Vasileios P., Davis, James C., Yang, Junfeng
Machine learning model repositories such as the Hugging Face Model Hub facilitate model exchanges. However, bad actors can deliver malware through compromised models. Existing defenses such as safer model formats, restrictive (but inflexible) loading policies, and model scanners have shortcomings: 44.9% of popular models on Hugging Face still use the insecure pickle format, 15% of these cannot be loaded by restrictive loading policies, and model scanners have both false positives and false negatives. Pickle remains the de facto standard for model exchange, and the ML community lacks a tool that offers transparent safe loading. We present PickleBall to help machine learning engineers load pickle-based models safely. PickleBall statically analyzes the source code of a given machine learning library and computes a custom policy that specifies a safe load-time behavior for benign models. PickleBall then dynamically enforces the policy during load time as a drop-in replacement for the pickle module. PickleBall generates policies that correctly load 79.8% of benign pickle-based models in our dataset, while rejecting all (100%) malicious examples in our dataset. In comparison, evaluated model scanners fail to identify known malicious models, and the state-of-art loader loads 22% fewer benign models than PickleBall. PickleBall removes the threat of arbitrary function invocation from malicious pickle-based models, raising the bar for attackers to depend on code reuse techniques.
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > Rhode Island > Providence County > Providence (0.04)
- North America > United States > Indiana > Tippecanoe County > West Lafayette (0.04)
- (8 more...)
- Leisure & Entertainment > Sports (1.00)
- Information Technology > Security & Privacy (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (0.87)
A Comparative Study of YOLOv8 to YOLOv11 Performance in Underwater Vision Tasks
Hung, Gordon, Rodriguez, Ivan Felipe
Autonomous underwater vehicles (AUVs) increasingly rely on on-board computer-vision systems for tasks such as habitat mapping, ecological monitoring, and infrastructure inspection. However, underwater imagery is hindered by light attenuation, turbidity, and severe class imbalance, while the computational resources available on AUVs are limited. One-stage detectors from the YOLO family are attractive because they fuse localization and classification in a single, low-latency network; however, their terrestrial benchmarks (COCO, PASCAL-VOC, Open Images) leave open the question of how successive YOLO releases perform in the marine domain. We curate two openly available datasets that span contrasting operating conditions: a Coral Disease set (4,480 images, 18 classes) and a Fish Species set (7,500 images, 20 classes). For each dataset, we create four training regimes (25 %, 50 %, 75 %, 100 % of the images) while keeping balanced validation and test partitions fixed. We train YOLOv8-s, YOLOv9-s, YOLOv10-s, and YOLOv11-s with identical hyperparameters (100 epochs, 640 px input, batch = 16, T4 GPU) and evaluate precision, recall, mAP50, mAP50-95, per-image inference time, and frames-per-second (FPS). Post-hoc Grad-CAM visualizations probe feature utilization and localization faithfulness. Across both datasets, accuracy saturates after YOLOv9, suggesting architectural innovations primarily target efficiency rather than accuracy. Inference speed, however, improves markedly. Our results (i) provide the first controlled comparison of recent YOLO variants on underwater imagery, (ii) show that lightweight YOLOv10 offers the best speed-accuracy trade-off for embedded AUV deployment, and (iii) deliver an open, reproducible benchmark and codebase to accelerate future marine-vision research.
- South America > Colombia > Bogotá D.C. > Bogotá (0.04)
- Asia > Taiwan (0.04)
- North America > United States > Rhode Island > Providence County > Providence (0.04)
- (3 more...)
The Impact of Image Resolution on Face Detection: A Comparative Analysis of MTCNN, YOLOv XI and YOLOv XII models
Ömercikoğlu, Ahmet Can, Yönügül, Mustafa Mansur, Erdoğmuş, Pakize
Face detection is a crucial component in many AI-driven applications such as surveillance, biometric authentication, and human-computer interaction. However, real-world conditions like low-resolution imagery present significant challenges that degrade detection performance. In this study, we systematically investigate the impact of input resolution on the accuracy and robustness of three prominent deep learning-based face detectors: YOLOv11, YOLOv12, and MTCNN. Using the WIDER FACE dataset, we conduct extensive evaluations across multiple image resolutions (160x160, 320x320, and 640x640) and assess each model's performance using metrics such as precision, recall, mAP50, mAP50-95, and inference time. Results indicate that YOLOv11 outperforms YOLOv12 and MTCNN in terms of detection accuracy, especially at higher resolutions, while YOLOv12 exhibits slightly better recall. MTCNN, although competitive in landmark localization, lags in real-time inference speed. Our findings provide actionable insights for selecting resolution-aware face detection models suitable for varying operational constraints.
Autonomous UAV Navigation for Search and Rescue Missions Using Computer Vision and Convolutional Neural Networks
Šiktar, Luka, Ćaran, Branimir, Šekoranja, Bojan, Švaco, Marko
In this paper, we present a subsystem, using Unmanned Aerial Vehicles (UAV), for search and rescue missions, focusing on people detection, face recognition and tracking of identified individuals. The proposed solution integrates a UAV with ROS2 framework, that utilizes multiple convolutional neural networks (CNN) for search missions. System identification and PD controller deployment are performed for autonomous UAV navigation. The ROS2 environment utilizes the YOLOv11 and YOLOv11-pose CNNs for tracking purposes, and the dlib library CNN for face recognition. The system detects a specific individual, performs face recognition and starts tracking. If the individual is not yet known, the UAV operator can manually locate the person, save their facial image and immediately initiate the tracking process. The tracking process relies on specific keypoints identified on the human body using the YOLOv11-pose CNN model. These keypoints are used to track a specific individual and maintain a safe distance. To enhance accurate tracking, system identification is performed, based on measurement data from the UAVs IMU. The identified system parameters are used to design PD controllers that utilize YOLOv11-pose to estimate the distance between the UAVs camera and the identified individual. The initial experiments, conducted on 14 known individuals, demonstrated that the proposed subsystem can be successfully used in real time. The next step involves implementing the system on a large experimental UAV for field use and integrating autonomous navigation with GPS-guided control for rescue operations planning.
- North America > United States > Nevada > Clark County > Las Vegas (0.04)
- Europe > Croatia > Zagreb County > Zagreb (0.04)
- Asia > China > Chongqing Province > Chongqing (0.04)
- Information Technology (0.50)
- Health & Medicine (0.34)
Vision Controlled Orthotic Hand Exoskeleton
Blais, Connor, Sarker, Md Abdul Baset, Imtiaz, Masudul H.
This paper presents the design and implementation of an AI vision-controlled orthotic hand exoskeleton to enhance rehabilitation and assistive functionality for individuals with hand mobility impairments. The system leverages a Google Coral Dev Board Micro with an Edge TPU to enable real-time object detection using a customized MobileNet\_V2 model trained on a six-class dataset. The exoskeleton autonomously detects objects, estimates proximity, and triggers pneumatic actuation for grasp-and-release tasks, eliminating the need for user-specific calibration needed in traditional EMG-based systems. The design prioritizes compactness, featuring an internal battery. It achieves an 8-hour runtime with a 1300 mAh battery. Experimental results demonstrate a 51ms inference speed, a significant improvement over prior iterations, though challenges persist in model robustness under varying lighting conditions and object orientations. While the most recent YOLO model (YOLOv11) showed potential with 15.4 FPS performance, quantization issues hindered deployment. The prototype underscores the viability of vision-controlled exoskeletons for real-world assistive applications, balancing portability, efficiency, and real-time responsiveness, while highlighting future directions for model optimization and hardware miniaturization.
- North America > United States (0.04)
- Europe > Sweden (0.04)
- Europe > Germany > Brandenburg > Potsdam (0.04)
- Electrical Industrial Apparatus (0.88)
- Health & Medicine > Therapeutic Area > Neurology (0.68)
- Energy (0.66)
Detecting Glioma, Meningioma, and Pituitary Tumors, and Normal Brain Tissues based on Yolov11 and Yolov8 Deep Learning Models
Taha, Ahmed M., Aly, Salah A., Darwish, Mohamed F.
Detecting Glioma, Meningioma, and Pituitary Tumors, and Normal Brain Tissues based on Y olov11 and Y olov8 Deep Learning Models Ahmed M. Taha a, Salah A. Aly b,c, Mohamed F. Darwish d a Dept. of CE, Faculty of Engineering, Egypt University of Informatics, Cairo, Egypt b Faculty of Computing and Data Science, Badya University, Giza, Egypt c CS&Math Branch, Faculty of Science, Fayoum University, Fayoum, Egypt d Dept. of Pathology, Faculty of Medicine, Badya University, Giza, Egypt Abstract --Accurate and quick diagnosis of normal brain tissue Glioma, Meningioma, and Pituitary T umors is crucial for optimal treatment planning and improved medical results. Magnetic Resonance Imaging (MRI) is widely used as a non-invasive diagnostic tool for detecting brain abnormalities, including tumors. However, manual interpretation of MRI scans is often time-consuming, prone to human error, and dependent on highly specialized expertise. This paper proposes an advanced AI-driven technique to detecting glioma, meningioma, and pituitary brain tumors using Y oloV11 and Y oloV8 deep learning models. Methods: Using a transfer learning-based fine-tuning approach, we integrate cutting-edge deep learning techniques with medical imaging to classify brain tumors into four categories: No-T umor, Glioma, Meningioma, and Pituitary T umors.
- Africa > Middle East > Egypt > Giza Governorate > Giza (0.44)
- Africa > Middle East > Egypt > Cairo Governorate > Cairo (0.24)
- Africa > Sub-Saharan Africa (0.04)
- Asia > Japan (0.04)
- Health & Medicine > Therapeutic Area > Neurology (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
- Health & Medicine > Therapeutic Area > Oncology > Head & Neck Cancer (0.60)
Acute Lymphoblastic Leukemia Diagnosis Employing YOLOv11, YOLOv8, ResNet50, and Inception-ResNet-v2 Deep Learning Models
Thousands of individuals succumb annually to leukemia alone. As artificial intelligence-driven technologies continue to evolve and advance, the question of their applicability and reliability remains unresolved. This study aims to utilize image processing and deep learning methodologies to achieve state-of-the-art results for the detection of Acute Lymphoblastic Leukemia (ALL) using data that best represents real-world scenarios. ALL is one of several types of blood cancer, and it is an aggressive form of leukemia. In this investigation, we examine the most recent advancements in ALL detection, as well as the latest iteration of the YOLO series and its performance. We address the question of whether white blood cells are malignant or benign. Additionally, the proposed models can identify different ALL stages, including early stages. Furthermore, these models can detect hematogones despite their frequent misclassification as ALL. By utilizing advanced deep learning models, namely, YOLOv8, YOLOv11, ResNet50 and Inception-ResNet-v2, the study achieves accuracy rates as high as 99.7%, demonstrating the effectiveness of these algorithms across multiple datasets and various real-world situations.
- Asia > Singapore (0.04)
- Asia > Japan (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (4 more...)
- Health & Medicine > Therapeutic Area > Oncology > Leukemia (1.00)
- Health & Medicine > Therapeutic Area > Hematology (1.00)
Polyp detection in colonoscopy images using YOLOv11
Sahoo, Alok Ranjan, Sahoo, Satya Sangram, Chakraborty, Pavan
Colorectal cancer (CRC) is one of the most commonly diagnosed cancers all over the world. It starts as a polyp in the inner lining of the colon. To prevent CRC, early polyp detection is required. Colonosopy is used for the inspection of the colon. Generally, the images taken by the camera placed at the tip of the endoscope are analyzed by the experts manually. Various traditional machine learning models have been used with the rise of machine learning. Recently, deep learning models have shown more effectiveness in polyp detection due to their superiority in generalizing and learning small features. These deep learning models for object detection can be segregated into two different types: single-stage and two-stage. Generally, two stage models have higher accuracy than single stage ones but the single stage models have low inference time. Hence, single stage models are easy to use for quick object detection. YOLO is one of the singlestage models used successfully for polyp detection. It has drawn the attention of researchers because of its lower inference time. The researchers have used Different versions of YOLO so far, and with each newer version, the accuracy of the model is increasing. This paper aims to see the effectiveness of the recently released YOLOv11 to detect polyp. We analyzed the performance for all five models of YOLOv11 (YOLO11n, YOLO11s, YOLO11m, YOLO11l, YOLO11x) with Kvasir dataset for the training and testing. Two different versions of the dataset were used. The first consisted of the original dataset, and the other was created using augmentation techniques. The performance of all the models with these two versions of the dataset have been analysed.
- Asia > India (0.05)
- Asia > Middle East > Republic of Türkiye > Karaman Province > Karaman (0.05)
- North America > United States > New York > New York County > New York City (0.04)
- Europe > Norway (0.04)
- Health & Medicine > Diagnostic Medicine (1.00)
- Health & Medicine > Therapeutic Area > Oncology > Colorectal Cancer (0.88)