Goto

Collaborating Authors

 vision-based tactile sensor


3D Cal: An Open-Source Software Library for Calibrating Tactile Sensors

Kota, Rohan, Shah, Kaival, Colgate, J. Edward, Reardon, Gregory

arXiv.org Artificial Intelligence

Tactile sensing plays a key role in enabling dexterous and reliable robotic manipulation, but realizing this capability requires substantial calibration to convert raw sensor readings into physically meaningful quantities. Despite its near-universal necessity, the calibration process remains ad hoc and labor-intensive. Here, we introduce 3D Cal, an open-source library that transforms a low-cost 3D printer into an automated probing device capable of generating large volumes of labeled training data for tactile sensor calibration. We demonstrate the utility of 3D Cal by calibrating two commercially available vision-based tactile sensors, DIGIT and GelSight Mini, to reconstruct high-quality depth maps using the collected data and a custom convolutional neural network. In addition, we perform a data ablation study to determine how much data is needed for accurate calibration, providing practical guidelines for researchers working with these specific sensors, and we benchmark the trained models on previously unseen objects to evaluate calibration accuracy and generalization performance. By automating tactile sensor calibration, 3D Cal can accelerate tactile sensing research, simplify sensor deployment, and promote the practical integration of tactile sensing in robotic platforms.


Classification of Vision-Based Tactile Sensors: A Review

Li, Haoran, Lin, Yijiong, Lu, Chenghua, Yang, Max, Psomopoulou, Efi, Lepora, Nathan F

arXiv.org Artificial Intelligence

-- Vision-based tactile sensors (VBTS) have gained widespread application in robotic hands, grippers and prosthetics due to their high spatial resolution, low manufacturing costs, and ease of customization. While VBTSs have common design features, such as a camera module, they can differ in a rich diversity of sensing principles, material compositions, multimodal approaches, and data interpretation methods. Here, we propose a novel classification of VBTS that categorizes the technology into two primary sensing principles based on the underlying transduction of contact into a tactile image: the Marker-Based Transduction Principle and the Intensity-Based Transduction Principle. Marker-Based Transduction interprets tactile information by detecting marker displacement and changes in marker density. Depending on the design of the contact module, Marker-Based Transduction can be further divided into two subtypes: Simple Marker-Based (SMB) and Morphological Marker-Based (MMB) mechanisms. Similarly, the Intensity-Based Transduction Principle encompasses the Reflective Layer-based (RLB) and Transparent Layer-Based (TLB) mechanisms. This paper provides a comparative study of the hardware characteristics of these four types of sensors including various combination types, and discusses the commonly used methods for interpreting tactile information. This comparison reveals some current challenges faced by VBTS technology and directions for future research. In robotic systems, tactile sensing is fundamental for enabling robots to interact with their environment through physical contact. By delivering real-time tactile feedback, such as object stiffness, local force, slip and contact position feedback, this capability empowers robotic systems to achieve precise object manipulation while preventing damage [1]-[4]. CL, HL and YL were supported by the the China Scholarship Council and Bristol joint scholarship. EP and NL were supported by the Horizon Europe research and innovation program under grant agreement No. 101120823 (MANiBOT) and the Royal Society International Collaboration Awards (South Korea). NL was also supported by an award from ARIA on'Democratising Hardware And Control For Robot Dexterity'. Lepora) HL is with School of Robotics, Xi'an Jiaotong-Liverpool University, China, and was with the School of Engineering Mathematics and T ech-nology, and Bristol Robotics Laboratory, University of Bristol, Bristol, U.K. (Email: haoran.li@xjtlu.edu.cn). YL, CL, MY, EP, and NL are with the School of Engineering Mathematics and T echnology, and Bristol Robotics Laboratory, University of Bristol, Bristol, U.K. (Email: {yijiong.lin, Traditional electronic technologies such as piezoelectric and piezoresistive sensor arrays have been considered promising due to their high temporal resolution and thin profiles.


SuperMag: Vision-based Tactile Data Guided High-resolution Tactile Shape Reconstruction for Magnetic Tactile Sensors

Hou, Peiyao, Sun, Danning, Wang, Meng, Huang, Yuzhe, Zhang, Zeyu, Liu, Hangxin, Li, Wanlin, Jiao, Ziyuan

arXiv.org Artificial Intelligence

-- Magnetic-based tactile sensors (MBTS) combine the advantages of compact design and high-frequency operation but suffer from limited spatial resolution due to their sparse taxel arrays. This paper proposes SuperMag, a tactile shape reconstruction method that addresses this limitation by leveraging high-resolution vision-based tactile sensor (VBTS) data to supervise MBTS super-resolution. Co-designed, open-source VBTS and MBTS with identical contact modules enable synchronized data collection of high-resolution shapes and magnetic signals via a symmetric calibration setup. The MBTS achieves a sampling frequency of 125 Hz, whereas the shape reconstruction sustains an inference time within 2.5 ms. Tactile sensing is essential in robotics, enabling agents to perceive and interact with their environment through physical contact [1, 2]. Inspired by the biological sense of touch, tactile sensors detect mechanical stimuli such as contact force, texture, slip, and vibrations. Common sensing technologies include capacitive [3], resistive [4], piezoresistive [5], piezoelectric [6], triboelectric [7], barometric [8], optical [9], and magnetic [10] sensors, each offering unique advantages for tactile perception and robotic applications.


In-Hand Object Pose Estimation via Visual-Tactile Fusion

Nonnengießer, Felix, Kshirsagar, Alap, Belousov, Boris, Peters, Jan

arXiv.org Artificial Intelligence

-- Accurate in-hand pose estimation is crucial for robotic object manipulation, but visual occlusion remains a major challenge for vision-based approaches. This paper presents an approach to robotic in-hand object pose estimation, combining visual and tactile information to accurately determine the position and orientation of objects grasped by a robotic hand. We address the challenge of visual occlusion by fusing visual information from a wrist-mounted RGB-D camera with tactile information from vision-based tactile sensors mounted on the fingertips of a robotic gripper . Our approach employs a weighting and sensor fusion module to combine point clouds from heterogeneous sensor types and control each modality's contribution to the pose estimation process. We use an augmented Iterative Closest Point (ICP) algorithm adapted for weighted point clouds to estimate the 6D object pose. Our experiments show that incorporating tactile information significantly improves pose estimation accuracy, particularly when occlusion is high. Our method achieves an average pose estimation error of 7.5 mm and 16.7 degrees, outperforming vision-only baselines by up to 20%. We also demonstrate the ability of our method to perform precise object manipulation in a real-world insertion task. In-hand pose estimation describes the process of determining the position and orientation of an object held within a robotic hand.


A Comparative Study of Human Activity Recognition: Motion, Tactile, and multi-modal Approaches

Belcamino, Valerio, Le, Nhat Minh Dinh, Luu, Quan Khanh, Carfì, Alessandro, Ho, Van Anh, Mastrogiovanni, Fulvio

arXiv.org Artificial Intelligence

Human activity recognition (HAR) is essential for effective Human-Robot Collaboration (HRC), enabling robots to interpret and respond to human actions. This study evaluates the ability of a vision-based tactile sensor to classify 15 activities, comparing its performance to an IMU-based data glove. Additionally, we propose a multi-modal framework combining tactile and motion data to leverage their complementary strengths. We examined three approaches: motion-based classification (MBC) using IMU data, tactile-based classification (TBC) with single or dual video streams, and multi-modal classification (MMC) integrating both. Offline validation on segmented datasets assessed each configuration's accuracy under controlled conditions, while online validation on continuous action sequences tested online performance. Results showed the multi-modal approach consistently outperformed single-modality methods, highlighting the potential of integrating tactile and motion sensing to enhance HAR systems for collaborative robotics.


Quantitative Hardness Assessment with Vision-based Tactile Sensing for Fruit Classification and Grasping

Liao, Zhongyuan, Du, Yipai, Duan, Jianghua, Liang, Haobo, Wang, Michael Yu

arXiv.org Artificial Intelligence

Accurate estimation of fruit hardness is essential for automated classification and handling systems, particularly in determining fruit variety, assessing ripeness, and ensuring proper harvesting force. This study presents an innovative framework for quantitative hardness assessment utilizing vision-based tactile sensing, tailored explicitly for robotic applications in agriculture. The proposed methodology derives normal force estimation from a vision-based tactile sensor, and, based on the dynamics of this normal force, calculates the hardness. This approach offers a rapid, non-destructive evaluation through single-contact interaction. The integration of this framework into robotic systems enhances real-time adaptability of grasping forces, thereby reducing the likelihood of fruit damage. Moreover, the general applicability of this approach, through a universal criterion based on average normal force dynamics, ensures its effectiveness across a wide variety of fruit types and sizes. Extensive experimental validation conducted across different fruit types and ripeness-tracking studies demonstrates the efficacy and robustness of the framework, marking a significant advancement in the domain of automated fruit handling.


A Modularized Design Approach for GelSight Family of Vision-based Tactile Sensors

Agarwal, Arpit, Mirzaee, Mohammad Amin, Sun, Xiping, Yuan, Wenzhen

arXiv.org Artificial Intelligence

GelSight family of vision-based tactile sensors has proven to be effective for multiple robot perception and manipulation tasks. These sensors are based on an internal optical system and an embedded camera to capture the deformation of the soft sensor surface, inferring the high-resolution geometry of the objects in contact. However, customizing the sensors for different robot hands requires a tedious trial-and-error process to re-design the optical system. In this paper, we formulate the GelSight sensor design process as a systematic and objective-driven design problem and perform the design optimization with a physically accurate optical simulation. The method is based on modularizing and parameterizing the sensor's optical components and designing four generalizable objective functions to evaluate the sensor. We implement the method with an interactive and easy-to-use toolbox called OptiSense Studio. With the toolbox, non-sensor experts can quickly optimize their sensor design in both forward and inverse ways following our predefined modules and steps. We demonstrate our system with four different GelSight sensors by quickly optimizing their initial design in simulation and transferring it to the real sensors.


VET: A Visual-Electronic Tactile System for Immersive Human-Machine Interaction

Zhang, Cong, Yang, Yisheng, Mu, Shilong, Lyu, Chuqiao, Li, Shoujie, Chai, Xinyue, Ding, Wenbo

arXiv.org Artificial Intelligence

In the pursuit of deeper immersion in human-machine interaction, achieving higher-dimensional tactile input and output on a single interface has become a key research focus. This study introduces the Visual-Electronic Tactile (VET) System, which builds upon vision-based tactile sensors (VBTS) and integrates electrical stimulation feedback to enable bidirectional tactile communication. We propose and implement a system framework that seamlessly integrates an electrical stimulation film with VBTS using a screen-printing preparation process, eliminating interference from traditional methods. While VBTS captures multi-dimensional input through visuotactile signals, electrical stimulation feedback directly stimulates neural pathways, preventing interference with visuotactile information. The potential of the VET system is demonstrated through experiments on finger electrical stimulation sensitivity zones, as well as applications in interactive gaming and robotic arm teleoperation. This system paves the way for new advancements in bidirectional tactile interaction and its broader applications.


Enhance Vision-based Tactile Sensors via Dynamic Illumination and Image Fusion

Redkin, Artemii, Dugonjic, Zdravko, Lambeta, Mike, Calandra, Roberto

arXiv.org Artificial Intelligence

Vision-based tactile sensors use structured light to measure deformation in their elastomeric interface. Until now, vision-based tactile sensors such as DIGIT and GelSight have been using a single, static pattern of structured light tuned to the specific form factor of the sensor. In this work, we investigate the effectiveness of dynamic illumination patterns, in conjunction with image fusion techniques, to improve the quality of sensing of vision-based tactile sensors. Specifically, we propose to capture multiple measurements, each with a different illumination pattern, and then fuse them together to obtain a single, higher-quality measurement. Experimental results demonstrate that this type of dynamic illumination yields significant improvements in image contrast, sharpness, and background difference. This discovery opens the possibility of retroactively improving the sensing quality of existing vision-based tactile sensors with a simple software update, and for new hardware designs capable of fully exploiting dynamic illumination.


Tacchi 2.0: A Low Computational Cost and Comprehensive Dynamic Contact Simulator for Vision-based Tactile Sensors

Sun, Yuhao, Zhang, Shixin, Li, Wenzhuang, Zhao, Jie, Shan, Jianhua, Shen, Zirong, Chen, Zixi, Sun, Fuchun, Guo, Di, Fang, Bin

arXiv.org Artificial Intelligence

With the development of robotics technology, some tactile sensors, such as vision-based sensors, have been applied to contact-rich robotics tasks. However, the durability of vision-based tactile sensors significantly increases the cost of tactile information acquisition. Utilizing simulation to generate tactile data has emerged as a reliable approach to address this issue. While data-driven methods for tactile data generation lack robustness, finite element methods (FEM) based approaches require significant computational costs. To address these issues, we integrated a pinhole camera model into the low computational cost vision-based tactile simulator Tacchi that used the Material Point Method (MPM) as the simulated method, completing the simulation of marker motion images. We upgraded Tacchi and introduced Tacchi 2.0. This simulator can simulate tactile images, marked motion images, and joint images under different motion states like pressing, slipping, and rotating. Experimental results demonstrate the reliability of our method and its robustness across various vision-based tactile sensors.