Uppalapati, Naveen Kumar
Precision Harvesting in Cluttered Environments: Integrating End Effector Design with Dual Camera Perception
Koe, Kendall, Shah, Poojan Kalpeshbhai, Walt, Benjamin, Westphal, Jordan, Marri, Samhita, Kamtikar, Shivani, Nam, James Seungbum, Uppalapati, Naveen Kumar, Krishnan, Girish, Chowdhary, Girish
Abstract-- Due to labor shortages in specialty crop industries, a need for robotic automation to increase agricultural efficiency and productivity has arisen. Previous manipulation systems perform well in harvesting in uncluttered and structured environments. High tunnel environments are more compact and cluttered in nature, requiring a rethinking of the large form factor systems and grippers. We propose a novel codesigned framework incorporating a global detection camera and a local eye-in-hand camera that demonstrates precise localization of small fruits via closed-loop visual feedback and reliable error handling. Field experiments in high tunnels show our system can reach an average of 85.0% of cherry tomato fruit in 10.98s on average. I. INTRODUCTION Decreasing food miles and increasing sustainable agricultural practices have prompted interest in urban agriculture Figure 1: Robot picking cherry tomatoes with our Detect2Grasp in recent years.
Active Semantic Mapping with Mobile Manipulator in Horticultural Environments
Cuaran, Jose, Ahluwalia, Kulbir Singh, Koe, Kendall, Uppalapati, Naveen Kumar, Chowdhary, Girish
Semantic maps are fundamental for robotics tasks such as navigation and manipulation. They also enable yield prediction and phenotyping in agricultural settings. In this paper, we introduce an efficient and scalable approach for active semantic mapping in horticultural environments, employing a mobile robot manipulator equipped with an RGB-D camera. Our method leverages probabilistic semantic maps to detect semantic targets, generate candidate viewpoints, and compute corresponding information gain. We present an efficient ray-casting strategy and a novel information utility function that accounts for both semantics and occlusions. The proposed approach reduces total runtime by 8% compared to previous baselines. Furthermore, our information metric surpasses other metrics in reducing multi-class entropy and improving surface coverage, particularly in the presence of segmentation noise. Real-world experiments validate our method's effectiveness but also reveal challenges such as depth sensor noise and varying environmental conditions, requiring further research.
Visual Servoing for Pose Control of Soft Continuum Arm in a Structured Environment
Kamtikar, Shivani, Marri, Samhita, Walt, Benjamin, Uppalapati, Naveen Kumar, Krishnan, Girish, Chowdhary, Girish
For soft continuum arms, visual servoing is a popular control strategy that relies on visual feedback to close the control loop. However, robust visual servoing is challenging as it requires reliable feature extraction from the image, accurate control models and sensors to perceive the shape of the arm, both of which can be hard to implement in a soft robot. This letter circumvents these challenges by presenting a deep neural network-based method to perform smooth and robust 3D positioning tasks on a soft arm by visual servoing using a camera mounted at the distal end of the arm. A convolutional neural network is trained to predict the actuations required to achieve the desired pose in a structured environment. Integrated and modular approaches for estimating the actuations from the image are proposed and are experimentally compared. A proportional control law is implemented to reduce the error between the desired and current image as seen by the camera. The model together with the proportional feedback control makes the described approach robust to several variations such as new targets, lighting, loads, and diminution of the soft arm. Furthermore, the model lends itself to be transferred to a new environment with minimal effort.