Schneider, Eric
Toward Semantic Scene Understanding for Fine-Grained 3D Modeling of Plants
Qadri, Mohamad, Freeman, Harry, Schneider, Eric, Kantor, George
Agricultural robotics is an active research area due to global population growth and expectations of food and labor shortages. Robots can potentially help with tasks such as pruning, harvesting, phenotyping, and plant modeling. However, agricultural automation is hampered by the difficulty in creating high resolution 3D semantic maps in the field that would allow for safe manipulation and navigation. In this paper, we build toward solutions for this issue and showcase how the use of semantics and environmental priors can help in constructing accurate 3D maps for the target application of sorghum. Specifically, we 1) use sorghum seeds as semantic landmarks to build a visual Simultaneous Localization and Mapping (SLAM) system that enables us to map 78\\% of a sorghum range on average, compared to 38% with ORB-SLAM2; and 2) use seeds as semantic features to improve 3D reconstruction of a full sorghum panicle from images taken by a robotic in-hand camera.
3D Skeletonization of Complex Grapevines for Robotic Pruning
Schneider, Eric, Jayanth, Sushanth, Silwal, Abhisesh, Kantor, George
Robotic pruning of dormant grapevines is an area of active research in order to promote vine balance and grape quality, but so far robotic efforts have largely focused on planar, simplified vines not representative of commercial vineyards. This paper aims to advance the robotic perception capabilities necessary for pruning in denser and more complex vine structures by extending plant skeletonization techniques. The proposed pipeline generates skeletal grapevine models that have lower reprojection error and higher connectivity than baseline algorithms. We also show how 3D and skeletal information enables prediction accuracy of pruning weight for dense vines surpassing prior work, where pruning weight is an important vine metric influencing pruning site selection.
3D Reconstruction-Based Seed Counting of Sorghum Panicles for Agricultural Inspection
Freeman, Harry, Schneider, Eric, Kim, Chung Hee, Lee, Moonyoung, Kantor, George
In this paper, we present a method for creating high-quality 3D models of sorghum panicles for phenotyping in breeding experiments. This is achieved with a novel reconstruction approach that uses seeds as semantic landmarks in both 2D and 3D. To evaluate the performance, we develop a new metric for assessing the quality of reconstructed point clouds without having a ground-truth point cloud. Finally, a counting method is presented where the density of seed centers in the 3D model allows 2D counts from multiple views to be effectively combined into a whole-panicle count. We demonstrate that using this method to estimate seed count and weight for sorghum outperforms count extrapolation from 2D images, an approach used in most state of the art methods for seeds and grains of comparable size.
Toward Human/Multi-Robot Systems to Support Emergency Services Agencies
Sklar, Elizabeth (University of Liverpool) | Schneider, Eric (University of Liverpool) | Ozgelen, A. Tuna (City University of New York) | Azhar, M. Q. (City University of New York)
The ability to make decisions that balance conflicting needs and variable-quality inputs is a skill that is inherently human. In emergency situations, such capabilities are tested under pressure, as needs and inputs change---often rapidly---and deliberation must take place quickly or else opportunities are lost. This short paper identifies challenges faced when emergency services personnel are supported by human/multi-robot systems. Several strategies are proposed to address these challenges, with deployment geared toward emergency services agencies within the next 5-10 years.
Learning to Avoid Collisions
Sklar, Elizabeth (Brooklyn College, City University of New York) | Parsons, Simon (Brooklyn College, City University of New York) | Epstein, Susan L. (Hunter College, City University of New York) | Ozgelen, Arif Tuna (The Graduate Center, City University of New York) | Munoz, Juan Pablo (The Graduate Center, City University of New York) | Abbasi, Farah (College of Staten Island, City University of New York) | Schneider, Eric (Hunter College, City University of New York) | Costantino, Michael (College of Staten Island, City University of New York)
Members of a multi-robot team, operating within close quarters, need to avoid crashing into each other. Simple collision avoidance methods can be used to prevent such collisions, typically by computing the distance to other robots and stopping, perhaps moving away, when this distance falls below a certain threshold. While this approach may avoid disaster, it may also reduce the team's efficiency if robots halt for a long time to let others pass by or if they travel further to move around one another. This paper reports on experiments where a human operator, through a graphical user interface, watches robots perform an exploration task. The operator can manually suspend robots' movements before they crash into each other, and then resume their movements when their paths are clear. Experiment logs record the robots' states when they are paused and resumed. A behavior pattern for collision avoidance is learned, by classifying the states of the robots' environment when the human operator issues "wait" and "resume" commands. Preliminary results indicate that it is possible to learn a classifier which models these behavior patterns, and that different human operators consider different factors when making decisions about stopping and starting robots.