PRISM: Probabilistic Real-Time Inference in Spatial World Models

Mirchev, Atanas, Kayalibay, Baris, Agha, Ahmed, van der Smagt, Patrick, Cremers, Daniel, Bayer, Justin

arXiv.org Artificial Intelligence 

Moving agents perceive streams of information, typically a mix of RGB images, depth and inertial measurements. Probabilistic generative models [1] are a principled way to formalise the synthesis of this data, and from these models inference can be derived through Bayes' rule. We focus on exactly such inference and target the agent states and the scene map, a problem known as simultaneous localisation and mapping (SLAM). We treat it as a posterior approximation for a given state-space model, such that the combination is useful for model-based control: the posterior inference serves as a state estimator and the predictive state-space model as a simulator with which to plan ahead [2]. To pave the way towards decision making, we believe an inference method should have: a compatible predictive model for both RGB-D images and 6-DoF dynamics; principled state and map uncertainty; real-time performance on commodity hardware; state-of-the-art localisation accuracy. We motivate these requirements further in appendix J. Prominent methods like LSD-SLAM [3], ORB-SLAM [4], DSO [5] have propelled visual SLAM forward, with heavy focus on large-scale localisation. The core of modern large-scale SLAM is maximum a-posteriori (MAP) smoothing in a probabilistic factor graph [6, 7]. At present this demands sparsity assumptions for computational feasibility, which obstructs the tight integration of dense maps and rendering.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found