Learning a Compact State Representation for Navigation Tasks by Autoencoding 2D-Lidar Scans

by   Christopher Gebauer, et al.

In this paper, we address the problem of generating a compact representation of 2D-lidar scans for reinforcement learning in navigation tasks. By now only little work focuses on the compactness of the provided state, which is a necessary condition to successfully and efficiently train a navigation agent. Our approach works in three stages. First, we propose a novel preprocessing of the distance measurements and compute a local, egocentric, binary grid map based on the current range measurements. We then autoencode the local map using a variational autoencoder, where the latent space serves as state representation. An important key for a compact and, at the same time, meaningful representation is the degree of disentanglement, which describes the correlation between each latent dimension. Therefore, we finally apply state-of-the-art disentangling methods to improve the representation power. Furthermore, we investige the possibilities of incorporating time-dependent information into the latent space. In particular, we incorporate the relation of consecutive scans, especially ego-motion, by applying a memory model. We implemented our approach in python using tensorflow. Our datasets are simulated with pybullet as well as recorded using a slamtec rplidar A3. The experiments show the capability of our approach to highly compress lidar data, maintain a meaningful distribution of the latent space, and even incorporate time-depended information.


Deep Generative Modeling of LiDAR Data

Building models capable of generating structured output is a key challen...

Latent Image Animator: Learning to Animate Images via Latent Space Navigation

Due to the remarkable progress of deep generative models, animating imag...

Mapless Humanoid Navigation Using Learned Latent Dynamics

In this paper, we propose a novel Deep Reinforcement Learning approach t...

IR-MCL: Implicit Representation-Based Online Global Localization

Determining the state of a mobile robot is an essential building block o...

DSLR: Dynamic to Static LiDAR Scan Reconstruction Using Adversarially Trained Autoencoder

Accurate reconstruction of static environments from LiDAR scans of scene...

Learning a Representation Map for Robot Navigation using Deep Variational Autoencoder

The aim of this work is to use Variational Autoencoder (VAE) to learn a ...

Laser2Vec: Similarity-based Retrieval for Robotic Perception Data

As mobile robot capabilities improve and deployment times increase, tool...

Please sign up or login with your details

Forgot password? Click here to reset