Autonomous exploration of mobile robots through deep neural networks

文章来源

这是一篇选自《INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS》的文章。2017年July-August七八月份发表的:
July-August 2017: 1–9
DOI: 10.1177/1729881417703571
journals.sagepub.com/home/arx


通过DNN进行移动机器人的自动exploration
作者:Lei Tai [^1] , Shaohua Li [^2] and Ming Liu[^2]

摘要

机器人自主探索问题往往是解决机器人探索未知环境的需求。我们describe了一个室内探索算法,这个算法的hierarchical(adj. 按等级划分的; 等级制度的;)structure 混合了几种CNN层和decision-making process。The whole system is trained end to end by taking only visual information (RGB-D information) as input and generates a sequence of main moving direction as output so that the robot achieves autonomous exploration ability.The robot is a TurtleBot with a Kinect mounted on it. The model is trained and tested in a real world environment. And the training data set is provided for download. The outputs of the test data are compared with the human decision.我们使用Gaussian process latent variable model来可视化最后convolutional layer的feature map,以此证明了我们这种DNN形式探索的有效性。
我们同时展示了一个novel和lightweight的深度学习library libcnn来解决机器人的深度学习process。

Keywords

Robot exploration, deep learning, CNN

Date received: 28 July 2016; accepted: 23 February 2017

Introduction【背景】

----Previous approaches for exploration are mostly based on a probabilistic model and calculated cost maps, like the work of Stachniss et al1.The advantage of probability based models is that they take into consideration uncertainty to describe the real-world cases. However, most of these approaches only utilize geometric information without any cognitive process.【过去师兄概率模型和计算后的costmap,比如工作1,这样的好处是有考虑到现实世界的不确定性。但是这些东西大概率只是使用到了geometric information 但是没有使用cognitive process。】2 3 From the perspective of intelligence, dealing with input directly to generate output without further processing of input information is a kind of low-level intelligence. It will be more satisfactory if a mobile robot could imitate the way human beings deal with such a task.Fortunately, deep learning, with its advantage in hierarchical feature extraction,【分层特征提取】 provides a potential solution for this problem.

Motivation and bio-inspired perception【动机和生物启发的感知?】

【首先bb了一些ANN的知识,此处略去】 Regarding the decision-making, a recent work by Google DeepMind (http://www.deepmind.com) has shown that the decision-making process could be learned using a deep reinforcement learning model on the Q-functions.
至于决策,谷歌DeepMind一些工作表明DMP可以被深度强化学习(model on Q-functions的)所学习。4 5
Note that all these state-of-art methods tried to solve only one aspect of perception, or with strong assumptions on the distinctiveness(独特性) of the input features (e.g., pixels from a gameplay)
However, a practical robotic application is usually conducted with uncertainties in observations. This requirement makes it hard to design an effective and unified deep network that is able to realize a complete – though maybe seemingly simple – robotic task. Despite these considerations, we present in this article a unified deep network that is able to perform efficient and human-like robotic exploration in real time.【之前的工作都只有perception方面或者对输入的图像有强独立性的假设,我们就这些问题展现了一个同意的DNN 能够表现出高效且human-like的实时robotic exploration能力】
本工作试图将感知和控制与单一的深层网络相结合。proposed structure fuses convolutional neural network(CNN) with the decision-making process.

The CNN structure is used to detect and comprehend visual features and the fully connected layers are for decision-making. Except for that, no additional modules or algorithms are required for the execution of the task. Overall, we present a complete solution to the autonomous exploration based on a single network structure.

Although there are several libraries for deep learning in computer vision and speech
recognition, we still need an ultra-fast and reliable library for robotic applications.
As part of the contributions, we hereby present a novel deep-learning library – libcnn,which is optimized for robotics in terms of lightweight and flexibility. It is used to support the implementation of all the related modules in this article

Contributions

We address the following contributions of this article:
 We present a novel deep-learning library especially
optimized for robotic applications. Its featured modules include scene labelling, object recognition and decision-making.
 We present a deep-network solution towards humanlike exploration for a mobile robot. It results in high similarity between the robotic and human decisions,
leading to effective and efficient robotic exploration. This is the first work to en-couple both robotic perception and control in a real environment with a single network.
 A large indoor corridor data set with human decision label is provided for download.
 The result of a test is compared with human decision quantitatively and visualization of feature maps are shown by Gaussian process latent variable model (GPLVM).

Related Work

Following dropout, a more generalized algorithm,namely ‘Drop-Connect’ was proposed.Instead of dropping out nodes in a network, a drop-connect network drops out
connections, that is, weights of a neural network and proves that the dropping of nodes is just a special case of the proposed network.
比起drop out 是drop层中的node,drop-connect操作是将部分连接drop掉。drop out去掉节点只是drop-connect的一个特例。
2014年 a fully CNN was proposed by Long6that highly reduces the computation redundancy and could be adapted to inputs of an arbitrary size.

然后就开始介绍自己的模型了,就是简单的用CNN加上soft_max这样输出不同方向的概率,然后不同的概率还更具不同的cofidence进行加权。

Experiments and results

就turtlebot和没有GPU的interCPU电脑

Human instruction and data gathering

就人开turtlebot然后收集指令和画面

网络设置

原来的depth map size from Kinect is 640 x480.input depth map is first downsampled to 1/4 of the original size, that is, 160 x120 。 这在很大程度上降低了计算成本,而不引入许多错误操作。The first convolution layer uses 32 convolution kernels of size 5 x5, followed by a ReLU layer and a 2 x2 pooling layer with stride 2.The second stage of convolution+ activation + pooling is the same as the first stage .For the third stage, 64 convolutional kernels of size 5 x are used, with no change of the ReLU layer and pooling layer .This results in 64 feature maps of size 20 x15.

Sample results and evaluation

为了评价代理决策和人类决策的相似性,我们在以下两个方面对我们的方法进行了评价:第一,机器人生成命令和人的控制命令的一致性 感知命令;其次,代理探索轨迹与人类探索轨迹的相似性



  1. ↩︎ ↩︎
  2. ↩︎
  3. ↩︎
  4. ↩︎
  5. ↩︎
  6. ↩︎

你可能感兴趣的:(论文相关)