Notice of retraction
Vol. 32, No. 8(2), S&M2292

ISSN (print) 0914-4935
ISSN (online) 2435-0869
Sensors and Materials
is an international peer-reviewed open access journal to provide a forum for researchers working in multidisciplinary fields of sensing technology.
Sensors and Materials
is covered by Science Citation Index Expanded (Clarivate Analytics), Scopus (Elsevier), and other databases.

Instructions to authors
English    日本語

Instructions for manuscript preparation
English    日本語


 Sensors and Materials
 1-23-3-303 Sendagi,
 Bunkyo-ku, Tokyo 113-0022, Japan
 Tel: 81-3-3827-8549
 Fax: 81-3-3827-8547

MYU Research, a scientific publisher, seeks a native English-speaking proofreader with a scientific background. B.Sc. or higher degree is desirable. In-office position; work hours negotiable. Call 03-3827-8549 for further information.

MYU Research

(proofreading and recording)

(translation service)

The Art of Writing Scientific Papers

(How to write scientific papers)
(Japanese Only)

Sensors and Materials, Volume 32, Number 4(1) (2020)
Copyright(C) MYU K.K.
pp. 1245-1259
S&M2174 Research Paper of Special Issue
Published in advance: February 5, 2020
Published: April 10, 2020

Image-similarity-based Convolutional Neural Network for Robot Visual Relocalization [PDF]

Li Wang, Ruifeng Li, Jingwen Sun, Hock Soon Seah, Chee Kwang Quah, Lijun Zhao, and Budianto Tandianus

(Received July 31, 2019; Accepted December 2, 2019)

Keywords: visual relocalization, CNN, image similarity

Convolutional neural network (CNN)-based methods, which train an end-to-end model to regress a six degree of freedom (DoF) pose of a robot from a single red–green–blue (RGB) image, have been developed to overcome the poor robustness of robot visual relocalization recently. However, the pose precision becomes low when the test image is dissimilar to training images. In this paper, we propose a novel method, named image-similarity-based CNN, which considers the image similarity of an input image during the CNN training. The higher the similarity of the input image, the higher precision we can achieve. Therefore, we crop the input image into several small image blocks, and the similarity between each cropped image block and training dataset images is measured by employing a feature vector in a fully connected CNN layer. Finally, the most similar image is selected to regress the pose. A genetic algorithm is utilized to determine the cropped position. Experiments on both open-source dataset 7-Scenes and two actual indoor environments are conducted. The results show that the proposed algorithm leads to better results and reduces large regression errors effectively compared with existing solutions.

Corresponding author: Ruifeng Li, Lijun Zhao

Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 International License.

Cite this article
Li Wang, Ruifeng Li, Jingwen Sun, Hock Soon Seah, Chee Kwang Quah, Lijun Zhao, and Budianto Tandianus, Image-similarity-based Convolutional Neural Network for Robot Visual Relocalization, Sens. Mater., Vol. 32, No. 4, 2020, p. 1245-1259.

Forthcoming Regular Issues

Forthcoming Special Issues

Special Issue on International Conference on BioSensors, BioElectronics, BioMedical Devices, BioMEMS/NEMS and Applications 2019 (Bio4Apps 2019) (1)
Guest editor, Hirofumi Nogami and Masaya Miyazaki (Kyushu University)
Conference website

Special Issue on Optical Sensors: Novel Materials, Approaches, and Applications
Guest editor, Yap Wing Fen (Universiti Putra Malaysia)

Special Issue on Intelligent Sensing Control Analysis, Optimization, and Automation (2)
Guest editor, Cheng-Chi Wang (National Chin-Yi University of Technology)

Special Issue on Geomatics Technologies for the Realization of Smart Cities (2)
Guest editor, He Huang and XiangLei Liu (Beijing University of Civil Engineering and Architecture)

Special Issue on Cyber–Physical Systems (CPS) and Internet of Things (IoT)
Guest editor, Yutaka Arakawa (Kyushu University)

Special Issue on Sensors and Materials Emerging Investigators in Japan
Guest editor, Tsuyoshi Minami (The University of Tokyo)

Copyright(C) MYU K.K. All Rights Reserved.