toward image-based localization for aibo using wavelet transform department of information...

1
Toward image-based localization for AIBO using wavelet transform Department of Information Engineering University of Padua, Italy A. Pretto, E. Menegatti, E. Pagello, Y. Jitsukawa, R. Ueda, T. Arai ime™ and a essed) decompressor to see this picture. Dept. of Precision Engineering University of Tokyo, Japan Abstract This paper describes a similarity measure to be used in image-based localization by autonomous robots with low computational resources. We propose a novel signature which allows memory saving and fast similarity calculation. The signature is based on the calculation of the 2D Haar Wavelet Transform of the gray- level image. We present experiments showing the effectiveness of the proposed image similarity measure. The used images were collected using the AIBOs ERS-7 of the RoboCup Team Araibo of the University of Tokyo on a RoboCup field. However, the proposed image similarity measure does not use any information on the structure of the environment and do not exploit the peculiar features of the RoboCup environment. Introduction Any techniques of image-based localization has to solve: (i)how to reduce the number of images necessary to fully describe the environment; (i)how to efficiently store a large data set of reference images (it is common to have several hundred reference images for typical environments); (ii)how to calculate in a fast and efficient way the similarity Each of the referred works addressed one (or all) of these problems. References 1. R. Cassinis, D. Duina, S. Inelli, and A. Rizzi. Unsupervised matching of visual landmarks for robotic homing using Fourier-Mellin transform. Robotics and Autonomous Systems, 40(2-3), August 2002. 4. M. N. Do and M. Vetterli. Wavelet-based texture retrieval using generalized Gaussian density and Kullback-Leibler distance. IEEE Tr. Im. Proc., 11(2):146–158, February 2002. 6. E. Frontoni, P. Zingaretti. An efficient similarity metric for omnidirectional vision sensors. Robotics and Autonomous Systems, 54(9):750–757, 2006. 7. J. Gaspar, N. Winters, and J. Santos-Victor. Vision- based navigation and environmental representations with an omnidirectional camera. IEEE Transaction on Robotics and Automation, Vol 16(number 6), Dec. 2000. 8. H.-M. Gross, A. Koenig, Ch. Schroeter, and H.-J. Boehme. Omnivision-based prob-abilistic self-localization for a mobile shopping assistant continued. In IEEE/RSJ Int. Conference on Intelligent Robots and Systems (IROS 2003), pages pp. 1505–1511, October 2003, Las Vegas USA. 11. B.J.A. Koerse, N. Vlassis, R. Bunschoten, and Y. Motomura. A probabilistic model for appareance-based robot localization. Image and Vision Computing, vol.19(6):pp. 381–391, April 2001. 12. Emanuele Menegatti, Takeshi Maeda, and Hiroshi Ishiguro. Image-based memory for robot navigation using properties of the omnidirectional images. Robotics and Autonomous Systems, Elsevier, 47(4):pp. 251–267, July 2004. 16. M. Vetterli and J Kovacevic. Wavelets and Subband Our approach To minimize the reference images to be stored, we keep as reference images two 180 degree panoramic views of the environment at every reference location We developed an algorithm that allows the ERS-7 robot to autonomously build two 180 degrees panorama images using its standard camera and to stitch them together. We use as image signature a 2-D Haar Wavelet Transform of the grey-level values of the image. We decide to stop at 4-th level decomposition, and to characterize images only by the detailed coefficients (horizontal, vertical and diagonal) of this level. The problem How to implement an image-based localization approach on a robot with limited storage memory and limited computational resources, as the Sony AIBO ERS-7? How to store in a memory-saving way the reference images? How to efficiently compare them with the input images? Which feature should we relay on, for a general approach not targeted to a particular environment? QuickTime™ and a TIFF (Uncompressed) decompressor are needed to see this picture. Wavelet signature We use as image signature a 2-D Haar Wavelet Transform of the grey-level values of the image. We decide to stop at 4-th level decomposition, and to characterize images only by the detailed coefficient (horizontal, vertical and diagonal) of this level. Haar Wavelet because: very effective in detecting discontinuity (one of the most important features in image-based localization) easily implemented and very fast to compute If one is interested in image reconstruction phase, the Haar Wavelets are not the good choice, because they tend to produce a lot of squared artifacts Memory saving: a gray-scale omnidirectional image 720x160 = 115.2 Kbyte ==> only 1.3 Kbyte. Experiments We tested the system in a RoboCup Four-legged League 540x360 cm soccer field, using a grid of 13 by 9 reference images. However, the proposed systems do not relay on RoboCup features - 90° +90° Input Image best match ==> second best match ==> <== Input Images <== resulting probability distribution 2-D Haar Wavelet

Upload: gavin-mccoy

Post on 02-Jan-2016

229 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Toward image-based localization for AIBO using wavelet transform Department of Information Engineering University of Padua, Italy A. Pretto, E. Menegatti,

Toward image-based localization for AIBO using wavelet transform

Department of Information Engineering

University of Padua, Italy

A. Pretto, E. Menegatti, E. Pagello, Y. Jitsukawa, R. Ueda, T. Arai

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

Dept. of Precision Engineering University of Tokyo, Japan

Abstract

This paper describes a similarity measure to be used in image-based localization by autonomous robots with low computational resources.

We propose a novel signature which allows memory saving and fast similarity calculation. The signature is based on the calculation of the 2D Haar Wavelet Transform of the gray-level image.

We present experiments showing the effectiveness of the proposed image similarity measure. The used images were collected using the AIBOs ERS-7 of the RoboCup Team Araibo of the University of Tokyo on a RoboCup field. However, the proposed image similarity measure does not use any information on the structure of the environment and do not exploit the peculiar features of the RoboCup environment.

Introduction

Any techniques of image-based localization has to solve: (i) how to reduce the number of images necessary to fully describe the environment; (i) how to efficiently store a large data set of reference images

(it is common to have several hundred reference images for typical environments);

(ii) how to calculate in a fast and efficient way the similarity

Each of the referred works addressed one (or all) of these problems.

References1. R. Cassinis, D. Duina, S. Inelli, and A. Rizzi. Unsupervised matching of visual

landmarks for robotic homing using Fourier-Mellin transform. Robotics and Autonomous Systems, 40(2-3), August 2002.

4. M. N. Do and M. Vetterli. Wavelet-based texture retrieval using generalized Gaussian density and Kullback-Leibler distance. IEEE Tr. Im. Proc., 11(2):146–158, February 2002.

6. E. Frontoni, P. Zingaretti. An efficient similarity metric for omnidirectional vision sensors. Robotics and Autonomous Systems, 54(9):750–757, 2006.

7. J. Gaspar, N. Winters, and J. Santos-Victor. Vision-based navigation and environmental representations with an omnidirectional camera. IEEE Transaction on Robotics and Automation, Vol 16(number 6), Dec. 2000.

8. H.-M. Gross, A. Koenig, Ch. Schroeter, and H.-J. Boehme. Omnivision-based prob-abilistic self-localization for a mobile shopping assistant continued. In IEEE/RSJ Int. Conference on Intelligent Robots and Systems (IROS 2003), pages pp. 1505–1511, October 2003, Las Vegas USA.

11. B.J.A. Koerse, N. Vlassis, R. Bunschoten, and Y. Motomura. A probabilisticmodel for appareance-based robot localization. Image and Vision Computing,

vol.19(6):pp. 381–391, April 2001.12. Emanuele Menegatti, Takeshi Maeda, and Hiroshi Ishiguro. Image-based

memory for robot navigation using properties of the omnidirectional images. Robotics and Autonomous Systems, Elsevier, 47(4):pp. 251–267, July 2004.

16. M. Vetterli and J Kovacevic. Wavelets and Subband Coding. Signal Processing Series. 1995.

17. J. Wolf, W. Burgard, and H. Burkhardt. Robust vision-based localization by combining an image retrieval system with monte carlo localization. IEEE Transactions on Robotics, 21(2):208–216, 2005.

Our approach

To minimize the reference images to be stored, we keep as reference images two 180 degree panoramic views of the environment at every reference location

We developed an algorithm that allows the ERS-7 robot to autonomously build two 180 degrees panorama images using its standard camera and to stitch them together.

We use as image signature a 2-D Haar Wavelet Transform of the grey-level

values of the image. We decide to stop at 4-th level decomposition, and to characterize images only by the detailed coefficients (horizontal, vertical and diagonal) of this level.

The problem

• How to implement an image-based localization approach on a robot with limited storage memory and limited computational resources, as the Sony AIBO ERS-7?

• How to store in a memory-saving way the reference images?

• How to efficiently compare them with the input images?

• Which feature should we relay on, for a general approach not targeted to a particular environment?

QuickTime™ and aTIFF (Uncompressed) decompressor

are needed to see this picture.

Wavelet signature

We use as image signature a 2-D Haar Wavelet Transform of the grey-level values of the image. We decide to stop at 4-th level decomposition, and to characterize images only by the detailed coefficient (horizontal, vertical and diagonal) of this level.

Haar Wavelet because:• very effective in detecting discontinuity (one of the most important features in

image-based localization)• easily implemented and very fast to compute

If one is interested in image reconstruction phase, the Haar Wavelets are not the good choice, because they tend to produce a lot of squared artifacts

Memory saving: a gray-scale omnidirectional image 720x160 = 115.2 Kbyte ==> only 1.3 Kbyte.

Experiments

We tested the system in a RoboCup Four-legged League 540x360 cm soccer field, using a grid of 13 by 9 reference images.

However, the proposed systems do not relay on RoboCup features

-90° +90°

Input Image

best match ==>

second best match ==>

<== Input Images

<== resulting probability distribution

2-D Haar Wavelet