4
Object Tracking for Autonomous Mobile Robot based on Feedback of Monocular-vision Xiaogang Guo, Changhong Wang, Zhenshen Qu Space Control and Inertial Technology Research Center, Harbin Institute of Technology Xi Da Zhi Street 92, Harbin, 150001, People’s Republic of China Email: [email protected] Abstract - Object tracking is the key issue for autonomous mobile robot navigation. In this paper, a method of mobile robot object tracking is put forward based on feedback of monocular- vision. This method implements object tracking through identifying the special sign. And this method can also simulate the spacecraft RVD. Experiments show: the method is effective for object tracking of autonomous mobile robot navigation, as well as robustness against environment disturbance. Index Terms - Mobile Robot, Monocular-vision, Object Tracking, Image processing, RVD I. INTRODUCTION An autonomous mobile robot(AMR) is an intelligent machine system, which can run autonomously in the room or the outside. Now the research on AMR is still the hot field in the world since it is widely used in many fields such as military, space exploration, dangerous environment and other fields. The identifying and tracking of an object is one of the key tasks for an autonomous mobile robot. Many different kinds of sensors are used to mobile navigation system such as infrared ray, laser telemeter, sonar, and visual systems. The method of vision-based navigation is widely used now for mobile robot[5]. It has higher distinguishability for space and gray, the scope of detecting abroad, higher precision[2, 3, 4]. For a vision-driven mobile robot, it has a function of a vision system to provide most of navigation information. The conventional approach to vision-based object tracking is stereovision and computation of range from image disparity. Adoption of the monocular vision requires only a single imager and exploits mobility to perform this task, thus dramatically reducing the complexity of the image processing system[8]. The measurement of monocular gets the passive imaging of object through single camera, it implements measurement making use of the prior information of object and the model of geometric line[1]. In the system of object tracking for autonomous mobile, the method of monocular- vision predigests the structure of system maximally; it achieves the algorithm of object tracking easily, and it guarantees the real time of object tracking. Space craft rendezvous and docking(RVD) is a complex system of engineering. And further, we can use this mobile robot object tracking system to simulate the RVD system. In this system, the object can run freely, so we can use the object to replace the target spacecraft and the mobile robot for the chaser. So the spacecraft RVD can be easily replaced by the course of object tracking for autonomous mobile robot. Though this system is simpler than the RVD, it can be used for the first step of study. In the following research, we will put emphases on this. The remainder of this paper is organized as follows: In Section II, the component of object tracking system is presented, and each of four layers is explained detailedly. Section III is focused on the object identifying so that the changes in the system can be tracked more effectively. The algorithm of object tracking is discussed. In Section IV, the object tracking is carried out and the fuzzy control method is used. In Section V, it is the system implement and experiment, an AS-R robot is used for the flat of experiment. Finally, conclusions are drawn in Section VI. II. COMPONENT OF OBJECT TRACKING SYSTEM Component of the autonomous tracking experiment system is shown in the picture below. This system includes four parts: the bottom control system of mobile robot, object identifying system, autonomous control system, and remote console with wireless communication. Fig. 1. architecture of object tracking system for autonomous mobile robot The bottom control system is embedded in the mobile robot. It consists of the mechanical system and a motor control adapter. The robot is controlled by the motor control adapter. The speed of the left and right wheels at any time may be detected by counting the pulse from the left and right of micro sensors which detect the slots in the encoder disks. The object identifying system is made up of a monocular CCD camera and a Matrox II video card. The real-time image of object is captured by the CCD camera, and then it is sent to the autonomous control system through the conversion of Matrox II card. 467 1-4244-0737-0/07/$20.00 c 2007 IEEE

Object Tracking for Autonomous Mobile Robot Based on Feedback of Monocular-Vision

Embed Size (px)

Citation preview

Page 1: Object Tracking for Autonomous Mobile Robot Based on Feedback of Monocular-Vision

Object Tracking for Autonomous Mobile Robot based

on Feedback of Monocular-vision Xiaogang Guo, Changhong Wang, Zhenshen Qu

Space Control and Inertial Technology Research Center, Harbin Institute of Technology

Xi Da Zhi Street 92, Harbin, 150001, People’s Republic of China

Email: [email protected]

Abstract - Object tracking is the key issue for autonomous

mobile robot navigation. In this paper, a method of mobile robot

object tracking is put forward based on feedback of monocular-

vision. This method implements object tracking through

identifying the special sign. And this method can also simulate

the spacecraft RVD. Experiments show: the method is effective

for object tracking of autonomous mobile robot navigation, as

well as robustness against environment disturbance.

Index Terms - Mobile Robot, Monocular-vision, Object

Tracking, Image processing, RVD

I. INTRODUCTION

An autonomous mobile robot(AMR) is an intelligent

machine system, which can run autonomously in the room or

the outside. Now the research on AMR is still the hot field in

the world since it is widely used in many fields such as

military, space exploration, dangerous environment and other

fields. The identifying and tracking of an object is one of the

key tasks for an autonomous mobile robot. Many different

kinds of sensors are used to mobile navigation system such as

infrared ray, laser telemeter, sonar, and visual systems. The

method of vision-based navigation is widely used now for

mobile robot[5]. It has higher distinguishability for space and

gray, the scope of detecting abroad, higher precision[2, 3, 4].

For a vision-driven mobile robot, it has a function of a vision

system to provide most of navigation information. The

conventional approach to vision-based object tracking is

stereovision and computation of range from image disparity.

Adoption of the monocular vision requires only a single

imager and exploits mobility to perform this task, thus

dramatically reducing the complexity of the image processing

system[8]. The measurement of monocular gets the passive

imaging of object through single camera, it implements

measurement making use of the prior information of object

and the model of geometric line[1]. In the system of object

tracking for autonomous mobile, the method of monocular-

vision predigests the structure of system maximally; it

achieves the algorithm of object tracking easily, and it

guarantees the real time of object tracking.

Space craft rendezvous and docking(RVD) is a complex

system of engineering. And further, we can use this mobile

robot object tracking system to simulate the RVD system. In

this system, the object can run freely, so we can use the object

to replace the target spacecraft and the mobile robot for the

chaser. So the spacecraft RVD can be easily replaced by the

course of object tracking for autonomous mobile robot.

Though this system is simpler than the RVD, it can be used

for the first step of study. In the following research, we will

put emphases on this.

The remainder of this paper is organized as follows: In

Section II, the component of object tracking system is

presented, and each of four layers is explained detailedly.

Section III is focused on the object identifying so that the

changes in the system can be tracked more effectively. The

algorithm of object tracking is discussed. In Section IV, the

object tracking is carried out and the fuzzy control method is

used. In Section V, it is the system implement and

experiment, an AS-R robot is used for the flat of experiment.

Finally, conclusions are drawn in Section VI.

II. COMPONENT OF OBJECT TRACKING SYSTEM

Component of the autonomous tracking experiment system

is shown in the picture below. This system includes four

parts: the bottom control system of mobile robot, object

identifying system, autonomous control system, and remote

console with wireless communication.

Fig. 1. architecture of object tracking system for autonomous mobile robot

The bottom control system is embedded in the mobile robot.

It consists of the mechanical system and a motor control

adapter. The robot is controlled by the motor control adapter.

The speed of the left and right wheels at any time may be

detected by counting the pulse from the left and right of micro

sensors which detect the slots in the encoder disks.

The object identifying system is made up of a monocular

CCD camera and a Matrox II video card. The real-time image

of object is captured by the CCD camera, and then it is sent to

the autonomous control system through the conversion of

Matrox II card.

467

1-4244-0737-0/07/$20.00 c©2007 IEEE

Page 2: Object Tracking for Autonomous Mobile Robot Based on Feedback of Monocular-Vision

The autonomous control system is a most important part in

the system. It calculates the status information of the mobile

robot, and according to the information it gives commands to

the motor control adapter. The course of controlling is a

closed-loop control.

The remote wireless console system includes a remote

computer control center, a wireless access point and two

wireless adapters. The wireless console can control robot

manually.

III. OBJECT IDENTIFYING

A. the character of object

The object is the special pattern for image processing and

identifying, which is shown in the picture below.

Fig. 2. the mark of object

In the experiment system, we design the object that is a

target with three different radial concentric circles. Using the

fixed different color to distinguish the different circle, we

adopt black, white and black in this system, which has such

advantages. This object can expand or contract comparatively

with the center circle, and it doesn’t influence the algorithm

of image processing.

B. Algorithm of object tracking

As shown in the picture below, the CCD camera is fixed at

the upside of mobile robot. The background is complexity in

the visual field of camera, and various disturber as the

illumination exchange or the shadow of people and object[9].

Therefore, at the beginning of algorithm, it needs to pop out

the object for identifying. In this paper, the identifier of robot

above is a foundation for standing out the object.

Fig. 3. object identifying

To stand out the object, the common method is that we

transform gray image to binary image[7]. But the fixed value

of threshold is unacceptable, for the environment of mobile

robot changing. We choose adaptive threshold algorithm

based on average gray-scale of each block. This method can

give different threshold according to different background, in

order to stand out object and to eliminate local disturber.

The course of algorithm processing is shown in the picture

below.

Fig. 4. algorithm flow of adaptive threshold based on blocks of

average gray level

Firstly this algorithm divides the collection of 394*288

gray-image into m*m blocks, then calculates the average gray

level of each block and gives the corresponding binary

threshold. The value of m can be confirmed by itself.

In the experiment system, the global average value is

effective to filter the disturbance of the local, and the average

value of blocks is effective to stand out the object. But the

disturbance of the local influences the stability and rapidity of

algorithm severely. Thus in this system, we adopt the

approximate global average value in order to ensure the

stability of system.

Fig. 5. algorithm of ellipse drawing up

468 2007 Second IEEE Conference on Industrial Electronics and Applications

Page 3: Object Tracking for Autonomous Mobile Robot Based on Feedback of Monocular-Vision

Binary image can be acquired through the algorithm of

prior processing. Then it is necessary to draw up the curve in

binary image.

Because the ellipse is a special circle, and the circle can be

changed into the ellipse in the condition of none-envisage[6].

Based on the above considering, we choose the algorithm of

ellipse drawing up in the experiment system.

The course of processing is shown in the Fig.5.

C. Data of vision feedback

The algorithm of mobile robot control needs the references

including the center image coordinate of object (pixel) and

the radial of three different circles. These data present the

position and orientation information of object in general

conditions. In complex conditions, for instance, the angle is

partial between the visual angle of robot and the object, the

control program can give the special algorithm of searching

for tracking.

IV. OBJECT TRACKING

When the algorithm of tracking doesn’t need to have the

higher precision of location, it requires that object tracking be

quickly and steadily[10]. In this system, we choose the

monocular-vision so that the algorithm of distance measuring

is simple. So the system can not get the precise distance

information of the object. But the system gets the fuzzy

distance information through the radial variety of object. The

controlled variable is calculated by the polling list of

fuzzification. And it is chosen through the experiment at a

great level.

The basic structure of algorithm is shown in the picture

below.

Fig. 6. algorithm structure of tracking

When the object enters into the visual field of mobile robot,

namely vision feedback presents the data of location, the

algorithm of control implements the course of stable tracking.

The system acquires the precise position and fuzzy distance

information of object through processing the data of vision

feedback. So it can give the fuzzy control field: LEFT, MID,

RIGHT, TOOFAR, FAR, NEAR, in order to control robot

with closed-loop. According to the difference of control field,

the different algorithm of control is presented. So the robot

control is smooth, and the tracking is stable.

V. SYSTEM IMPLEMENT AND EXPERIMENT

In this system, the experiment object we choose is an AS-R

robot that it is the research version of the Power Strom robot

produced by the Shanghai Grandar Robotics Co. Ltd. Using

the CCD gray camera upside the mobile robot to locate the

object, we can get the position and orientation information of

robot. At the same time, we send the real-time data of position

and orientation to the control program of robot tracking, then

the algorithm of fuzzy control can control the position and

orientation of robot according to the different situation, at last

complete the experiment by tracking the object.

In the experiment, the object is placed 4 meters far from the

robot. By the remote control center, the robot was

commanded to implement the course of tracking. Firstly, if

the object was not in the visual scope of the robot, the robot

would rotate slowly to look for the object. Until the object

was found, the distance and visual angle between the robot

and the object was calculated. According to the information

of distance and orientation, the robot reaches to the object and

tracking it. The robot was controlled by the fuzzy control

method in this course.

VI. CONCLUSION

In this paper we present a method of object tracking for

autonomous mobile based on the feedback of monocular-

vision. This method can acquire the precise orientation and

fuzzy distance information of object, so as to implement the

object tracking for mobile robot through identifying the

special mark and processing the data of vision feedback. And

further, this system is the basic study for the behind research

of RVD. Proved by our experiments, this method can track

the object with real-time successfully, and it can also quickly

and precisely locate the object. So the simulation method of

mobile robot object tracking is effectively for the RVD

system. This method, which is able to restrain noise, is robust.

Thus, it is a reference for real-time tracking of autonomous

mobile robots system in complex environment.

REFERENCES

[1] M. Bertozzi, A. Broggi, and A. Fascioli, “Vision-based intelligent

vehicles: state of the art and perspectives,” Robotics and Autonomous

Systems, 2000 (3): 1 16Proceedings of 6th Conference on Optical 3-D

Measurement Techniques[C], Zurich: ETH, 2003.

[2] W. Choi, C. Ryu, and H. Kim, “Navigation of a mobile robot using

mono-vision and mono-audition,” Proceedings of the IEEE International

Conference on Systems, Man and Cybernetics, pp. IV-686 - IV-691, v4,

1999.

[3] Davidson, C. James, Hutchinson, and A. Seth, “Recognition of

Traversable Areas for Mobile Robotic Navigation in Outdoor

Environments,” IEEE International Conference on Intelligent Robots and

Systems, pp. 297-304, v1, 2003.

[4] R. Gregor, M. Lutzeler, and M. Pellkofer, “EMS-Vision: a perceptual

system for autonomous vehicles,” IEEE Transactions on Intelligent

Transportation Systems, pp. 48-59, 3 (1), 2002.

[5] X. G. Guo, Z. S. Qu, and B. Q. Xi, “Research on Location Method for

Mobile Robots Formation based on Global-camera,” ISSCAA2006, pp.

347-349, 2006.

[6] L. J. John, C. H. Declan, and P. W. Karim, “Vision-based navigation for

rendezvous, docking and proximity operations,” [R]. AAS, 99-102, 1999.

[7] Pettersen, Alf, Rotvold, Oyvind. Method and system for point

measurement of spatial coordinates [P]. U. S. Patent: 5440392, 1995.

2007 Second IEEE Conference on Industrial Electronics and Applications 469

Page 4: Object Tracking for Autonomous Mobile Robot Based on Feedback of Monocular-Vision

[8] M. Poteser, K. Kral, “Visual Distance Discrimination Between Stationary

Targets in Praying Mantis: an Index of The Use of Motion Parallax,” The

Journal of Experimental Biology 198, pp. 2127-2137, 1995.

[9] M. Subbaro, G. Surya, “Depth from defocus: a spatial domain approach,”

[J]. Int J Computer Vision, pp. 271 -294, 13 (3), 1994.

[10]H. Zhang, J. P. Ostrowski, “Visual motion planning for mobile robots,”

[J]. IEEE Transactions on Robotics and Automation, pp. 199-208, 18 (2),

2002.

470 2007 Second IEEE Conference on Industrial Electronics and Applications