VISION BASED AUTOMATED FORMATION FOR
MULTI ROBOT COOPERATION
MOHD MUHAIMIN BIN MAHMOOD
UNIVERSITI TEKNOLOGI MALAYSIA
VISION BASED AUTOMATED FORMATION FOR
MULTI ROBOT COOPERATION
MOHD MUHAIMIN BIN MAHMOOD
A thesis submitted in fulfilment of the
requirement for the award of the degree of
Master of Engineering (Electrical)
Faculty of Electrical Engineering
Universiti Teknologi Malaysia
AUGUST 2016
iii
Specially dedicated to:
My beloved family, lecturers and all friends for their external support,
encouragement, and inspiration throughout my journey of education.
MAY ALLAH BLESS US
iv
ACKNOWLEDGEMENT
First of all, I would like to thank Allah SWT for giving me strength to complete
this project successfully on time. I would like to express my gratitude to my project
supervisor, Prof Dr Shamsudin H. M. Amin who willing to accept me as his student to
carry out project under him. Besides, I would like to thank Prof Dr Shamsudin for the
passionate guidance and advises he gave upon my throughout entire project. It would be
difficult to complete this project without his support and understanding.
However, I wish to deeply indebted to my family member for giving me spirit
along with the support assisting my project and throughout years in UTM. Their
blessings were the main effort for me to overcome all the hardships and obstacles that I
faced. Not forgotten to their hardworking to provide me financial support to ensure my
successfulness in this project.
Next, my sincerest appreciation goes to my friends and especially to my
laboratory mates who always giving me valuable guidance, suggestions, kindness, and
valuable time during the accomplishment of this project. Lastly, I would like to extend
my deep appreciation for the laboratory technicians for giving me as mush patient and
guidance and support to enhance my project. Thank you so much.
v
ABSTRACT
In a multi robot system, robots are required to cooperate with each other, and
therefore should have the ability to make their own decision based on multiple input
sensors not only from the robots, but also from nearby robots. The task of carrying
oversized objects of different shapes poses a challenge in selecting an appropriate group
formation. Hence, the main objective of this project is to establish an algorithm that
enables multi robot system to carry large load by automatically selecting the required
group formation to successfully execute the task. At first, a robot will need to identify
the shape of the object (oversized bar, rectangular, square or circular shapes). Then, the
robots will form a suitable formation to carry the object. There are two main problems in
this project. First, the capability of the robot to identify the shape of the object because
the object’s image will be a bit skew form the actual shape, due to the slanting angle of
the camera used to detect the shapes of the objects. The second challenge is maintaining
the formation of the robots, while carrying the load on top of the robots, to a specified
destination. A multi robot system, built in-house is used in the experiments to investigate
the performance of the algorithm proposed. Algorithms implemented in this project are
leader-follower and behaviour based strategy. One of the robots will operate as the
command giver or the leader to the other robots. The algorithm consists of
communication strategies and autonomous decision making capability. The robot will be
communicating with each other using Xbee wireless modules and extracting the
behaviour of the other robots. Sensors placed around the body of the robots are utilized
to detect their relative distance, and hence, used to maintain their formation, so as to
prevent the load from falling down. All the decisions are made by the robots
autonomously via the onboard controllers. The multi robot system is shown to be able to
autonomously determine the shape of the different oversized objects, thus appropriately
change into formations capable of transporting large objects to a specified destination
point autonomously, with no outside intervention.
vi
ABSTRAK
Dalam sistem berbilang robot, robot-robot perlu untuk saling bekerjasama,
dilengkapi dengan kemampuan membuat keputusan sendiri bergantung kepada peranti-
peranti input yang bukan sahaja daripada robot tersebut, bahkan daripada robot
berdekatan. Tugasan untuk membawa objek besar yang terdiri daripada beberapa bentuk
yang berlainan adalah mencabar dalam menentukan susunan robot yang sesuai. Objektif
utama projek ini adalah melaksanakan satu algoritma yang membolehkan berbilang
robot untuk mengangkut objek besar secara automatik dan menentukan susunan
kumpulan robot berdasarkan bentuk objek yang perlu diangkut. Pada peringkat awal,
sebuah robot perlu mengenal pasti bentuk objek yang akan diangkut (bentuk batang
panjang, segi empat tepat, segi empat sama atau bulatan). Seterusnya, robot-robot
tersebut akan membuat susunan yang sesuai dengan bentuk dan saiz objek, untuk
mengangkut objek tadi. Terdapat dua masalah utama dalam projek ini. Pertama,
kemampuan robot untuk mengenal pasti bentuk objek kerana imej objek adalah sedikit
senget daripada bentuk sebenar, berikutan sudut kamera tidak terus dari atas objek.
Cabaran kedua adalah mengekalkan susunan kumpulan robot tadi apabila objek diangkut
di atas robot, ke destinasi yang ditetapkan. Sistem berbilang robot yang dibina sendiri
digunakan dalam eksperimen ini bagi mengenal pasti kemampuan algoritma yang
dicadangkan. Algoritma yang digunakan dalam eksperimen ini adalah ketua-pengikut
dan strategi berdasarkan kelakuan. Satu robot akan berfungsi sebagai pemberi arahan
kepada robot lain. Algoritma ini ada strategi komunikasi dan kemampuan membuat
keputusan secara automatik. Robot-robot ini akan berhubung dengan menggunakan
modul tanpa wayar Xbee dan ekstrak kelakuan robot satu lagi. Peranti yang diletakkan
sekeliling badan robot diguna untuk mengetahui jarak relatif yang digunakan untuk
mengekalkan susunan robot bagi mengelakkan beban daripada jatuh. Semua keputusan
dibuat robot itu sendiri secara automatik dengan kawalan litar papan. Sistem berbilang
robot menunjukkan ia mampu secara automatik untuk menentukan bentuk objek besar
dan menentukan susunan yang mampu membawa objek tersebut daripada titik destinasi
yang ditentukan tanpa bantuan daripada luar.
vii
TABLE OF CONTENTS
CHAPTER TITLE PAGE
DECLARATION ii
DEDICATION iii
ACKNOWLEDGEMENT iv
ABSTRACT v
ABSTRAK vi
TABLE OF CONTENTS vii
LIST OF FIGURES xi
LIST OF TABLE xv
LIST OF APPENDICES xvi
1 INTRODUCTION 1
1.1 Overview 2
1.2 Problem Statement 4
1.3 Research Questions 4
1.4 Research Objectives 5
1.5 Scope of Research 5
1.6 Thesis Outline 6
viii
2 LITERATURE REVIEW 7
2.1 Introduction 8
2.2 Centralized Group Control 9
2.2.1 Leader-follower type cooperative
control
10
2.3 Environment Adaptive 12
2.3.1 Modular Robot 13
2.4 Image Comparing Algorithm 14
2.4.1 Sift Method 15
2.5 UTM Multi Agent 17
2.6 Summary 19
3 METHODOLOGY 21
3.1 Robot Software Programming 21
3.2 Robot Architecture 22
3.3 Robot Component 24
3.3.1 A Laptop 24
3.3.2 Webcam 25
3.3.3 Ultrasonic Range Rinder (MAXsonar
EZI)
25
3.3.4 Analog Distance Sensor (GP2YOA21) 26
3.3.5 Incremental Rotary Encoder 27
3.3.6 USB Hub and USB Hub Power 27
3.3.7 Vexta brushless motor with wheel and
motor driver
28
3.3.8 Battery Charger 29
3.3.9 SKXbee 29
3.3.10 Interface Free Controller, IFC 30
ix
3.3.11 Switches and LED 31
3.4 Automated Formation Selection using Vision
Based on Scale Invariant Formation Transform
(SIFT)
32
3.5 Robot Formation 36
3.5.1 Square Object 36
3.5.2 Triangular Object 38
3.5.3 Rectangular Bar 39
3.5.4 Circular Object 41
3.6 Summary 43
4 RESULTS AND DISCUSSION 45
4.1 Controller 45
4.2 Obstacle Anti-Colliding Behavior 46
4.3 Reaching Destination 47
4.4 Experiment 47
4.4.1 Experiment 1 48
4.4.2 Experiment 2 50
4.4.3 Experiment 3 53
4.4.4 Experiment 4 55
4.5 Object Shape Recognition 57
4.6 Robot Tracing 62
4.6.1 Square Shape Object 62
4.6.2 Triangular Shape Object 63
4.6.3 Rectangular Bar Shape Object 64
4.6.4 Circular Shape Object 65
4.7 Result Comparison with Theoretical Expectation 66
4.7.1 Square Object 66
x
4.7.2 Triangular Object 67
4.7.3 Rectangular Bar Object 68
4.7.4 Circular Object 69
4.8 Discussion 70
5 CONCLUSION 73
5.1 Conclusion 73
5.2 Achievement and Contribution 74
5.3 Future Work 74
REFERENCES 76
Appendices A-C 82
xi
LIST OF FIGURES
FIGURE NO. TITLE PAGE
2.1 Environment-adaptation using tilt sensors and 3D topology of
module with multiple tilt constraints
13
2.2 Top: A skeleton view of the modular robot. A: Pivot module; B:
Supporting group. A and B are jointly viewed as an agent.
Surface group C is viewed as an inter-agent link. Bottom: The
desired shape (level surface) is specified with a set of local
desired tilt angles (constraints).
14
2.3 Extracting all of the SIFT features that it can find in both
images, match point
17
2.4 UTM Multi Agent Robot 19
3.1 UTM Multi Agent Robot System Wiring 22
3.2 UTM Multi Agent Robot Front and Rear View 23
3.3 UTM Multi Agent Robot Laptop 24
3.4 UTM Multi Agent Robot Webcam 25
3.5 UTM Multi Agent Robot ultrasonic Range Finder 25
3.6 UTM Multi Agent Robot Analog Distance Sensor 26
3.7 UTM Multi Agent Robot Rotary Encoder 27
3.8 UTM Multi Agent Robot USB Hub 27
3.9 UTM Multi Agent Robot Brushless Motor 28
3.10 UTM Multi Agent Robot Battery Charger 29
3.11 UTM Multi Agent Robot SKXBee 29
xii
3.12 UTM Multi Agent Robot IFC 30
3.13 UTM Multi Agent Robot Top View 32
3.14 Environment-adaptation using tilt sensors and 3D topology of
module with multiple tilt constraints
33
3.15 Flow leader and follower robots 34
3.16 Limit switch position on top of multi robot 35
3.17 Sample of Square Object 36
3.18 Formation of the robot if the object is square shape 37
3.19 Sample of Triangular Object 38
3.20 Formation of the robots if the object is triangle shape 39
3.21 Sample of Rectangular Bar Object 40
3.22 Formation of the robots if the object is rectangular bar shape 40
3.23 Sample of Circular Object 41
3.24 Formation of the robots if the object is circular shape 42
4.1 Multi Robot Controller 46
4.2 Obstacle Avoidance Flowchart 46
4.3 Initial multi robot position 48
4.4 Leader robot inspect square object 48
4.5 Multi robot make suitable position 49
4.6 Square object is putted on the robots 49
4.7 Robots move forward 50
4.8 Robots safely arrived at finishing line 50
4.9 (a) Robot positioned in straight line
(b) Leader Robot recognize shape of the object
(c) Follower robot move beside leader robot
51
4.10 Triangular Object placed on robots 51
4.11 Robots move forward and maintain the formation 52
4.12 Robots successfully arrived at the destination 52
4.13 (a) Robot positioned in straight line
(b) Leader Robot recognize shape of the object
53
xiii
(c) Follower robot move beside leader robot
4.14 Rectangular Bar object placed on robots 54
4.15 Robot moved slowly to ensure the object not fall down 54
4.16 Robots successfully arrived at the destination 55
4.17 (a) Robot positioned in straight line
(b) Leader Robot recognize shape of the object
(c) Follower robot moved and make formation suitable for
circular object
56
4.18 Circular shape object putted on robots 56
4.19 (a) Robots moved toward destination
(b) Robots successfully arrived at the destination
57
4.20 (a) Image of square in computer database
(b) Image of square shape object capture by robot webcam
58
4.21 Sentences print on the laptop monitor when analyzing square
shape object
58
4.22 (a) Image of triangular in computer database
(b) Image of triangular shape object capture by robot webcam
59
4.23 Sentences print on the laptop monitor when analyzing triangular
shape object
59
4.24 (a) Image of rectangular in computer database
(b) Image of rectangular bar shape object capture by robot
webcam
60
4.25 Sentences print on the laptop monitor when analyzing
rectangular shape object
60
4.26 (a) Image of circular in computer database
(b) Image of circular shape object capture by robot webcam
61
4.27 Sentences print on the laptop monitor when analyzing circular
shape object
61
4.28 Tracing of UTM Multi Robot, when it detect square shape
object until it reach finishing line safely
62
4.29 Tracing of UTM Multi Robot, when it detect triangular shape 63
xiv
object until it reach finishing line safely
4.30 Tracing of UTM Multi Robot, when it detect rectangular bar
shape object until it reach finishing line safely
64
4.31 Tracing of UTM Multi Robot, when it detect circular shape
object until it reach finishing line safely
66
4.32 (a) Theoretical Square Shape object
(b) Image of square shape object capture by robot webcam
66
4.33 (a) Theoretical Square Shape Formation
(b) Actual UTM Multi Robot Formation for square shape object
67
4.34 (a) Theoretical Triangular Shape object
(b) Image of triangular shape object capture by robot webcam
67
4.35 (a) Theoretical Triangular Shape Formation
(b) Actual UTM Multi Robot Formation for triangular shape
object
68
4.36 (a) Theoretical Rectangular Bar Shape object
(b) Image of rectangular bar shape object capture by robot
webcam
68
4.37 (a) Theoretical Rectangular Bar Shape Formation
(b) Actual UTM Multi Robot Formation for rectangular bar
shape object
69
4.38 (a) Theoretical Circular Shape object
(b) Image of circular shape object capture by robot webcam
69
4.39 (a) Theoretical Square Shape Formation
(b) Actual UTM Multi Robot Formation for square shape object
70
xvi
LIST OF APPENDICES
APPENDIX TITLE PAGE
A UTM Multi Agent Robot Hardware 82
B UTM Multi Agent Robot Software – Leader 116
C UTM Multi Agent Robot Software – Follower 127
CHAPTER 1
INTRODUCTION
Distributed robotic system became an interesting field of study, which bloom
around recent millennium era because it is more convenient against single robot system
[1]. It can be classified to many areas such as communication, dealing with difficult task
[2], task allocation and control [3], localization [4], exploration, large object
transportation [5], motion coordination and biological inspiration. All these topics are
discussed in multi robot system.
Object transport and manipulation
Robots will be collaborated to carry, move or manipulate object. Most of the projects
only use simulation or software verification. The projects on object transportation
normally study on constrain motion, numbers of robot, mechanism to hold and grasp
load, undefined workspace and so on. There are also various methods in handling the
object [6]. It can be pushed, griped, and pull using rope and others.
Motion coordination
In this application, it involves path planning [7,8], traffic control, formation generation
and maintaining the formation. The issues that commonly come out are about target
2
tracking, target search and multi-robot docking behavior. The approach used normally
marking some checkpoints for the robot, using dynamic priority assignment, generation
of cooperative gait and vision based robot. [9]
Formation control is the popular subject in multi robot area of study where the robot
communicates and able to move to destination point in desired robot orientation. The
formation control could be a bit complex in term of stability and convergence when the
multi robot is tasked to transport an object to a target point.
Reconfiguration Robotics
In this research area, reconfiguration robotic is about multi robot that can be adjusted or
adjusting itself either manual or automatic depend on the current situation [10,11]. These
types of robot give a great versatility and robustness, because it can generate desire
shape to fulfill certain function which changes from time to time.
1.1 Overview
Since the past decades, many robots have been built to fulfill human needs and
interests. In movies also, the development of robotics have grown, for example the
movie ‘Transformer’. The movie has been made in many series, which indicates that
many people around the world like robots and machinery thing. Competitions involving
robots nowadays are numerous to improve the design of robotic system and to generate
new ideas in the robotics field such as ‘Robot Soccer’, ‘ABU Robocon’, ‘Robo Sumo’,
‘Robofest’ and many more.
There are many objectives on robotic research because of the usage of robot is
very wide. These include the usage in dangerous environment, complete difficult tasks
[4] given, assist human in daily work or routine, transport large objects [5], and doing
3
repeated task over and over again. In some cases use of a single robot is not sufficient
and can only cover very limited tasks. So, it requires more than one robot to work in
collaboration, so that the task will be completed successfully. In the real world
environment, the robots must alter their formation by simple navigation strategies [1],
behavior based control [10,11], virtual structures [7,8], formation vector [3], hierarchical
formation, omnidirectional vision [9], etc.
There is an increasing research interest in systems composed of multiple
autonomous mobile robots exhibiting cooperative behaviors. Groups of mobile robots
are constructed, with the aim of studying such issues as group architecture, resource
conflict, origin of cooperation, learning, and geometric problems. Below are some of
the advantages of using multi robot systems against single robot systems [2]:
1) A large range of task domains
2) Greater efficiency [12]
3) Improved system performance
4) Fault tolerance
5) Robustness [12]
6) Lower economic cost [13]
7) Ease of development
8) Distributed sensing and action
As yet, few applications of cooperative robotics have been reported, these
include cooperative manipulation [10], navigation, and planning that is used for a group
of airplane maneuvers [7]. Supporting theory is still in its formative stages [14].
4
1.2 Problem Statement
Navigation strategy, maneuvering, manipulation, and mission planning are main
problem in mobile research study [15]. The implementation of behavior-based system
rather than common mobility systems for a mobile robot field gives a challenging in part
of research aspect to this multi robot study. Existing UTM researches only deal with
robot navigation and load transportation. Designing controller for behavior based multi
robot system is challenging because of two factors [16]. Firstly, a behavior-based system
has possibility to act on its own because the surrounding changes very fast and the robot
has many inputs to be processed. Next, its algorithm model is not complete yet and
mostly still unknown. So, in order to design a controller for this dynamic system, it
needs to use some approaches depending on knowledge, or model. In the case of a robot,
this system consists of the robot itself and the environment in which it operates for
example in a laboratory. However, the situation is different when a robot is worked in
the real world environment with many things need to be considered and take into
account.
1.3 Research Questions
In this project, researcher tries to find out how the multi robots cooperate
together to complete a given task that changes depending on the environment or
stimulate condition? Will the robots able to differentiate the dissimilarity of the situation
and from that able to make good respond together, not just a robot?
5
1.4 Research Objective
Multi agent robot system consists of mobile robots with similar attribute and can
move freely. This team can be operated by minimum presence of two-robot. The
outcomes of this collaboration of robots are:
1) to bring an oversized object to a fixed defined destination using a leader-follower
strategy automatically.
2) to identify what formation that the robots will make, depending on the image of
the object that it captured without external support.
3) to design a behavior based controller (software) for a team of multi agent robots
to do a single task regarding communication and cooperation between both robot.
1.5 Scope of Research
In the limitation of the time of the project, author will state some of scopes of the
project in order for the project work perfectly and complete in the given time frame.
Firstly, in this project, UTM Multi Robot that is already available in Robotic Laboratory
is used. This is because it needs a vast amount of time and effort to come out with a
brand new robot.
Second, the technologies that have been used depend on what already build-in
with the UTM Multi Robot. The robot equipped with many sensors and system, but the
technologies are a bit outdated.
6
Next, the experiments to try out the early hypothesis have been carried out in the
Fakulti Kejuruteraan Elektrik, UTM Robotic Laboratory because the robot is heavy and
it easier to commence the experiment or doing modification on the spot.
Lastly, this project will use Visual Studio as programing software. Visual studio
enable programmer to interface the sensor with graphic on the screen. This attribute
indicates either the sensors working or not, by graphic interface on laptop.
1.6 Thesis Outline
This thesis consists of five chapters in total. In the second chapter, it will
describe the study of previous researches. It will discuss about leader follower technique
in multi robot system, environment adaptive method and image processing in
recognizing object that will be transported. The details about the robot used in this
project shall also be described in this chapter.
Next, Chapter 3 explains about the methodology of this project from hardware
part to software thing. It includes how the robot recognized the shape of the object in
front; gave command to the follower robot, formed suitable formation, and transported
load to the destination successfully.
Then results from the experiment are presented in Chapter 4. It also contains
discussion about the results, including the comparison of the result obtained with the
expected outcome (theory). The result will also be presented not just by the camera
snapshot in the experiment that been carried out. It also will be presented in trace form,
in order to clearly present what really happen in the experiment.
Last chapter, will conclude the project that has been done and proposed
suggestion for future project in multi robot system field of study.
76
REFERENCE
[1] Wang, P. K. C., “Navigation Strategies For Multiple Autonomous Mobile Robots Moving In
Formation”, J. Robotic Systems 8(2) (1991), pp. 177–195.
[2] Lynne, E.P., Tang, F., “Building Multirobot Coalitions Through Automated Task Solution
Synthesis”, Proceedings of the IEEE 94(7), 2006, pp. 1289–1305.
[3] Yamaguchi, H., Arai, T., and Beni, G., “A Distributed Control Scheme For Multiple Obotic
Vehicles To Make Group Formations”, Robotics Autonom. Systems 36(4), 2001, pp. 125–147.
[4] Calin, B., “Geometric Methods for Multi Robot Planning and Control”, Universiti of
Pennsyvania: Ph.D Thesis, 2003.
[5] M. Ridzuan A., “Development Of Reactive-Decentralized Control Algorithm For Multi Agent
Robotics System In Cooperative Task Achievement”, Universiti Teknologi Malaysia: Master
Thesis, 2003.
[6] Boon K. H., “Vehicle Following and Collision Avoidance Robot”, Degree Project,
UniversitiTeknologi Malaysia (UTM), Skudai, Johor, 2010.
[7] Beard, R. W., Lawton, J., and Hadaegh, F. Y., “A Coordination Architecture For Spacecraft
Formation Control”, IEEE Trans. Control Systems Technol. 9(6), 2001, pp. 777–790.
[8] Lewis, M. A. and Tan, K.-H., “High Precision Formation Control Of Mobile Robots Using
Virtual Structures”, Autonom. Robots 4(4), 1997, pp. 387–403.
[9] Vidal, R., Shakernia, O., and Sastry, S., “Formation Control of Nonholonomic Mobile Robots
with Omnidirectional Visual Serving and Motion Segmentation”, in: Proc. of IEEE Internat.
Conf. on Robotics and Automation, 2003, pp. 584–589.
77
[10] Balch, T. and Arkin, R. C., “Behavior-Based Formation Control for Multirobot Teams”, IEEE
Trans. Robotics Automat. 14(6), 1998, pp. 926–939.
[11] Balch, T. and Hybinette, M., “Social Potentials for Scalable Multi-Robot Formations”, in: IEEE
Internat. Conf. on Robotics and Automation (ICRA), 2000, 73-80.
[12] Parker, L. E., “Current State Of The Art In Distributed Autonomous Mobile Robotics”. In Parker,
L. E., Bekey, G., and Barhen, J., editors, Distributed Autonomous Robotic Systems, volume 4,
Springer, Tokio, 2000, pages 3–12.
[13] A. Isa A. Salameh, “Design and Construction of Robot Leg For Walking And Climbing Based on
Animal Locomotion Characteristics”, Universiti Teknologi Malaysia: Master Thesis, 2000.
[14] Tamio, A., Enrico, P., and Lynne, E. P., “Editorial: Advances in Multi – Robot Systems”, IEEE
Transactions on Robotics and Automation. 18(5), 2002, pp. 655-661.
[15] Rodeny. A. Brooks, “Cambrian Intelligence”, the Massachusetts Institute of Technology (MIT)
Press, 1986.
[16] Rosenblatt, Julio K. "DAMN: A Distributed Architecture for Mobile Navigation." Journal of
Experimental & Theoretical Artificial Intelligence. 9:2-3, 1997, pp. 339-360.
[17] ZhiDong W., Eiji N. and Takuji M., “Realizing Cooperative Object Manipulation using Multiple
Behavior-based Robots”, Proc. Intelligent Robot and Systems (IROS) 96, IEEE, 1996.
[18] Yang X., Watanabe K., Izumi K. and Kiguchi K., “A Decentralized Control System for
Cooperative Transportation by Multiple Non-Holonomic Mobile Robots”, International Journal
of Control, 77:10, 2004, pp. 949-963.
[19] Natsuki M., J. Ota, Aiyama, Y., Jun S., and Arai T.; “Cooperative Transport System with
Regrasping Car-like Mobile Robots”; Proc. Intelligent Robot and Systems (IROS) 97; IEEE,
1997.
[20] Ying W., and Clarence W. S.; “Cooperative Transportation by Multiple Robots with Machine
Learning”; 2006 IEEE Congress on Evolutionary Computation, Sheraton Vancouver Wall Centre
Hotel, Vancouver, BC, Canada, July 16-21, 2006.
78
[21] Inoue K., Ota J., and Arai T.; “Iterative Transportation by Multiple Mobile Robots Considering
Unknown Obstacles”; Journal of Robotics and Mechatronics Vol. 21 No.1, 2009; September 7,
2008.
[22] Yutaka I., Takahiro T. and Hitoshi I.; “Cooperative transportation system for humanoid robots
using simulation-based learning”; Applied Soft Computing 7, 2005 Elsevier, 2007, pp. 115–125.
[23] Kazuharu T. and Toshiyuki M.; “Load Distribution Control for Cooperative Transportation by
Multiple Mobile Robots”; Advance Motion Control (AMC) 2004 - Kawasaki, Japan , IEEE,
2004.
[24] Chin P. T., Rajankumar B., and Venkat K., “Decentralized Kinematic Control of Payload
Transport by a System of Mobile Manipulators”; Proceedings of the 2004 IEEE, International
Conference on Robotics & Automation, New Orleans, LA, April 2004.
[25] Mahitthidetch U. and Poj T., “Cooperative Overhead Transportation of a Box by Decentralized
Mobile Robots”; Robotics, Automation and Mechatronics (RAM) 2008, IEEE, 2008.
[26] Ahmad M.R., Amin S.H.M. and Mamat, R., “Development of Reactive Control Strategy for
Multi-Agent Mobile Robotics System”; Jurnal Teknologi, 35(D) Dis., Universiti Teknologi
Malaysia, 2001, pp. 99-114.
[27] ZhiDong W., Yugo T., Yasuhisa H., and Kazuhiro K.; “Decentralized Cooperative Object
Transportation by Multiple Mobile Robots with a Pushing Leader”, Springer 2008.
[28] Xin C. and Yangmin L.; “Cooperative Transportation by Multiple Mobile Manipulators Using
Adaptive NN Control”; 2006 International Joint Conference on Neural Networks, Sheraton
Vancouver Wall Centre Hotel, Vancouver, BC, Canada, July 16-21, 2006.
[29] Kazuhiro K., and Tomohiro O., “Decentralized Control of Multiple Robots Handling an Object”;
Proc. Intelligent Robot and Systems (IROS) 96, IEEE, 1996.
[30] Masakazu F., Wataru I., and Kazuhiro K., “Cooperative Control of Multiple Mobile Robots
Transporting a Single Object with Loose Handling”; Proceedings of the 2007 IEEE, International
Conference on Robotics and Biomimetics, December 15 -18, Sanya, China, 2007.
[31] Ahmad M.R., Amin S.H.M. and Mamat R., “Development of Decentralized Based Reactive
Control Strategy for Intelligent Multi-Agent Mobile-Robotics System”; Seventh international
79
Conference on Control, Automation, Robotics And Vision (ICARCV'OZ), Dec. 2002, Singapore;
Volume 1, 2-5 Dec., vol.1, 2002, pp. 220-227.
[32] Guilherme A. S. P., Bruno S. P., Luiz C., and Mario F. M. C., “Coordination of Multiple Mobile
Robots in an Object Carrying Task using Implicit Communication”; IEEE, International
Conference on Robotics & Automation Washington, DC, May 2002.
[33] Mahitthidetch U. and Poj T., “Cooperative Behavior-based Control of Decentralized Mobile
Robots on an Overhead Box Carrying Task”; Proceedings of International Conference on
Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology
(ECTI-CON) 2008; IEEE, 2008.
[34] Williams B. K. and Robert C. S., Carl D. Shapiro, “Adaptive Management”, The U.S.
Department of the Interior Technical Guide. US Department of the Interior. ISBN 1-4113-1760-
2, 2007.
[35] Larranaga. J., Muguira L., Lopez-Garde J. M. and Vazquez J. I., "An Environment Adaptive
Zigbee-based Indoor Positioning Algorithm", IEEE Indoor Positioning and Indoor Navigation
(IPIN), 2010 International Conference, Zurich, 2010, page 1-8.
[36] Willmott. S. and Faltings. B., "Benefits of Environment Adaptive Organisations for Agent
Coordination and Network Routing Problems", Multi Agent System, Proceedings. Forth
International Conference, Boston, 2000, pp. 333-340.
[37] Lacroix. S., Chatila R., Fleury S. Herrb M. and Simeon T., "Autonomous Navigation in Outdoor
Environment: Adaptive Approach and Experiment", Robotics and Automation, Proceedings.,
1994 IEEE International Conference, 1994, San Diego, CA, Vol. 1, 1994, pp. 426-432.
[38] Dongmei Z., Huadong M., Liang L. and Dan T., "EAAR: An Approach to Environment Adaptive
Application Reconfiguration in Sensor Network", Mobile Ad-hoc and Sensor Network, Volume
3794 of series Lecture Notes in Computer Science, pp. 259-268.
[39] C. Yu and R. Nagpal, “Sensing-based Shape Formation Tasks on Modular Multi-Robot
Systems”, A theoretical study. In Proc. Autonomous Agents and Multi Agent Systems
(AAMAS), 2008, pp. 71-78.
[40] Riesenhuber M. and Tomaso P., "Hierarchical Models of Object Recognition in Cortex." Nature
Neuroscience 2, no. 11, 1999, pp. 1019-1025.
80
[41] http://stackoverflow.com/questions/843972/image-comparison-fast-algorithm.
[42] Shotton J., Matthew J., and Roberto C., "Semantic Texton Forests for Image Categorization and
Segmentation." in Computer Vision and Pattern Recognition, 2008. Computer Vision and Pattern
Recognition (CVPR) 2008. IEEE Conference on, IEEE, 2008, pp. 1-8.
[43] Lowe, David G. "Three-dimensional Object Recognition from Single Two-dimensional Images."
Artificial intelligence, 31:3, 1987, pp. 355-395.
[44] Schmid C., and R. Mohr, "Local Grayvalue Invariants for Image Retrieval," IEEE Pattern
Analysis and Machine Intelligent (PAMI), Vol 19, No 5, 1997, pp. 149-167.
[45] Lowe D. G., "Object Recognition from Local Sscale Invariant Features," International
Conference on Computer Vision, Corfu, Greece, September 1999, pp. 1150-1157.
[46] Serre T., Kouh M., Cadieu C., Knoblich U., Kreiman G., and Poggio T., “A Theory of Object
Recognition: Computations and Circuits in the Feedforward Path of the Ventral Stream in
Primate Visual Cortex”, Computer Science and Artificial Intelligence Laboratory Technical
Report, MIT-CSAIL-TR-2005-082, the Massachusetts Institute of Technology’s Computer
Science and Artificial Intelligence Laboratory (MIT-CSAIL), December 19, 2005.
[47] Beis J., and Lowe D.G., “Shape Indexing using Approximate Nearest-Neighbour Search in High-
Dimensional Spaces”, Conference on Computer Vision and Pattern Recognition, Puerto Rico,
1997, pp. 1000–1006.
[48] Lowe D. G., “Local Feature View Clustering for 3D Object Recognition", Proc. of the IEEE
Conference on Computer Vision and Pattern Recognition, Kauai, Hawaii, December 2001, pp. 1-
7.
[49] Lowe D. G., “Distinctive Image Features from Scale-Invariant Keypoints”, International Journal
of Computer Vision, 60, 2, 2004, pp. 91-110.
[50] Chris M. C., https://chrisjmccormick.files.wordpress.com/2013/01/siftmatching.png, January 24,
2013.
[51] Sami AK., "Collaboration Strategies For Multi-Robot System (Robot Navigation)", Universiti
Teknologi Malaysia (UTM), PHD Thesis, 2011.